var/home/core/zuul-output/0000755000175000017500000000000015067354264014541 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015067365432015504 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000004732157015067365425017723 0ustar rootrootOct 02 01:46:49 crc systemd[1]: Starting Kubernetes Kubelet... Oct 02 01:46:50 crc restorecon[4672]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 01:46:50 crc restorecon[4672]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 01:46:50 crc restorecon[4672]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Oct 02 01:46:51 crc kubenswrapper[4885]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 02 01:46:51 crc kubenswrapper[4885]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Oct 02 01:46:51 crc kubenswrapper[4885]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 02 01:46:51 crc kubenswrapper[4885]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 02 01:46:51 crc kubenswrapper[4885]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Oct 02 01:46:51 crc kubenswrapper[4885]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.713800 4885 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.721141 4885 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.721184 4885 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.721200 4885 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.721212 4885 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.721222 4885 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.721234 4885 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.721245 4885 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.721255 4885 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.721300 4885 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.721313 4885 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.721326 4885 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.721337 4885 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.721348 4885 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.721356 4885 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.721367 4885 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.721375 4885 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.721385 4885 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.721395 4885 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.721405 4885 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.721415 4885 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.721425 4885 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.721436 4885 feature_gate.go:330] unrecognized feature gate: Example Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.721446 4885 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.721456 4885 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.721466 4885 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.721475 4885 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.721483 4885 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.721491 4885 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.721501 4885 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.721510 4885 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.721520 4885 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.721530 4885 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.721540 4885 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.721550 4885 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.721560 4885 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.721571 4885 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.721582 4885 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.721591 4885 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.721601 4885 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.721611 4885 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.721621 4885 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.721630 4885 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.721640 4885 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.721650 4885 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.721662 4885 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.721671 4885 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.721682 4885 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.721692 4885 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.721707 4885 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.721720 4885 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.721732 4885 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.721746 4885 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.721760 4885 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.721771 4885 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.721779 4885 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.721787 4885 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.721796 4885 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.721804 4885 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.721812 4885 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.721820 4885 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.721828 4885 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.721836 4885 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.721844 4885 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.721852 4885 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.721861 4885 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.721869 4885 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.721877 4885 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.721885 4885 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.721893 4885 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.721901 4885 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.721909 4885 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.723116 4885 flags.go:64] FLAG: --address="0.0.0.0" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.723140 4885 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.723154 4885 flags.go:64] FLAG: --anonymous-auth="true" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.723167 4885 flags.go:64] FLAG: --application-metrics-count-limit="100" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.723179 4885 flags.go:64] FLAG: --authentication-token-webhook="false" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.723189 4885 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.723201 4885 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.723212 4885 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.723224 4885 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.723233 4885 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.723243 4885 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.723253 4885 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.723285 4885 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.723294 4885 flags.go:64] FLAG: --cgroup-root="" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.723303 4885 flags.go:64] FLAG: --cgroups-per-qos="true" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.723312 4885 flags.go:64] FLAG: --client-ca-file="" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.723321 4885 flags.go:64] FLAG: --cloud-config="" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.723329 4885 flags.go:64] FLAG: --cloud-provider="" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.723338 4885 flags.go:64] FLAG: --cluster-dns="[]" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.723349 4885 flags.go:64] FLAG: --cluster-domain="" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.723358 4885 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.723367 4885 flags.go:64] FLAG: --config-dir="" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.723376 4885 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.723385 4885 flags.go:64] FLAG: --container-log-max-files="5" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.723396 4885 flags.go:64] FLAG: --container-log-max-size="10Mi" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.723405 4885 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.723414 4885 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.723424 4885 flags.go:64] FLAG: --containerd-namespace="k8s.io" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.723433 4885 flags.go:64] FLAG: --contention-profiling="false" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.723445 4885 flags.go:64] FLAG: --cpu-cfs-quota="true" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.723454 4885 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.723463 4885 flags.go:64] FLAG: --cpu-manager-policy="none" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.723473 4885 flags.go:64] FLAG: --cpu-manager-policy-options="" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.723487 4885 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.723498 4885 flags.go:64] FLAG: --enable-controller-attach-detach="true" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.723509 4885 flags.go:64] FLAG: --enable-debugging-handlers="true" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.723521 4885 flags.go:64] FLAG: --enable-load-reader="false" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.723532 4885 flags.go:64] FLAG: --enable-server="true" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.723543 4885 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.723556 4885 flags.go:64] FLAG: --event-burst="100" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.723568 4885 flags.go:64] FLAG: --event-qps="50" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.723579 4885 flags.go:64] FLAG: --event-storage-age-limit="default=0" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.723591 4885 flags.go:64] FLAG: --event-storage-event-limit="default=0" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.723600 4885 flags.go:64] FLAG: --eviction-hard="" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.723613 4885 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.723623 4885 flags.go:64] FLAG: --eviction-minimum-reclaim="" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.723632 4885 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.723641 4885 flags.go:64] FLAG: --eviction-soft="" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.723649 4885 flags.go:64] FLAG: --eviction-soft-grace-period="" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.723658 4885 flags.go:64] FLAG: --exit-on-lock-contention="false" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.723667 4885 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.723677 4885 flags.go:64] FLAG: --experimental-mounter-path="" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.723686 4885 flags.go:64] FLAG: --fail-cgroupv1="false" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.723696 4885 flags.go:64] FLAG: --fail-swap-on="true" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.723706 4885 flags.go:64] FLAG: --feature-gates="" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.723719 4885 flags.go:64] FLAG: --file-check-frequency="20s" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.723730 4885 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.723741 4885 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.723750 4885 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.723759 4885 flags.go:64] FLAG: --healthz-port="10248" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.723770 4885 flags.go:64] FLAG: --help="false" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.723780 4885 flags.go:64] FLAG: --hostname-override="" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.723789 4885 flags.go:64] FLAG: --housekeeping-interval="10s" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.723798 4885 flags.go:64] FLAG: --http-check-frequency="20s" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.723807 4885 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.723815 4885 flags.go:64] FLAG: --image-credential-provider-config="" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.723824 4885 flags.go:64] FLAG: --image-gc-high-threshold="85" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.723833 4885 flags.go:64] FLAG: --image-gc-low-threshold="80" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.723842 4885 flags.go:64] FLAG: --image-service-endpoint="" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.723850 4885 flags.go:64] FLAG: --kernel-memcg-notification="false" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.723858 4885 flags.go:64] FLAG: --kube-api-burst="100" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.723867 4885 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.723877 4885 flags.go:64] FLAG: --kube-api-qps="50" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.723885 4885 flags.go:64] FLAG: --kube-reserved="" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.723894 4885 flags.go:64] FLAG: --kube-reserved-cgroup="" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.723903 4885 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.723912 4885 flags.go:64] FLAG: --kubelet-cgroups="" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.723921 4885 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.723929 4885 flags.go:64] FLAG: --lock-file="" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.723938 4885 flags.go:64] FLAG: --log-cadvisor-usage="false" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.723948 4885 flags.go:64] FLAG: --log-flush-frequency="5s" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.723956 4885 flags.go:64] FLAG: --log-json-info-buffer-size="0" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.723969 4885 flags.go:64] FLAG: --log-json-split-stream="false" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.723979 4885 flags.go:64] FLAG: --log-text-info-buffer-size="0" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.723988 4885 flags.go:64] FLAG: --log-text-split-stream="false" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.723996 4885 flags.go:64] FLAG: --logging-format="text" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.724035 4885 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.724046 4885 flags.go:64] FLAG: --make-iptables-util-chains="true" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.724056 4885 flags.go:64] FLAG: --manifest-url="" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.724065 4885 flags.go:64] FLAG: --manifest-url-header="" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.724076 4885 flags.go:64] FLAG: --max-housekeeping-interval="15s" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.724086 4885 flags.go:64] FLAG: --max-open-files="1000000" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.724098 4885 flags.go:64] FLAG: --max-pods="110" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.724107 4885 flags.go:64] FLAG: --maximum-dead-containers="-1" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.724116 4885 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.724125 4885 flags.go:64] FLAG: --memory-manager-policy="None" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.724134 4885 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.724142 4885 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.724152 4885 flags.go:64] FLAG: --node-ip="192.168.126.11" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.724161 4885 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.724181 4885 flags.go:64] FLAG: --node-status-max-images="50" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.724190 4885 flags.go:64] FLAG: --node-status-update-frequency="10s" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.724198 4885 flags.go:64] FLAG: --oom-score-adj="-999" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.724207 4885 flags.go:64] FLAG: --pod-cidr="" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.724216 4885 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.724230 4885 flags.go:64] FLAG: --pod-manifest-path="" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.724239 4885 flags.go:64] FLAG: --pod-max-pids="-1" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.724248 4885 flags.go:64] FLAG: --pods-per-core="0" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.724257 4885 flags.go:64] FLAG: --port="10250" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.724294 4885 flags.go:64] FLAG: --protect-kernel-defaults="false" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.724303 4885 flags.go:64] FLAG: --provider-id="" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.724311 4885 flags.go:64] FLAG: --qos-reserved="" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.724320 4885 flags.go:64] FLAG: --read-only-port="10255" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.724330 4885 flags.go:64] FLAG: --register-node="true" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.724339 4885 flags.go:64] FLAG: --register-schedulable="true" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.724348 4885 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.724364 4885 flags.go:64] FLAG: --registry-burst="10" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.724372 4885 flags.go:64] FLAG: --registry-qps="5" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.724381 4885 flags.go:64] FLAG: --reserved-cpus="" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.724390 4885 flags.go:64] FLAG: --reserved-memory="" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.724401 4885 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.724410 4885 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.724419 4885 flags.go:64] FLAG: --rotate-certificates="false" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.724428 4885 flags.go:64] FLAG: --rotate-server-certificates="false" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.724437 4885 flags.go:64] FLAG: --runonce="false" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.724446 4885 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.724456 4885 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.724464 4885 flags.go:64] FLAG: --seccomp-default="false" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.724473 4885 flags.go:64] FLAG: --serialize-image-pulls="true" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.724482 4885 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.724491 4885 flags.go:64] FLAG: --storage-driver-db="cadvisor" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.724500 4885 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.724510 4885 flags.go:64] FLAG: --storage-driver-password="root" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.724519 4885 flags.go:64] FLAG: --storage-driver-secure="false" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.724528 4885 flags.go:64] FLAG: --storage-driver-table="stats" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.724537 4885 flags.go:64] FLAG: --storage-driver-user="root" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.724546 4885 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.724555 4885 flags.go:64] FLAG: --sync-frequency="1m0s" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.724564 4885 flags.go:64] FLAG: --system-cgroups="" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.724573 4885 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.724588 4885 flags.go:64] FLAG: --system-reserved-cgroup="" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.724596 4885 flags.go:64] FLAG: --tls-cert-file="" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.724605 4885 flags.go:64] FLAG: --tls-cipher-suites="[]" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.724618 4885 flags.go:64] FLAG: --tls-min-version="" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.724626 4885 flags.go:64] FLAG: --tls-private-key-file="" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.724636 4885 flags.go:64] FLAG: --topology-manager-policy="none" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.724644 4885 flags.go:64] FLAG: --topology-manager-policy-options="" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.724653 4885 flags.go:64] FLAG: --topology-manager-scope="container" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.724662 4885 flags.go:64] FLAG: --v="2" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.724674 4885 flags.go:64] FLAG: --version="false" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.724685 4885 flags.go:64] FLAG: --vmodule="" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.724695 4885 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.724706 4885 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.724911 4885 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.724923 4885 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.724933 4885 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.724941 4885 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.724950 4885 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.724958 4885 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.724965 4885 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.724973 4885 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.724981 4885 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.724989 4885 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.724996 4885 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.725004 4885 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.725012 4885 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.725020 4885 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.725028 4885 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.725036 4885 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.725043 4885 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.725051 4885 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.725059 4885 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.725067 4885 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.725076 4885 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.725083 4885 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.725091 4885 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.725099 4885 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.725107 4885 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.725117 4885 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.725127 4885 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.725136 4885 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.725146 4885 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.725155 4885 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.725165 4885 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.725173 4885 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.725181 4885 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.725189 4885 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.725197 4885 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.725207 4885 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.725217 4885 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.725228 4885 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.725237 4885 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.725247 4885 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.725255 4885 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.725290 4885 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.725298 4885 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.725306 4885 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.725315 4885 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.725323 4885 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.725331 4885 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.725341 4885 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.725350 4885 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.725369 4885 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.725379 4885 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.725389 4885 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.725398 4885 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.725408 4885 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.725418 4885 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.725428 4885 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.725438 4885 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.725446 4885 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.725454 4885 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.725461 4885 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.725470 4885 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.725477 4885 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.725490 4885 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.725502 4885 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.725512 4885 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.725522 4885 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.725531 4885 feature_gate.go:330] unrecognized feature gate: Example Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.725541 4885 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.725559 4885 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.725568 4885 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.725578 4885 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.725596 4885 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.741221 4885 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.741325 4885 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.741503 4885 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.741537 4885 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.741549 4885 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.741560 4885 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.741571 4885 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.741583 4885 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.741593 4885 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.741603 4885 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.741613 4885 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.741623 4885 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.741633 4885 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.741643 4885 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.741655 4885 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.741665 4885 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.741673 4885 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.741683 4885 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.741692 4885 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.741702 4885 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.741713 4885 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.741723 4885 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.741734 4885 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.741745 4885 feature_gate.go:330] unrecognized feature gate: Example Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.741757 4885 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.741800 4885 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.741812 4885 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.741823 4885 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.741834 4885 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.741864 4885 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.741879 4885 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.741890 4885 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.741900 4885 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.741911 4885 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.741921 4885 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.741932 4885 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.741942 4885 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.741952 4885 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.741963 4885 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.741973 4885 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.741984 4885 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.741993 4885 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.742017 4885 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.742031 4885 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.742040 4885 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.742049 4885 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.742059 4885 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.742070 4885 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.742080 4885 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.742091 4885 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.742103 4885 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.742113 4885 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.742127 4885 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.742141 4885 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.742153 4885 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.742164 4885 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.742175 4885 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.742183 4885 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.742194 4885 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.742202 4885 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.742210 4885 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.742217 4885 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.742228 4885 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.742237 4885 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.742247 4885 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.742284 4885 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.742294 4885 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.742303 4885 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.742311 4885 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.742319 4885 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.742327 4885 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.742337 4885 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.742346 4885 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.742362 4885 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.742622 4885 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.742640 4885 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.742650 4885 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.742659 4885 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.742668 4885 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.742676 4885 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.742685 4885 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.742693 4885 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.742703 4885 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.742711 4885 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.742720 4885 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.742729 4885 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.742737 4885 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.742747 4885 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.742759 4885 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.742768 4885 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.742781 4885 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.742791 4885 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.742801 4885 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.742811 4885 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.742820 4885 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.742829 4885 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.742838 4885 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.742846 4885 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.742891 4885 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.742900 4885 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.742908 4885 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.742916 4885 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.742924 4885 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.742932 4885 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.742940 4885 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.742948 4885 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.742956 4885 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.742964 4885 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.742972 4885 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.742980 4885 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.742988 4885 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.742999 4885 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.743010 4885 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.743020 4885 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.743030 4885 feature_gate.go:330] unrecognized feature gate: Example Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.743040 4885 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.743050 4885 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.743060 4885 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.743070 4885 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.743079 4885 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.743089 4885 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.743099 4885 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.743110 4885 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.743124 4885 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.743137 4885 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.743149 4885 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.743160 4885 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.743171 4885 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.743181 4885 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.743191 4885 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.743201 4885 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.743211 4885 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.743222 4885 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.743235 4885 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.743247 4885 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.743296 4885 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.743308 4885 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.743319 4885 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.743329 4885 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.743339 4885 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.743349 4885 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.743359 4885 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.743370 4885 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.743381 4885 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.743390 4885 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.743406 4885 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.743697 4885 server.go:940] "Client rotation is on, will bootstrap in background" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.749769 4885 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.749913 4885 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.751921 4885 server.go:997] "Starting client certificate rotation" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.751972 4885 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.752372 4885 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-11-10 10:01:40.390363172 +0000 UTC Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.752517 4885 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 944h14m48.637851964s for next certificate rotation Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.829929 4885 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.835038 4885 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.855396 4885 log.go:25] "Validated CRI v1 runtime API" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.894859 4885 log.go:25] "Validated CRI v1 image API" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.897825 4885 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.906356 4885 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-10-02-01-42-23-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.906405 4885 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.935861 4885 manager.go:217] Machine: {Timestamp:2025-10-02 01:46:51.931129129 +0000 UTC m=+0.742876568 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2799886 MemoryCapacity:33654124544 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:50ddf2c2-ea95-4960-ae48-d17e9484ee2f BootID:92603ae5-ba38-4201-98e9-979fd56cb50d Filesystems:[{Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365408768 Type:vfs Inodes:821633 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108169 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827060224 Type:vfs Inodes:4108169 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:b2:fe:b7 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:b2:fe:b7 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:a5:d6:22 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:14:18:3c Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:83:6f:6d Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:1b:df:85 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:26:eb:23:17:e9:30 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:ce:85:0e:9c:05:52 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654124544 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.936376 4885 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.936614 4885 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.939095 4885 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.939530 4885 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.939581 4885 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.939919 4885 topology_manager.go:138] "Creating topology manager with none policy" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.939940 4885 container_manager_linux.go:303] "Creating device plugin manager" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.940863 4885 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.940919 4885 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.941636 4885 state_mem.go:36] "Initialized new in-memory state store" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.941791 4885 server.go:1245] "Using root directory" path="/var/lib/kubelet" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.948820 4885 kubelet.go:418] "Attempting to sync node with API server" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.948864 4885 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.948914 4885 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.948937 4885 kubelet.go:324] "Adding apiserver pod source" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.948957 4885 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.956444 4885 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.39:6443: connect: connection refused Oct 02 01:46:51 crc kubenswrapper[4885]: E1002 01:46:51.956588 4885 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.39:6443: connect: connection refused" logger="UnhandledError" Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.956579 4885 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.39:6443: connect: connection refused Oct 02 01:46:51 crc kubenswrapper[4885]: E1002 01:46:51.956707 4885 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.39:6443: connect: connection refused" logger="UnhandledError" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.957078 4885 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.958513 4885 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.962363 4885 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.964726 4885 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.964827 4885 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.964846 4885 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.964889 4885 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.964911 4885 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.964924 4885 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.964938 4885 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.964989 4885 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.965007 4885 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.965022 4885 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.965119 4885 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.965162 4885 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.967684 4885 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.969327 4885 server.go:1280] "Started kubelet" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.970035 4885 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.39:6443: connect: connection refused Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.971626 4885 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Oct 02 01:46:51 crc systemd[1]: Started Kubernetes Kubelet. Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.971689 4885 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.972545 4885 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.973132 4885 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.973185 4885 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.973235 4885 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-21 14:09:20.835404519 +0000 UTC Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.973380 4885 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 1932h22m28.862029592s for next certificate rotation Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.973729 4885 volume_manager.go:287] "The desired_state_of_world populator starts" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.973764 4885 volume_manager.go:289] "Starting Kubelet Volume Manager" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.973895 4885 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Oct 02 01:46:51 crc kubenswrapper[4885]: E1002 01:46:51.974340 4885 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 02 01:46:51 crc kubenswrapper[4885]: E1002 01:46:51.975204 4885 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.39:6443: connect: connection refused" interval="200ms" Oct 02 01:46:51 crc kubenswrapper[4885]: W1002 01:46:51.975642 4885 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.39:6443: connect: connection refused Oct 02 01:46:51 crc kubenswrapper[4885]: E1002 01:46:51.975793 4885 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.39:6443: connect: connection refused" logger="UnhandledError" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.975851 4885 factory.go:55] Registering systemd factory Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.975879 4885 factory.go:221] Registration of the systemd container factory successfully Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.976297 4885 factory.go:153] Registering CRI-O factory Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.976316 4885 factory.go:221] Registration of the crio container factory successfully Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.976418 4885 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.976450 4885 factory.go:103] Registering Raw factory Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.976475 4885 manager.go:1196] Started watching for new ooms in manager Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.983295 4885 manager.go:319] Starting recovery of all containers Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.985168 4885 server.go:460] "Adding debug handlers to kubelet server" Oct 02 01:46:51 crc kubenswrapper[4885]: E1002 01:46:51.983965 4885 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.39:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.186a89572e005b2c default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-10-02 01:46:51.96853534 +0000 UTC m=+0.780282769,LastTimestamp:2025-10-02 01:46:51.96853534 +0000 UTC m=+0.780282769,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.999029 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.999133 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.999158 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.999181 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.999201 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.999220 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.999240 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.999288 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.999311 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.999330 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.999352 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.999370 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.999429 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.999452 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.999470 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.999489 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.999513 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.999531 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Oct 02 01:46:51 crc kubenswrapper[4885]: I1002 01:46:51.999552 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:51.999570 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:51.999589 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:51.999609 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:51.999656 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:51.999676 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:51.999694 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:51.999713 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:51.999735 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:51.999755 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:51.999775 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:51.999794 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:51.999811 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:51.999830 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:51.999853 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:51.999871 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:51.999892 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:51.999912 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:51.999931 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:51.999950 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:51.999967 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:51.999987 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.000004 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.000023 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.000043 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.000062 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.000080 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.000099 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.000121 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.000140 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.000159 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.000179 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.000200 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.000220 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.000245 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.000298 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.000322 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.000344 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.000373 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.000394 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.000416 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.000435 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.000455 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.000474 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.000492 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.000512 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.000532 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.000552 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.000571 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.000590 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.000609 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.000627 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.000650 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.000669 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.000690 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.000710 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.000733 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.000752 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.000774 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.000794 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.000816 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.000836 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.000855 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.000873 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.000893 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.000912 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.000930 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.000948 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.000969 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.000988 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.001007 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.001025 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.001042 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.001060 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.001077 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.001096 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.001114 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.001132 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.001151 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.001170 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.001189 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.001216 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.001233 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.001252 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.001293 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.001314 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.001354 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.001376 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.001395 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.001413 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.001433 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.001453 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.001472 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.001493 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.001513 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.001533 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.001552 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.001573 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.001590 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.001607 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.001626 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.001646 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.001665 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.001683 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.001703 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.001721 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.001743 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.001764 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.001784 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.001802 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.001820 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.001838 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.001855 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.001873 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.001891 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.001911 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.001928 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.001947 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.001965 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.001983 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.002001 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.002021 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.002040 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.002056 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.002074 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.002092 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.002112 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.002129 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.002146 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.002164 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.002182 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.002199 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.002215 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.002233 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.002251 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.002295 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.002313 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.002332 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.002350 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.002368 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.002386 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.002404 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.002423 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.002442 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.002460 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.002481 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.002501 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.002521 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.004629 4885 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.004671 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.004694 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.004718 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.004740 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.004763 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.004783 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.004803 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.004825 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.004844 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.004869 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.004888 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.004910 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.004929 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.004957 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.004975 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.004993 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.005011 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.005029 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.005048 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.005067 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.005085 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.005105 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.005125 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.005143 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.005161 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.005183 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.005201 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.005223 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.005241 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.005287 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.005308 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.005326 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.005347 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.005364 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.005383 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.005404 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.005424 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.005459 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.005478 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.005497 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.005549 4885 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.005568 4885 reconstruct.go:97] "Volume reconstruction finished" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.005581 4885 reconciler.go:26] "Reconciler: start to sync state" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.019948 4885 manager.go:324] Recovery completed Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.037321 4885 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.040673 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.040736 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.040755 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.041814 4885 cpu_manager.go:225] "Starting CPU manager" policy="none" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.041962 4885 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.042105 4885 state_mem.go:36] "Initialized new in-memory state store" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.042203 4885 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.045132 4885 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.045225 4885 status_manager.go:217] "Starting to sync pod status with apiserver" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.045321 4885 kubelet.go:2335] "Starting kubelet main sync loop" Oct 02 01:46:52 crc kubenswrapper[4885]: E1002 01:46:52.045425 4885 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Oct 02 01:46:52 crc kubenswrapper[4885]: W1002 01:46:52.046512 4885 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.39:6443: connect: connection refused Oct 02 01:46:52 crc kubenswrapper[4885]: E1002 01:46:52.046617 4885 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.39:6443: connect: connection refused" logger="UnhandledError" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.058719 4885 policy_none.go:49] "None policy: Start" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.060389 4885 memory_manager.go:170] "Starting memorymanager" policy="None" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.060435 4885 state_mem.go:35] "Initializing new in-memory state store" Oct 02 01:46:52 crc kubenswrapper[4885]: E1002 01:46:52.074428 4885 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.133073 4885 manager.go:334] "Starting Device Plugin manager" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.133159 4885 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.133181 4885 server.go:79] "Starting device plugin registration server" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.134574 4885 eviction_manager.go:189] "Eviction manager: starting control loop" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.134618 4885 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.134911 4885 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.135163 4885 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.135195 4885 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.145923 4885 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc"] Oct 02 01:46:52 crc kubenswrapper[4885]: E1002 01:46:52.145975 4885 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.146079 4885 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.147729 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.147800 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.147825 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.148069 4885 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.148652 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.148777 4885 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.149976 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.150025 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.150044 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.150126 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.150162 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.150180 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.150210 4885 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.150827 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.150919 4885 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.151351 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.151397 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.151410 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.151546 4885 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.151791 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.151858 4885 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.153590 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.153618 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.153651 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.153628 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.153672 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.153680 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.153698 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.153654 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.153832 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.153890 4885 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.154046 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.154100 4885 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.155104 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.155144 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.155161 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.155218 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.155245 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.155297 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.155479 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.155531 4885 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.156498 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.156533 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.156552 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:46:52 crc kubenswrapper[4885]: E1002 01:46:52.176412 4885 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.39:6443: connect: connection refused" interval="400ms" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.208529 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.208583 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.208621 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.208655 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.208687 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.208719 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.208746 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.208774 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.208819 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.208878 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.208907 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.208938 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.208967 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.208998 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.209035 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.234807 4885 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.236124 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.236200 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.236225 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.236306 4885 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 02 01:46:52 crc kubenswrapper[4885]: E1002 01:46:52.237150 4885 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.39:6443: connect: connection refused" node="crc" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.310771 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.310841 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.310885 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.310919 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.310952 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.310992 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.311003 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.311054 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.311106 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.311155 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.311160 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.311105 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.311019 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.311317 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.311352 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.311386 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.311418 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.311429 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.311436 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.311449 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.311098 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.311479 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.311489 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.311527 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.311863 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.311900 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.311946 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.312054 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.311978 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.312147 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.437580 4885 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.439123 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.439172 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.439190 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.439223 4885 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 02 01:46:52 crc kubenswrapper[4885]: E1002 01:46:52.440012 4885 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.39:6443: connect: connection refused" node="crc" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.496676 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.509989 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.532421 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 01:46:52 crc kubenswrapper[4885]: W1002 01:46:52.550119 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-c587306f043cbd25966cbe3ae1adee582f1b44d5c111bee7b9de00e0f5bb2502 WatchSource:0}: Error finding container c587306f043cbd25966cbe3ae1adee582f1b44d5c111bee7b9de00e0f5bb2502: Status 404 returned error can't find the container with id c587306f043cbd25966cbe3ae1adee582f1b44d5c111bee7b9de00e0f5bb2502 Oct 02 01:46:52 crc kubenswrapper[4885]: W1002 01:46:52.550593 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-ad4ffa999ae83a23fa6ad6753b2cf4ebb6997c63e3f8b4a505ad01e01cb00019 WatchSource:0}: Error finding container ad4ffa999ae83a23fa6ad6753b2cf4ebb6997c63e3f8b4a505ad01e01cb00019: Status 404 returned error can't find the container with id ad4ffa999ae83a23fa6ad6753b2cf4ebb6997c63e3f8b4a505ad01e01cb00019 Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.554678 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 02 01:46:52 crc kubenswrapper[4885]: W1002 01:46:52.560770 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-b4b3eb2234ef7337699f0655275b84328330324c5f0a3be8b36a49cc2efe7b3f WatchSource:0}: Error finding container b4b3eb2234ef7337699f0655275b84328330324c5f0a3be8b36a49cc2efe7b3f: Status 404 returned error can't find the container with id b4b3eb2234ef7337699f0655275b84328330324c5f0a3be8b36a49cc2efe7b3f Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.562603 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 02 01:46:52 crc kubenswrapper[4885]: E1002 01:46:52.577429 4885 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.39:6443: connect: connection refused" interval="800ms" Oct 02 01:46:52 crc kubenswrapper[4885]: W1002 01:46:52.578180 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-74259bb019f874dc709833ec99ba8b833fdc80c1cb4476e2f812935afd64f599 WatchSource:0}: Error finding container 74259bb019f874dc709833ec99ba8b833fdc80c1cb4476e2f812935afd64f599: Status 404 returned error can't find the container with id 74259bb019f874dc709833ec99ba8b833fdc80c1cb4476e2f812935afd64f599 Oct 02 01:46:52 crc kubenswrapper[4885]: W1002 01:46:52.585953 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-f6a991e3da92e69d2a7dd2735c8325d7c9c4618c8ecf906bfe98eebf1b4fcb05 WatchSource:0}: Error finding container f6a991e3da92e69d2a7dd2735c8325d7c9c4618c8ecf906bfe98eebf1b4fcb05: Status 404 returned error can't find the container with id f6a991e3da92e69d2a7dd2735c8325d7c9c4618c8ecf906bfe98eebf1b4fcb05 Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.840465 4885 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.842483 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.842546 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.842568 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.842657 4885 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 02 01:46:52 crc kubenswrapper[4885]: E1002 01:46:52.843367 4885 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.39:6443: connect: connection refused" node="crc" Oct 02 01:46:52 crc kubenswrapper[4885]: W1002 01:46:52.888454 4885 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.39:6443: connect: connection refused Oct 02 01:46:52 crc kubenswrapper[4885]: E1002 01:46:52.888569 4885 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.39:6443: connect: connection refused" logger="UnhandledError" Oct 02 01:46:52 crc kubenswrapper[4885]: I1002 01:46:52.971290 4885 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.39:6443: connect: connection refused Oct 02 01:46:53 crc kubenswrapper[4885]: I1002 01:46:53.049890 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"c587306f043cbd25966cbe3ae1adee582f1b44d5c111bee7b9de00e0f5bb2502"} Oct 02 01:46:53 crc kubenswrapper[4885]: I1002 01:46:53.051321 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"ad4ffa999ae83a23fa6ad6753b2cf4ebb6997c63e3f8b4a505ad01e01cb00019"} Oct 02 01:46:53 crc kubenswrapper[4885]: I1002 01:46:53.052776 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"f6a991e3da92e69d2a7dd2735c8325d7c9c4618c8ecf906bfe98eebf1b4fcb05"} Oct 02 01:46:53 crc kubenswrapper[4885]: I1002 01:46:53.054211 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"74259bb019f874dc709833ec99ba8b833fdc80c1cb4476e2f812935afd64f599"} Oct 02 01:46:53 crc kubenswrapper[4885]: I1002 01:46:53.057131 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"b4b3eb2234ef7337699f0655275b84328330324c5f0a3be8b36a49cc2efe7b3f"} Oct 02 01:46:53 crc kubenswrapper[4885]: W1002 01:46:53.068315 4885 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.39:6443: connect: connection refused Oct 02 01:46:53 crc kubenswrapper[4885]: E1002 01:46:53.068417 4885 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.39:6443: connect: connection refused" logger="UnhandledError" Oct 02 01:46:53 crc kubenswrapper[4885]: W1002 01:46:53.323175 4885 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.39:6443: connect: connection refused Oct 02 01:46:53 crc kubenswrapper[4885]: E1002 01:46:53.323816 4885 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.39:6443: connect: connection refused" logger="UnhandledError" Oct 02 01:46:53 crc kubenswrapper[4885]: E1002 01:46:53.378445 4885 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.39:6443: connect: connection refused" interval="1.6s" Oct 02 01:46:53 crc kubenswrapper[4885]: W1002 01:46:53.555424 4885 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.39:6443: connect: connection refused Oct 02 01:46:53 crc kubenswrapper[4885]: E1002 01:46:53.555561 4885 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.39:6443: connect: connection refused" logger="UnhandledError" Oct 02 01:46:53 crc kubenswrapper[4885]: I1002 01:46:53.643672 4885 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 01:46:53 crc kubenswrapper[4885]: I1002 01:46:53.645785 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:46:53 crc kubenswrapper[4885]: I1002 01:46:53.645859 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:46:53 crc kubenswrapper[4885]: I1002 01:46:53.645887 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:46:53 crc kubenswrapper[4885]: I1002 01:46:53.645941 4885 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 02 01:46:53 crc kubenswrapper[4885]: E1002 01:46:53.646718 4885 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.39:6443: connect: connection refused" node="crc" Oct 02 01:46:53 crc kubenswrapper[4885]: I1002 01:46:53.971849 4885 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.39:6443: connect: connection refused Oct 02 01:46:54 crc kubenswrapper[4885]: I1002 01:46:54.065514 4885 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="25c85ae265d3c8450cd878ca9e4452c40df15c44493535be916b7500f1c5bac0" exitCode=0 Oct 02 01:46:54 crc kubenswrapper[4885]: I1002 01:46:54.065666 4885 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 01:46:54 crc kubenswrapper[4885]: I1002 01:46:54.065684 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"25c85ae265d3c8450cd878ca9e4452c40df15c44493535be916b7500f1c5bac0"} Oct 02 01:46:54 crc kubenswrapper[4885]: I1002 01:46:54.068517 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:46:54 crc kubenswrapper[4885]: I1002 01:46:54.068580 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:46:54 crc kubenswrapper[4885]: I1002 01:46:54.068609 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:46:54 crc kubenswrapper[4885]: I1002 01:46:54.070524 4885 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="7b4cd17b64e5657ae05e71f3ca2340cb0d7965b1b49d1f7b4bfe4494292293ca" exitCode=0 Oct 02 01:46:54 crc kubenswrapper[4885]: I1002 01:46:54.070600 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"7b4cd17b64e5657ae05e71f3ca2340cb0d7965b1b49d1f7b4bfe4494292293ca"} Oct 02 01:46:54 crc kubenswrapper[4885]: I1002 01:46:54.070652 4885 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 01:46:54 crc kubenswrapper[4885]: I1002 01:46:54.071930 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:46:54 crc kubenswrapper[4885]: I1002 01:46:54.071982 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:46:54 crc kubenswrapper[4885]: I1002 01:46:54.072002 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:46:54 crc kubenswrapper[4885]: I1002 01:46:54.076322 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"e7c00db833a593ec40eb63a6b383455e19e960c2e98b9b66f76877ee993057e7"} Oct 02 01:46:54 crc kubenswrapper[4885]: I1002 01:46:54.076375 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"31c2f0193ffef54965b79a3569b0f6f4547e540bac0774155159a1c996c196c9"} Oct 02 01:46:54 crc kubenswrapper[4885]: I1002 01:46:54.076394 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"533e73e5f90b7258acbcdf018ee01323c118f9dc45035b25b065ba01cabecccf"} Oct 02 01:46:54 crc kubenswrapper[4885]: I1002 01:46:54.079518 4885 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="83acbc20919f7c1958b6d492fadb203855df02761662425df9f91a0757fe2bf7" exitCode=0 Oct 02 01:46:54 crc kubenswrapper[4885]: I1002 01:46:54.079631 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"83acbc20919f7c1958b6d492fadb203855df02761662425df9f91a0757fe2bf7"} Oct 02 01:46:54 crc kubenswrapper[4885]: I1002 01:46:54.079708 4885 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 01:46:54 crc kubenswrapper[4885]: I1002 01:46:54.081017 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:46:54 crc kubenswrapper[4885]: I1002 01:46:54.081075 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:46:54 crc kubenswrapper[4885]: I1002 01:46:54.081100 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:46:54 crc kubenswrapper[4885]: I1002 01:46:54.083400 4885 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="592f89a326c9c1559193a71904b47ca27b1dbe1a7b30e03bc48924bb2f35f921" exitCode=0 Oct 02 01:46:54 crc kubenswrapper[4885]: I1002 01:46:54.083459 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"592f89a326c9c1559193a71904b47ca27b1dbe1a7b30e03bc48924bb2f35f921"} Oct 02 01:46:54 crc kubenswrapper[4885]: I1002 01:46:54.083628 4885 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 01:46:54 crc kubenswrapper[4885]: I1002 01:46:54.085417 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:46:54 crc kubenswrapper[4885]: I1002 01:46:54.085474 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:46:54 crc kubenswrapper[4885]: I1002 01:46:54.085491 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:46:54 crc kubenswrapper[4885]: I1002 01:46:54.088026 4885 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 01:46:54 crc kubenswrapper[4885]: I1002 01:46:54.089421 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:46:54 crc kubenswrapper[4885]: I1002 01:46:54.089470 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:46:54 crc kubenswrapper[4885]: I1002 01:46:54.089489 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:46:54 crc kubenswrapper[4885]: W1002 01:46:54.848738 4885 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.39:6443: connect: connection refused Oct 02 01:46:54 crc kubenswrapper[4885]: E1002 01:46:54.848847 4885 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.39:6443: connect: connection refused" logger="UnhandledError" Oct 02 01:46:54 crc kubenswrapper[4885]: I1002 01:46:54.971969 4885 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.39:6443: connect: connection refused Oct 02 01:46:54 crc kubenswrapper[4885]: E1002 01:46:54.979874 4885 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.39:6443: connect: connection refused" interval="3.2s" Oct 02 01:46:55 crc kubenswrapper[4885]: I1002 01:46:55.088053 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"614992bc938987be5169d990ded500acb5120eb0ddfc221168faa0d8c16e0972"} Oct 02 01:46:55 crc kubenswrapper[4885]: I1002 01:46:55.088117 4885 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 01:46:55 crc kubenswrapper[4885]: I1002 01:46:55.089778 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:46:55 crc kubenswrapper[4885]: I1002 01:46:55.089819 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:46:55 crc kubenswrapper[4885]: I1002 01:46:55.089828 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:46:55 crc kubenswrapper[4885]: I1002 01:46:55.096464 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"620f54e02c463f6e603808e5e57e9e5770d125578cc258ba16117a2416e57d3c"} Oct 02 01:46:55 crc kubenswrapper[4885]: I1002 01:46:55.096520 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"298c2ace58c4a9d09cfdc864316317e370bb0ac5735d631af6d85c6298f686d8"} Oct 02 01:46:55 crc kubenswrapper[4885]: I1002 01:46:55.096543 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"b742e4b288bb50569d2fd23f3f9857d649a46063998170eee582ecbbba64d190"} Oct 02 01:46:55 crc kubenswrapper[4885]: I1002 01:46:55.096632 4885 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 01:46:55 crc kubenswrapper[4885]: I1002 01:46:55.097979 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:46:55 crc kubenswrapper[4885]: I1002 01:46:55.098009 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:46:55 crc kubenswrapper[4885]: I1002 01:46:55.098018 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:46:55 crc kubenswrapper[4885]: I1002 01:46:55.101338 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"21e2f462a0deb0d740185663b2014f4a1e57b4fdbeb82571264b002f928035bb"} Oct 02 01:46:55 crc kubenswrapper[4885]: I1002 01:46:55.101486 4885 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 01:46:55 crc kubenswrapper[4885]: I1002 01:46:55.103896 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:46:55 crc kubenswrapper[4885]: I1002 01:46:55.103918 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:46:55 crc kubenswrapper[4885]: I1002 01:46:55.103927 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:46:55 crc kubenswrapper[4885]: I1002 01:46:55.103925 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"8e86ecd03835efc9a253b6804bc61e6ac690e75b8d1f3cc77f5647730e586574"} Oct 02 01:46:55 crc kubenswrapper[4885]: I1002 01:46:55.103843 4885 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="8e86ecd03835efc9a253b6804bc61e6ac690e75b8d1f3cc77f5647730e586574" exitCode=0 Oct 02 01:46:55 crc kubenswrapper[4885]: I1002 01:46:55.104095 4885 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 01:46:55 crc kubenswrapper[4885]: I1002 01:46:55.105344 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:46:55 crc kubenswrapper[4885]: I1002 01:46:55.105362 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:46:55 crc kubenswrapper[4885]: I1002 01:46:55.105372 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:46:55 crc kubenswrapper[4885]: I1002 01:46:55.110125 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"90535f97ab2ed4626d47540ebd7baada837fd332140323bec0f49aa38d3c0009"} Oct 02 01:46:55 crc kubenswrapper[4885]: I1002 01:46:55.110200 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"d06c5942805a498d9621a906088aa54538b4eed583d1b8d0e3ba2a6718378a36"} Oct 02 01:46:55 crc kubenswrapper[4885]: I1002 01:46:55.110254 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"06b9f6f7fe44d5771a838d0b89c5bdea3ebb79023ae2675ee907ec2766b3d35a"} Oct 02 01:46:55 crc kubenswrapper[4885]: I1002 01:46:55.110334 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"a89b49cbe08d927629564e607de13655ddf3d01ac4eb5d688b325210306c120c"} Oct 02 01:46:55 crc kubenswrapper[4885]: I1002 01:46:55.227581 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 01:46:55 crc kubenswrapper[4885]: I1002 01:46:55.247269 4885 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 01:46:55 crc kubenswrapper[4885]: I1002 01:46:55.249079 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:46:55 crc kubenswrapper[4885]: I1002 01:46:55.249139 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:46:55 crc kubenswrapper[4885]: I1002 01:46:55.249151 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:46:55 crc kubenswrapper[4885]: I1002 01:46:55.249190 4885 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 02 01:46:55 crc kubenswrapper[4885]: E1002 01:46:55.250026 4885 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.39:6443: connect: connection refused" node="crc" Oct 02 01:46:56 crc kubenswrapper[4885]: I1002 01:46:56.115786 4885 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="48f95af51feaf014a18496bdfadd59d7f1e11167c5c8b677fa7037e48368ff2a" exitCode=0 Oct 02 01:46:56 crc kubenswrapper[4885]: I1002 01:46:56.115885 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"48f95af51feaf014a18496bdfadd59d7f1e11167c5c8b677fa7037e48368ff2a"} Oct 02 01:46:56 crc kubenswrapper[4885]: I1002 01:46:56.116085 4885 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 01:46:56 crc kubenswrapper[4885]: I1002 01:46:56.117628 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:46:56 crc kubenswrapper[4885]: I1002 01:46:56.117677 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:46:56 crc kubenswrapper[4885]: I1002 01:46:56.117695 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:46:56 crc kubenswrapper[4885]: I1002 01:46:56.124553 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"4bb63bf30d6ecce9e71f0db0c16632bdfa6b33924b673378eb0c38ab229fa837"} Oct 02 01:46:56 crc kubenswrapper[4885]: I1002 01:46:56.124716 4885 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 01:46:56 crc kubenswrapper[4885]: I1002 01:46:56.124945 4885 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 01:46:56 crc kubenswrapper[4885]: I1002 01:46:56.125442 4885 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 01:46:56 crc kubenswrapper[4885]: I1002 01:46:56.126057 4885 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 01:46:56 crc kubenswrapper[4885]: I1002 01:46:56.126063 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 02 01:46:56 crc kubenswrapper[4885]: I1002 01:46:56.127305 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:46:56 crc kubenswrapper[4885]: I1002 01:46:56.127451 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:46:56 crc kubenswrapper[4885]: I1002 01:46:56.127490 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:46:56 crc kubenswrapper[4885]: I1002 01:46:56.127541 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:46:56 crc kubenswrapper[4885]: I1002 01:46:56.127490 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:46:56 crc kubenswrapper[4885]: I1002 01:46:56.127613 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:46:56 crc kubenswrapper[4885]: I1002 01:46:56.127623 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:46:56 crc kubenswrapper[4885]: I1002 01:46:56.127344 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:46:56 crc kubenswrapper[4885]: I1002 01:46:56.127686 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:46:56 crc kubenswrapper[4885]: I1002 01:46:56.127704 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:46:56 crc kubenswrapper[4885]: I1002 01:46:56.128019 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:46:56 crc kubenswrapper[4885]: I1002 01:46:56.128184 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:46:57 crc kubenswrapper[4885]: I1002 01:46:57.135173 4885 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 01:46:57 crc kubenswrapper[4885]: I1002 01:46:57.135422 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"4adcc537c23697a324d82af94c31059dd444245efdf7c978db58ede256e112ed"} Oct 02 01:46:57 crc kubenswrapper[4885]: I1002 01:46:57.135755 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"e3e634b3727c517cc48b17c2cfb38f26460f53d1bf25d631d08ec9acb8d780e7"} Oct 02 01:46:57 crc kubenswrapper[4885]: I1002 01:46:57.135540 4885 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 01:46:57 crc kubenswrapper[4885]: I1002 01:46:57.135821 4885 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 02 01:46:57 crc kubenswrapper[4885]: I1002 01:46:57.135782 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"1d2808bc658ea80d68469ed16422d56ef19447f4c90b4af7b2597c64406f8a44"} Oct 02 01:46:57 crc kubenswrapper[4885]: I1002 01:46:57.135956 4885 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 01:46:57 crc kubenswrapper[4885]: I1002 01:46:57.137144 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:46:57 crc kubenswrapper[4885]: I1002 01:46:57.137192 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:46:57 crc kubenswrapper[4885]: I1002 01:46:57.137211 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:46:57 crc kubenswrapper[4885]: I1002 01:46:57.137152 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:46:57 crc kubenswrapper[4885]: I1002 01:46:57.137336 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:46:57 crc kubenswrapper[4885]: I1002 01:46:57.137373 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:46:57 crc kubenswrapper[4885]: I1002 01:46:57.138193 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:46:57 crc kubenswrapper[4885]: I1002 01:46:57.138239 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:46:57 crc kubenswrapper[4885]: I1002 01:46:57.138262 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:46:57 crc kubenswrapper[4885]: I1002 01:46:57.942337 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 01:46:58 crc kubenswrapper[4885]: I1002 01:46:58.144442 4885 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 01:46:58 crc kubenswrapper[4885]: I1002 01:46:58.144526 4885 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 01:46:58 crc kubenswrapper[4885]: I1002 01:46:58.144543 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"97cb59adc2ddfa365ebc9dec649600dc90466ffe59134a6d15a9d3134a7c9b81"} Oct 02 01:46:58 crc kubenswrapper[4885]: I1002 01:46:58.144613 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"edcd1d99644a778dfb6d0a17d2a4abd906400fe8209e611ebf8d7a9abe7ccf28"} Oct 02 01:46:58 crc kubenswrapper[4885]: I1002 01:46:58.148574 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:46:58 crc kubenswrapper[4885]: I1002 01:46:58.148636 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:46:58 crc kubenswrapper[4885]: I1002 01:46:58.148783 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:46:58 crc kubenswrapper[4885]: I1002 01:46:58.149635 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:46:58 crc kubenswrapper[4885]: I1002 01:46:58.149689 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:46:58 crc kubenswrapper[4885]: I1002 01:46:58.149710 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:46:58 crc kubenswrapper[4885]: I1002 01:46:58.245289 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 01:46:58 crc kubenswrapper[4885]: I1002 01:46:58.245528 4885 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 02 01:46:58 crc kubenswrapper[4885]: I1002 01:46:58.245578 4885 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 01:46:58 crc kubenswrapper[4885]: I1002 01:46:58.247456 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:46:58 crc kubenswrapper[4885]: I1002 01:46:58.247506 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:46:58 crc kubenswrapper[4885]: I1002 01:46:58.247523 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:46:58 crc kubenswrapper[4885]: I1002 01:46:58.418445 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 01:46:58 crc kubenswrapper[4885]: I1002 01:46:58.450725 4885 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 01:46:58 crc kubenswrapper[4885]: I1002 01:46:58.452981 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:46:58 crc kubenswrapper[4885]: I1002 01:46:58.453040 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:46:58 crc kubenswrapper[4885]: I1002 01:46:58.453059 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:46:58 crc kubenswrapper[4885]: I1002 01:46:58.453096 4885 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 02 01:46:58 crc kubenswrapper[4885]: I1002 01:46:58.867910 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Oct 02 01:46:59 crc kubenswrapper[4885]: I1002 01:46:59.003447 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 01:46:59 crc kubenswrapper[4885]: I1002 01:46:59.147583 4885 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 01:46:59 crc kubenswrapper[4885]: I1002 01:46:59.147728 4885 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 01:46:59 crc kubenswrapper[4885]: I1002 01:46:59.148925 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:46:59 crc kubenswrapper[4885]: I1002 01:46:59.148976 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:46:59 crc kubenswrapper[4885]: I1002 01:46:59.148994 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:46:59 crc kubenswrapper[4885]: I1002 01:46:59.149227 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:46:59 crc kubenswrapper[4885]: I1002 01:46:59.149304 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:46:59 crc kubenswrapper[4885]: I1002 01:46:59.149321 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:00 crc kubenswrapper[4885]: I1002 01:47:00.151140 4885 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 01:47:00 crc kubenswrapper[4885]: I1002 01:47:00.151217 4885 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 01:47:00 crc kubenswrapper[4885]: I1002 01:47:00.152916 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:00 crc kubenswrapper[4885]: I1002 01:47:00.152962 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:00 crc kubenswrapper[4885]: I1002 01:47:00.152974 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:00 crc kubenswrapper[4885]: I1002 01:47:00.153010 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:00 crc kubenswrapper[4885]: I1002 01:47:00.153027 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:00 crc kubenswrapper[4885]: I1002 01:47:00.152980 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:00 crc kubenswrapper[4885]: I1002 01:47:00.701648 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 01:47:00 crc kubenswrapper[4885]: I1002 01:47:00.701852 4885 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 01:47:00 crc kubenswrapper[4885]: I1002 01:47:00.703450 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:00 crc kubenswrapper[4885]: I1002 01:47:00.703501 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:00 crc kubenswrapper[4885]: I1002 01:47:00.703521 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:00 crc kubenswrapper[4885]: I1002 01:47:00.708744 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 01:47:00 crc kubenswrapper[4885]: I1002 01:47:00.942478 4885 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 02 01:47:00 crc kubenswrapper[4885]: I1002 01:47:00.942563 4885 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 02 01:47:01 crc kubenswrapper[4885]: I1002 01:47:01.153637 4885 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 01:47:01 crc kubenswrapper[4885]: I1002 01:47:01.155145 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:01 crc kubenswrapper[4885]: I1002 01:47:01.155207 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:01 crc kubenswrapper[4885]: I1002 01:47:01.155226 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:02 crc kubenswrapper[4885]: E1002 01:47:02.146163 4885 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 02 01:47:02 crc kubenswrapper[4885]: I1002 01:47:02.147390 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 01:47:02 crc kubenswrapper[4885]: I1002 01:47:02.155304 4885 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 01:47:02 crc kubenswrapper[4885]: I1002 01:47:02.156933 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:02 crc kubenswrapper[4885]: I1002 01:47:02.156991 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:02 crc kubenswrapper[4885]: I1002 01:47:02.157009 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:04 crc kubenswrapper[4885]: I1002 01:47:04.202877 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Oct 02 01:47:04 crc kubenswrapper[4885]: I1002 01:47:04.203083 4885 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 01:47:04 crc kubenswrapper[4885]: I1002 01:47:04.204984 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:04 crc kubenswrapper[4885]: I1002 01:47:04.205034 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:04 crc kubenswrapper[4885]: I1002 01:47:04.205051 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:05 crc kubenswrapper[4885]: W1002 01:47:05.724529 4885 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout Oct 02 01:47:05 crc kubenswrapper[4885]: I1002 01:47:05.724678 4885 trace.go:236] Trace[959882458]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (02-Oct-2025 01:46:55.723) (total time: 10001ms): Oct 02 01:47:05 crc kubenswrapper[4885]: Trace[959882458]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (01:47:05.724) Oct 02 01:47:05 crc kubenswrapper[4885]: Trace[959882458]: [10.001438643s] [10.001438643s] END Oct 02 01:47:05 crc kubenswrapper[4885]: E1002 01:47:05.724712 4885 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Oct 02 01:47:05 crc kubenswrapper[4885]: W1002 01:47:05.828604 4885 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout Oct 02 01:47:05 crc kubenswrapper[4885]: I1002 01:47:05.828741 4885 trace.go:236] Trace[2071088839]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (02-Oct-2025 01:46:55.827) (total time: 10001ms): Oct 02 01:47:05 crc kubenswrapper[4885]: Trace[2071088839]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (01:47:05.828) Oct 02 01:47:05 crc kubenswrapper[4885]: Trace[2071088839]: [10.001206448s] [10.001206448s] END Oct 02 01:47:05 crc kubenswrapper[4885]: E1002 01:47:05.828774 4885 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Oct 02 01:47:05 crc kubenswrapper[4885]: I1002 01:47:05.976523 4885 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Oct 02 01:47:06 crc kubenswrapper[4885]: W1002 01:47:06.059057 4885 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout Oct 02 01:47:06 crc kubenswrapper[4885]: I1002 01:47:06.059192 4885 trace.go:236] Trace[491410208]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (02-Oct-2025 01:46:56.057) (total time: 10001ms): Oct 02 01:47:06 crc kubenswrapper[4885]: Trace[491410208]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (01:47:06.059) Oct 02 01:47:06 crc kubenswrapper[4885]: Trace[491410208]: [10.001443154s] [10.001443154s] END Oct 02 01:47:06 crc kubenswrapper[4885]: E1002 01:47:06.059225 4885 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Oct 02 01:47:06 crc kubenswrapper[4885]: I1002 01:47:06.658229 4885 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Oct 02 01:47:06 crc kubenswrapper[4885]: I1002 01:47:06.658727 4885 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 02 01:47:06 crc kubenswrapper[4885]: I1002 01:47:06.665551 4885 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Oct 02 01:47:06 crc kubenswrapper[4885]: I1002 01:47:06.665645 4885 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 02 01:47:07 crc kubenswrapper[4885]: I1002 01:47:07.171035 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 02 01:47:07 crc kubenswrapper[4885]: I1002 01:47:07.173618 4885 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="4bb63bf30d6ecce9e71f0db0c16632bdfa6b33924b673378eb0c38ab229fa837" exitCode=255 Oct 02 01:47:07 crc kubenswrapper[4885]: I1002 01:47:07.173654 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"4bb63bf30d6ecce9e71f0db0c16632bdfa6b33924b673378eb0c38ab229fa837"} Oct 02 01:47:07 crc kubenswrapper[4885]: I1002 01:47:07.173771 4885 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 01:47:07 crc kubenswrapper[4885]: I1002 01:47:07.174742 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:07 crc kubenswrapper[4885]: I1002 01:47:07.174800 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:07 crc kubenswrapper[4885]: I1002 01:47:07.174848 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:07 crc kubenswrapper[4885]: I1002 01:47:07.176130 4885 scope.go:117] "RemoveContainer" containerID="4bb63bf30d6ecce9e71f0db0c16632bdfa6b33924b673378eb0c38ab229fa837" Oct 02 01:47:08 crc kubenswrapper[4885]: I1002 01:47:08.180992 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 02 01:47:08 crc kubenswrapper[4885]: I1002 01:47:08.183600 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"178b4340ee6904c60c814b99d83ee225161aeb23e438d8d02f856579a13ec4e6"} Oct 02 01:47:08 crc kubenswrapper[4885]: I1002 01:47:08.183801 4885 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 01:47:08 crc kubenswrapper[4885]: I1002 01:47:08.184874 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:08 crc kubenswrapper[4885]: I1002 01:47:08.184910 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:08 crc kubenswrapper[4885]: I1002 01:47:08.184923 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:08 crc kubenswrapper[4885]: I1002 01:47:08.418501 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 01:47:09 crc kubenswrapper[4885]: I1002 01:47:09.012471 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 01:47:09 crc kubenswrapper[4885]: I1002 01:47:09.187112 4885 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 01:47:09 crc kubenswrapper[4885]: I1002 01:47:09.188701 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:09 crc kubenswrapper[4885]: I1002 01:47:09.188792 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:09 crc kubenswrapper[4885]: I1002 01:47:09.188824 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:09 crc kubenswrapper[4885]: I1002 01:47:09.194246 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 01:47:10 crc kubenswrapper[4885]: I1002 01:47:10.189579 4885 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 01:47:10 crc kubenswrapper[4885]: I1002 01:47:10.190832 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:10 crc kubenswrapper[4885]: I1002 01:47:10.190898 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:10 crc kubenswrapper[4885]: I1002 01:47:10.190916 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:10 crc kubenswrapper[4885]: I1002 01:47:10.765858 4885 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Oct 02 01:47:10 crc kubenswrapper[4885]: I1002 01:47:10.943503 4885 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 02 01:47:10 crc kubenswrapper[4885]: I1002 01:47:10.943628 4885 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 02 01:47:11 crc kubenswrapper[4885]: I1002 01:47:11.026731 4885 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Oct 02 01:47:11 crc kubenswrapper[4885]: I1002 01:47:11.028208 4885 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Oct 02 01:47:11 crc kubenswrapper[4885]: E1002 01:47:11.657666 4885 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Oct 02 01:47:11 crc kubenswrapper[4885]: I1002 01:47:11.659525 4885 trace.go:236] Trace[1963336893]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (02-Oct-2025 01:46:59.379) (total time: 12280ms): Oct 02 01:47:11 crc kubenswrapper[4885]: Trace[1963336893]: ---"Objects listed" error: 12280ms (01:47:11.659) Oct 02 01:47:11 crc kubenswrapper[4885]: Trace[1963336893]: [12.280107661s] [12.280107661s] END Oct 02 01:47:11 crc kubenswrapper[4885]: I1002 01:47:11.659582 4885 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Oct 02 01:47:11 crc kubenswrapper[4885]: I1002 01:47:11.662806 4885 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Oct 02 01:47:11 crc kubenswrapper[4885]: E1002 01:47:11.666506 4885 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Oct 02 01:47:11 crc kubenswrapper[4885]: I1002 01:47:11.967830 4885 apiserver.go:52] "Watching apiserver" Oct 02 01:47:11 crc kubenswrapper[4885]: I1002 01:47:11.971668 4885 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Oct 02 01:47:11 crc kubenswrapper[4885]: I1002 01:47:11.972147 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-kube-apiserver/kube-apiserver-crc","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h"] Oct 02 01:47:11 crc kubenswrapper[4885]: I1002 01:47:11.972640 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 02 01:47:11 crc kubenswrapper[4885]: I1002 01:47:11.972755 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:47:11 crc kubenswrapper[4885]: E1002 01:47:11.972821 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:47:11 crc kubenswrapper[4885]: I1002 01:47:11.972841 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:47:11 crc kubenswrapper[4885]: E1002 01:47:11.972914 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:47:11 crc kubenswrapper[4885]: I1002 01:47:11.972989 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 02 01:47:11 crc kubenswrapper[4885]: I1002 01:47:11.973033 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 02 01:47:11 crc kubenswrapper[4885]: I1002 01:47:11.973164 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:47:11 crc kubenswrapper[4885]: E1002 01:47:11.973202 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:47:11 crc kubenswrapper[4885]: I1002 01:47:11.974552 4885 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Oct 02 01:47:11 crc kubenswrapper[4885]: I1002 01:47:11.974978 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Oct 02 01:47:11 crc kubenswrapper[4885]: I1002 01:47:11.975170 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Oct 02 01:47:11 crc kubenswrapper[4885]: I1002 01:47:11.975386 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Oct 02 01:47:11 crc kubenswrapper[4885]: I1002 01:47:11.975706 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Oct 02 01:47:11 crc kubenswrapper[4885]: I1002 01:47:11.975716 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Oct 02 01:47:11 crc kubenswrapper[4885]: I1002 01:47:11.976055 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Oct 02 01:47:11 crc kubenswrapper[4885]: I1002 01:47:11.976237 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Oct 02 01:47:11 crc kubenswrapper[4885]: I1002 01:47:11.977145 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Oct 02 01:47:11 crc kubenswrapper[4885]: I1002 01:47:11.979188 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.003281 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.021725 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.034807 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.045442 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.058228 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.069459 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.069546 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.069599 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.069646 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.069700 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.069758 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.069803 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.069855 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.069904 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.069960 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.070013 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.070064 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.070115 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.070168 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.070219 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.070317 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.070375 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.070432 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.070484 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.071313 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.071398 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.071462 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.071516 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.071316 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.071540 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.071570 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.071644 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.071737 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.071562 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.071522 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.071577 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.071367 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.071679 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.071779 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: E1002 01:47:12.071818 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:47:12.571734195 +0000 UTC m=+21.383481644 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.072416 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.072010 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.072033 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.072283 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.072311 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.072428 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.072504 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.072526 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.072548 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.072569 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.072592 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.072612 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.072635 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.072657 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.072677 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.072700 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.072721 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.072742 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.072764 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.072789 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.072813 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.072837 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.072858 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.072879 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.072900 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.072921 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.073401 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.073423 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.073444 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.073465 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.073487 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.073507 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.073527 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.073508 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.073547 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.073622 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.073648 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.073674 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.073700 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.073805 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.073814 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.073829 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.073863 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.073888 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.073910 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.073933 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.073956 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.073981 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.074005 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.074029 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.074050 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.074096 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.074278 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.074401 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.074587 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.074614 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.075493 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.075528 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.075568 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.075590 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.075613 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.075634 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.075670 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.075693 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.075716 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.075739 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.075762 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.075783 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.075803 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.075826 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.075853 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.075875 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.075900 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.076337 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.076393 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.076418 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.076441 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.076468 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.076492 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.076526 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.076549 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.076571 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.076596 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.076619 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.076641 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.076665 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.076687 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.076707 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.076728 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.076749 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.076773 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.076796 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.076817 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.076844 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.076865 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.076886 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.077048 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.077076 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.077099 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.077121 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.077145 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.077165 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.077187 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.077208 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.077229 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.077251 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.077292 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.077316 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.077337 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.077360 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.077381 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.077402 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.077425 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.077446 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.077469 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.077491 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.077513 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.077535 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.077558 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.077579 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.077600 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.077624 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.077648 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.077670 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.077692 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.077722 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.077744 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.077766 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.077789 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.077862 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.077889 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.077912 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.077936 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.077959 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.077983 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.078004 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.078029 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.078052 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.078080 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.078103 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.078126 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.078149 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.078172 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.078193 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.078216 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.078239 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.078276 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.078300 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.078323 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.078454 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.078482 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.078504 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.078791 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.078818 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.078845 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.078867 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.078890 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.078932 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.078957 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.078980 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.079002 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.079024 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.079046 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.079069 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.079089 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.079112 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.079135 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.079160 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.079186 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.079209 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.079233 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.079276 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.079300 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.079323 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.079346 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.079368 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.079390 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.079412 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.079435 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.079476 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.079502 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.079528 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.079555 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.079575 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.079601 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.079629 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.079657 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.079702 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.079725 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.079752 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.079778 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.079800 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.079824 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.080898 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.080945 4885 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.082833 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.082860 4885 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.082879 4885 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.082899 4885 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.082918 4885 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.082934 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.082949 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.083012 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.083029 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.083044 4885 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.083059 4885 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.083109 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.083124 4885 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.083139 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.083410 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.083454 4885 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.083480 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.074586 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.074706 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.074875 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.074960 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.075133 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.075459 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.084404 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.075867 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.076082 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.077224 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.077251 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.077375 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.077822 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.077884 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.077946 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.084518 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.078528 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.078641 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.078738 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.079153 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.079245 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.079570 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.079604 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.079866 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.079878 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.079986 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.080338 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.080500 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.080861 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: E1002 01:47:12.081054 4885 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.081027 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.081349 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.081569 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.081813 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.081887 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.081900 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.081944 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.081991 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.082024 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.082190 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.081962 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.082973 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.083027 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.083093 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.083237 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.083821 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.084546 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: E1002 01:47:12.085063 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 01:47:12.585036317 +0000 UTC m=+21.396783756 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.086940 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.086939 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.087099 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.087143 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.087221 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.087322 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.088525 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.088666 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.089791 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.090122 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.090820 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.090868 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.090974 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.091475 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.091910 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.092252 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.092297 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.092370 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.092552 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.092797 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.092826 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.093315 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.093377 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.094058 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.094226 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.094417 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.094423 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.094441 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.094507 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.095603 4885 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.095724 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.095842 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.095989 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.096235 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.096367 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.096514 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.096816 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.096012 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.097196 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.097408 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.097423 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.097623 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.097648 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.097718 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.097825 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.098223 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: E1002 01:47:12.098569 4885 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 01:47:12 crc kubenswrapper[4885]: E1002 01:47:12.098733 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 01:47:12.598715909 +0000 UTC m=+21.410463328 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.102117 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a0bb885-2eb0-419c-8a7a-ef634cb3c144\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a89b49cbe08d927629564e607de13655ddf3d01ac4eb5d688b325210306c120c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06c5942805a498d9621a906088aa54538b4eed583d1b8d0e3ba2a6718378a36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b9f6f7fe44d5771a838d0b89c5bdea3ebb79023ae2675ee907ec2766b3d35a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://178b4340ee6904c60c814b99d83ee225161aeb23e438d8d02f856579a13ec4e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bb63bf30d6ecce9e71f0db0c16632bdfa6b33924b673378eb0c38ab229fa837\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T01:47:06Z\\\",\\\"message\\\":\\\"W1002 01:46:55.406027 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1002 01:46:55.406714 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759369615 cert, and key in /tmp/serving-cert-3795040473/serving-signer.crt, /tmp/serving-cert-3795040473/serving-signer.key\\\\nI1002 01:46:55.614670 1 observer_polling.go:159] Starting file observer\\\\nW1002 01:46:55.617662 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1002 01:46:55.618112 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 01:46:55.620347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3795040473/tls.crt::/tmp/serving-cert-3795040473/tls.key\\\\\\\"\\\\nF1002 01:47:06.224459 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90535f97ab2ed4626d47540ebd7baada837fd332140323bec0f49aa38d3c0009\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://592f89a326c9c1559193a71904b47ca27b1dbe1a7b30e03bc48924bb2f35f921\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://592f89a326c9c1559193a71904b47ca27b1dbe1a7b30e03bc48924bb2f35f921\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.102680 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.106797 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.106815 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.107728 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.107836 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.108296 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.114869 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.114882 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.115068 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.115487 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.116634 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.119028 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.123470 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.125825 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.126665 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.126806 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: E1002 01:47:12.126822 4885 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.126849 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: E1002 01:47:12.126859 4885 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 01:47:12 crc kubenswrapper[4885]: E1002 01:47:12.126884 4885 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.126921 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: E1002 01:47:12.126988 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-02 01:47:12.62695968 +0000 UTC m=+21.438707099 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 01:47:12 crc kubenswrapper[4885]: E1002 01:47:12.128657 4885 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 01:47:12 crc kubenswrapper[4885]: E1002 01:47:12.128702 4885 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 01:47:12 crc kubenswrapper[4885]: E1002 01:47:12.128717 4885 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 01:47:12 crc kubenswrapper[4885]: E1002 01:47:12.128796 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-02 01:47:12.628771074 +0000 UTC m=+21.440518463 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.133076 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.133172 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.133328 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.133479 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.133768 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.133847 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.134646 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.135807 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.137742 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.137851 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.138233 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.138588 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.139103 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.139234 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.139863 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.140072 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.140309 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.140327 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.140382 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.140581 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.144562 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.145159 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.146646 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.146971 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.146999 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.147053 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.148736 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.148731 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.149176 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.151054 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.151393 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.151898 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.152314 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.152859 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.153421 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.155873 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.156097 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.157512 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.157840 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.158983 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.161827 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.162717 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.162890 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.163609 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.163756 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.163724 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.166494 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.166307 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.163995 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.164027 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.165216 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.166802 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a0bb885-2eb0-419c-8a7a-ef634cb3c144\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a89b49cbe08d927629564e607de13655ddf3d01ac4eb5d688b325210306c120c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06c5942805a498d9621a906088aa54538b4eed583d1b8d0e3ba2a6718378a36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b9f6f7fe44d5771a838d0b89c5bdea3ebb79023ae2675ee907ec2766b3d35a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://178b4340ee6904c60c814b99d83ee225161aeb23e438d8d02f856579a13ec4e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bb63bf30d6ecce9e71f0db0c16632bdfa6b33924b673378eb0c38ab229fa837\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T01:47:06Z\\\",\\\"message\\\":\\\"W1002 01:46:55.406027 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1002 01:46:55.406714 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759369615 cert, and key in /tmp/serving-cert-3795040473/serving-signer.crt, /tmp/serving-cert-3795040473/serving-signer.key\\\\nI1002 01:46:55.614670 1 observer_polling.go:159] Starting file observer\\\\nW1002 01:46:55.617662 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1002 01:46:55.618112 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 01:46:55.620347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3795040473/tls.crt::/tmp/serving-cert-3795040473/tls.key\\\\\\\"\\\\nF1002 01:47:06.224459 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90535f97ab2ed4626d47540ebd7baada837fd332140323bec0f49aa38d3c0009\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://592f89a326c9c1559193a71904b47ca27b1dbe1a7b30e03bc48924bb2f35f921\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://592f89a326c9c1559193a71904b47ca27b1dbe1a7b30e03bc48924bb2f35f921\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.167313 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.167325 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.167535 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.167832 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.167949 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.168121 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.168152 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.168223 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.168258 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.168708 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.169317 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.169412 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.168430 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.170548 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.171236 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.171582 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.171736 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.171788 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.172094 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.172235 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.172435 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.172496 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.173083 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.173391 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.174302 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.178348 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.184597 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.184686 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.184857 4885 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.184890 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.184918 4885 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.184948 4885 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.184973 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.184999 4885 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.185027 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.185052 4885 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.185075 4885 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.185099 4885 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.185125 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.185149 4885 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.185175 4885 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.185201 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.185226 4885 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.185251 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.185351 4885 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.185378 4885 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.185404 4885 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.185431 4885 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.185456 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.185483 4885 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.185516 4885 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.185541 4885 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.185568 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.185593 4885 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.185617 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.185647 4885 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.185674 4885 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.185700 4885 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.185725 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.185750 4885 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.185775 4885 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.185800 4885 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.185825 4885 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.185910 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.185939 4885 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.185965 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.185990 4885 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.186015 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.186041 4885 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.186069 4885 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.186094 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.186118 4885 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.186145 4885 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.186169 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.186193 4885 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.186217 4885 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.186240 4885 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.186303 4885 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.186328 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.186351 4885 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.186374 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.186400 4885 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.186424 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.186447 4885 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.186472 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.186507 4885 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.186532 4885 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.186556 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.186580 4885 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.186663 4885 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.186691 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.186768 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.186794 4885 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.186858 4885 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.186882 4885 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.186949 4885 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.186976 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.187048 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.187072 4885 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.187137 4885 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.187166 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.187233 4885 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.187313 4885 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.187344 4885 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.187369 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.187434 4885 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.187459 4885 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.187484 4885 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.187509 4885 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.187534 4885 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.187561 4885 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.187586 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.187610 4885 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.187634 4885 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.187657 4885 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.187681 4885 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.187705 4885 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.187728 4885 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.187753 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.187777 4885 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.187801 4885 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.187852 4885 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.187882 4885 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.187940 4885 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.187964 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.187987 4885 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.188009 4885 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.188031 4885 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.188054 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.188078 4885 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.188102 4885 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.188124 4885 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.188147 4885 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.188174 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.188198 4885 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.188221 4885 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.188247 4885 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.188302 4885 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.188387 4885 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.188416 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.188492 4885 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.188519 4885 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.188544 4885 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.188568 4885 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.188568 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.188592 4885 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.188671 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.188689 4885 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.188703 4885 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.188707 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.188717 4885 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.188802 4885 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.188830 4885 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.188857 4885 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.188882 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.188909 4885 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.188933 4885 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.189144 4885 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.190331 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.190455 4885 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.190483 4885 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.190509 4885 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.190533 4885 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.190558 4885 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.190585 4885 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.190611 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.190638 4885 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.190661 4885 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.190687 4885 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.190710 4885 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.190733 4885 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.190756 4885 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.190779 4885 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.190802 4885 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.190826 4885 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.190852 4885 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.190877 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.190907 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.190932 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.190956 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.190982 4885 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.191005 4885 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.191025 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.191052 4885 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.191077 4885 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.191102 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.191129 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.191247 4885 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.191313 4885 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.190339 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.191338 4885 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.191587 4885 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.191608 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.191623 4885 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.191636 4885 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.191675 4885 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.191687 4885 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.191699 4885 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.191715 4885 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.191801 4885 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.191860 4885 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.191874 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.191886 4885 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.209330 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.230336 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.233608 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.235237 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.244468 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.244795 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.271233 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.285575 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.293020 4885 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.293051 4885 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.293061 4885 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.293070 4885 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.296100 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.296168 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"506a599f-fb1c-4d14-be88-0e4833f3d624\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31c2f0193ffef54965b79a3569b0f6f4547e540bac0774155159a1c996c196c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://533e73e5f90b7258acbcdf018ee01323c118f9dc45035b25b065ba01cabecccf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c00db833a593ec40eb63a6b383455e19e960c2e98b9b66f76877ee993057e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21e2f462a0deb0d740185663b2014f4a1e57b4fdbeb82571264b002f928035bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.309736 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.316315 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 01:47:12 crc kubenswrapper[4885]: W1002 01:47:12.318198 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-d1ee98ece8df4b6a3a17f1384a8b7d41a8db3eec4d60dd9b2c3dd62528cad299 WatchSource:0}: Error finding container d1ee98ece8df4b6a3a17f1384a8b7d41a8db3eec4d60dd9b2c3dd62528cad299: Status 404 returned error can't find the container with id d1ee98ece8df4b6a3a17f1384a8b7d41a8db3eec4d60dd9b2c3dd62528cad299 Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.321688 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.325239 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.333183 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 01:47:12 crc kubenswrapper[4885]: W1002 01:47:12.334413 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-b1d393c7163305b09de7069a616cdf7840488a261d43fd89973b61da5e081d19 WatchSource:0}: Error finding container b1d393c7163305b09de7069a616cdf7840488a261d43fd89973b61da5e081d19: Status 404 returned error can't find the container with id b1d393c7163305b09de7069a616cdf7840488a261d43fd89973b61da5e081d19 Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.341410 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.351521 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a0bb885-2eb0-419c-8a7a-ef634cb3c144\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a89b49cbe08d927629564e607de13655ddf3d01ac4eb5d688b325210306c120c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06c5942805a498d9621a906088aa54538b4eed583d1b8d0e3ba2a6718378a36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b9f6f7fe44d5771a838d0b89c5bdea3ebb79023ae2675ee907ec2766b3d35a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://178b4340ee6904c60c814b99d83ee225161aeb23e438d8d02f856579a13ec4e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bb63bf30d6ecce9e71f0db0c16632bdfa6b33924b673378eb0c38ab229fa837\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T01:47:06Z\\\",\\\"message\\\":\\\"W1002 01:46:55.406027 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1002 01:46:55.406714 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759369615 cert, and key in /tmp/serving-cert-3795040473/serving-signer.crt, /tmp/serving-cert-3795040473/serving-signer.key\\\\nI1002 01:46:55.614670 1 observer_polling.go:159] Starting file observer\\\\nW1002 01:46:55.617662 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1002 01:46:55.618112 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 01:46:55.620347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3795040473/tls.crt::/tmp/serving-cert-3795040473/tls.key\\\\\\\"\\\\nF1002 01:47:06.224459 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90535f97ab2ed4626d47540ebd7baada837fd332140323bec0f49aa38d3c0009\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://592f89a326c9c1559193a71904b47ca27b1dbe1a7b30e03bc48924bb2f35f921\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://592f89a326c9c1559193a71904b47ca27b1dbe1a7b30e03bc48924bb2f35f921\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.360250 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.369146 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.595520 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:47:12 crc kubenswrapper[4885]: E1002 01:47:12.595533 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:47:13.59550986 +0000 UTC m=+22.407257259 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.595730 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:47:12 crc kubenswrapper[4885]: E1002 01:47:12.595827 4885 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 01:47:12 crc kubenswrapper[4885]: E1002 01:47:12.595876 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 01:47:13.595868201 +0000 UTC m=+22.407615600 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.696887 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.696928 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:47:12 crc kubenswrapper[4885]: I1002 01:47:12.696962 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:47:12 crc kubenswrapper[4885]: E1002 01:47:12.697032 4885 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 01:47:12 crc kubenswrapper[4885]: E1002 01:47:12.697079 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 01:47:13.697065929 +0000 UTC m=+22.508813328 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 01:47:12 crc kubenswrapper[4885]: E1002 01:47:12.697113 4885 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 01:47:12 crc kubenswrapper[4885]: E1002 01:47:12.697164 4885 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 01:47:12 crc kubenswrapper[4885]: E1002 01:47:12.697185 4885 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 01:47:12 crc kubenswrapper[4885]: E1002 01:47:12.697285 4885 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 01:47:12 crc kubenswrapper[4885]: E1002 01:47:12.697338 4885 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 01:47:12 crc kubenswrapper[4885]: E1002 01:47:12.697360 4885 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 01:47:12 crc kubenswrapper[4885]: E1002 01:47:12.697292 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-02 01:47:13.697236784 +0000 UTC m=+22.508984223 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 01:47:12 crc kubenswrapper[4885]: E1002 01:47:12.697502 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-02 01:47:13.697472811 +0000 UTC m=+22.509220250 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.198276 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"b1d393c7163305b09de7069a616cdf7840488a261d43fd89973b61da5e081d19"} Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.199634 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"74adcf912ccedb6b5d4f2803fd215e3e2ef62367597aa6da528354c4511ee536"} Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.199687 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"a7403ef968d040ac09d722b7557438bef23892d578992826f42b700b370d495d"} Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.199698 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"34c481276c961cb4e62c6df5eb0594ce721126b9ab88ef120f8ccd54309c4bf0"} Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.200736 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"b087880e05039816dcab154643ac72c7e4a0a1124817c51dc2c639b62bbee0fb"} Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.200768 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"d1ee98ece8df4b6a3a17f1384a8b7d41a8db3eec4d60dd9b2c3dd62528cad299"} Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.214567 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:13Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.226514 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:13Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.238554 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:13Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.251241 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:13Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.265330 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a0bb885-2eb0-419c-8a7a-ef634cb3c144\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a89b49cbe08d927629564e607de13655ddf3d01ac4eb5d688b325210306c120c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06c5942805a498d9621a906088aa54538b4eed583d1b8d0e3ba2a6718378a36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b9f6f7fe44d5771a838d0b89c5bdea3ebb79023ae2675ee907ec2766b3d35a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://178b4340ee6904c60c814b99d83ee225161aeb23e438d8d02f856579a13ec4e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bb63bf30d6ecce9e71f0db0c16632bdfa6b33924b673378eb0c38ab229fa837\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T01:47:06Z\\\",\\\"message\\\":\\\"W1002 01:46:55.406027 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1002 01:46:55.406714 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759369615 cert, and key in /tmp/serving-cert-3795040473/serving-signer.crt, /tmp/serving-cert-3795040473/serving-signer.key\\\\nI1002 01:46:55.614670 1 observer_polling.go:159] Starting file observer\\\\nW1002 01:46:55.617662 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1002 01:46:55.618112 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 01:46:55.620347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3795040473/tls.crt::/tmp/serving-cert-3795040473/tls.key\\\\\\\"\\\\nF1002 01:47:06.224459 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90535f97ab2ed4626d47540ebd7baada837fd332140323bec0f49aa38d3c0009\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://592f89a326c9c1559193a71904b47ca27b1dbe1a7b30e03bc48924bb2f35f921\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://592f89a326c9c1559193a71904b47ca27b1dbe1a7b30e03bc48924bb2f35f921\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:13Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.275176 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:13Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.290038 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74adcf912ccedb6b5d4f2803fd215e3e2ef62367597aa6da528354c4511ee536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7403ef968d040ac09d722b7557438bef23892d578992826f42b700b370d495d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:13Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.300616 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-58lhj"] Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.300888 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-58lhj" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.305789 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.305789 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.306317 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.319991 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"506a599f-fb1c-4d14-be88-0e4833f3d624\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31c2f0193ffef54965b79a3569b0f6f4547e540bac0774155159a1c996c196c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://533e73e5f90b7258acbcdf018ee01323c118f9dc45035b25b065ba01cabecccf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c00db833a593ec40eb63a6b383455e19e960c2e98b9b66f76877ee993057e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21e2f462a0deb0d740185663b2014f4a1e57b4fdbeb82571264b002f928035bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:13Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.345841 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:13Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.363753 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:13Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.391114 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:13Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.406693 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/ac6a19c4-f259-4dbd-a016-d6c2c849c6cb-hosts-file\") pod \"node-resolver-58lhj\" (UID: \"ac6a19c4-f259-4dbd-a016-d6c2c849c6cb\") " pod="openshift-dns/node-resolver-58lhj" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.406751 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hzc4c\" (UniqueName: \"kubernetes.io/projected/ac6a19c4-f259-4dbd-a016-d6c2c849c6cb-kube-api-access-hzc4c\") pod \"node-resolver-58lhj\" (UID: \"ac6a19c4-f259-4dbd-a016-d6c2c849c6cb\") " pod="openshift-dns/node-resolver-58lhj" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.408313 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a0bb885-2eb0-419c-8a7a-ef634cb3c144\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a89b49cbe08d927629564e607de13655ddf3d01ac4eb5d688b325210306c120c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06c5942805a498d9621a906088aa54538b4eed583d1b8d0e3ba2a6718378a36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b9f6f7fe44d5771a838d0b89c5bdea3ebb79023ae2675ee907ec2766b3d35a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://178b4340ee6904c60c814b99d83ee225161aeb23e438d8d02f856579a13ec4e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bb63bf30d6ecce9e71f0db0c16632bdfa6b33924b673378eb0c38ab229fa837\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T01:47:06Z\\\",\\\"message\\\":\\\"W1002 01:46:55.406027 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1002 01:46:55.406714 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759369615 cert, and key in /tmp/serving-cert-3795040473/serving-signer.crt, /tmp/serving-cert-3795040473/serving-signer.key\\\\nI1002 01:46:55.614670 1 observer_polling.go:159] Starting file observer\\\\nW1002 01:46:55.617662 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1002 01:46:55.618112 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 01:46:55.620347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3795040473/tls.crt::/tmp/serving-cert-3795040473/tls.key\\\\\\\"\\\\nF1002 01:47:06.224459 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90535f97ab2ed4626d47540ebd7baada837fd332140323bec0f49aa38d3c0009\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://592f89a326c9c1559193a71904b47ca27b1dbe1a7b30e03bc48924bb2f35f921\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://592f89a326c9c1559193a71904b47ca27b1dbe1a7b30e03bc48924bb2f35f921\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:13Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.423056 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b087880e05039816dcab154643ac72c7e4a0a1124817c51dc2c639b62bbee0fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:13Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.433540 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-58lhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac6a19c4-f259-4dbd-a016-d6c2c849c6cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hzc4c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-58lhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:13Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.445027 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"506a599f-fb1c-4d14-be88-0e4833f3d624\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31c2f0193ffef54965b79a3569b0f6f4547e540bac0774155159a1c996c196c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://533e73e5f90b7258acbcdf018ee01323c118f9dc45035b25b065ba01cabecccf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c00db833a593ec40eb63a6b383455e19e960c2e98b9b66f76877ee993057e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21e2f462a0deb0d740185663b2014f4a1e57b4fdbeb82571264b002f928035bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:13Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.454525 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:13Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.464086 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74adcf912ccedb6b5d4f2803fd215e3e2ef62367597aa6da528354c4511ee536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7403ef968d040ac09d722b7557438bef23892d578992826f42b700b370d495d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:13Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.507416 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/ac6a19c4-f259-4dbd-a016-d6c2c849c6cb-hosts-file\") pod \"node-resolver-58lhj\" (UID: \"ac6a19c4-f259-4dbd-a016-d6c2c849c6cb\") " pod="openshift-dns/node-resolver-58lhj" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.507469 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hzc4c\" (UniqueName: \"kubernetes.io/projected/ac6a19c4-f259-4dbd-a016-d6c2c849c6cb-kube-api-access-hzc4c\") pod \"node-resolver-58lhj\" (UID: \"ac6a19c4-f259-4dbd-a016-d6c2c849c6cb\") " pod="openshift-dns/node-resolver-58lhj" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.507810 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/ac6a19c4-f259-4dbd-a016-d6c2c849c6cb-hosts-file\") pod \"node-resolver-58lhj\" (UID: \"ac6a19c4-f259-4dbd-a016-d6c2c849c6cb\") " pod="openshift-dns/node-resolver-58lhj" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.523500 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hzc4c\" (UniqueName: \"kubernetes.io/projected/ac6a19c4-f259-4dbd-a016-d6c2c849c6cb-kube-api-access-hzc4c\") pod \"node-resolver-58lhj\" (UID: \"ac6a19c4-f259-4dbd-a016-d6c2c849c6cb\") " pod="openshift-dns/node-resolver-58lhj" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.608281 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.608381 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:47:13 crc kubenswrapper[4885]: E1002 01:47:13.608505 4885 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 01:47:13 crc kubenswrapper[4885]: E1002 01:47:13.608937 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:47:15.608883166 +0000 UTC m=+24.420630605 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:47:13 crc kubenswrapper[4885]: E1002 01:47:13.609152 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 01:47:15.609133803 +0000 UTC m=+24.420881232 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.611691 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-58lhj" Oct 02 01:47:13 crc kubenswrapper[4885]: W1002 01:47:13.624795 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podac6a19c4_f259_4dbd_a016_d6c2c849c6cb.slice/crio-9a977fddf52cca1839c83bff425aeccc6ddbf13e80d607951ab123a5df6e05ce WatchSource:0}: Error finding container 9a977fddf52cca1839c83bff425aeccc6ddbf13e80d607951ab123a5df6e05ce: Status 404 returned error can't find the container with id 9a977fddf52cca1839c83bff425aeccc6ddbf13e80d607951ab123a5df6e05ce Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.697555 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-vhgbd"] Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.697838 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-vhgbd" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.698724 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-rttx8"] Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.698944 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.700335 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.700765 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.700862 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.701149 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.701314 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.701456 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.701788 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.701890 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.702032 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.702300 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-2ll9b"] Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.702742 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-2ll9b" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.703303 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.704795 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.705014 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.708632 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.708662 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.708693 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:47:13 crc kubenswrapper[4885]: E1002 01:47:13.708747 4885 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 01:47:13 crc kubenswrapper[4885]: E1002 01:47:13.708783 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 01:47:15.708772385 +0000 UTC m=+24.520519784 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 01:47:13 crc kubenswrapper[4885]: E1002 01:47:13.708964 4885 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 01:47:13 crc kubenswrapper[4885]: E1002 01:47:13.708988 4885 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 01:47:13 crc kubenswrapper[4885]: E1002 01:47:13.708999 4885 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 01:47:13 crc kubenswrapper[4885]: E1002 01:47:13.709040 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-02 01:47:15.709025543 +0000 UTC m=+24.520772942 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 01:47:13 crc kubenswrapper[4885]: E1002 01:47:13.708967 4885 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 01:47:13 crc kubenswrapper[4885]: E1002 01:47:13.709069 4885 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 01:47:13 crc kubenswrapper[4885]: E1002 01:47:13.709082 4885 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 01:47:13 crc kubenswrapper[4885]: E1002 01:47:13.709114 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-02 01:47:15.709106526 +0000 UTC m=+24.520853925 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.720886 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"506a599f-fb1c-4d14-be88-0e4833f3d624\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31c2f0193ffef54965b79a3569b0f6f4547e540bac0774155159a1c996c196c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://533e73e5f90b7258acbcdf018ee01323c118f9dc45035b25b065ba01cabecccf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c00db833a593ec40eb63a6b383455e19e960c2e98b9b66f76877ee993057e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21e2f462a0deb0d740185663b2014f4a1e57b4fdbeb82571264b002f928035bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:13Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.733917 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:13Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.749678 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74adcf912ccedb6b5d4f2803fd215e3e2ef62367597aa6da528354c4511ee536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7403ef968d040ac09d722b7557438bef23892d578992826f42b700b370d495d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:13Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.764747 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:13Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.775804 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:13Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.800406 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a0bb885-2eb0-419c-8a7a-ef634cb3c144\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a89b49cbe08d927629564e607de13655ddf3d01ac4eb5d688b325210306c120c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06c5942805a498d9621a906088aa54538b4eed583d1b8d0e3ba2a6718378a36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b9f6f7fe44d5771a838d0b89c5bdea3ebb79023ae2675ee907ec2766b3d35a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://178b4340ee6904c60c814b99d83ee225161aeb23e438d8d02f856579a13ec4e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bb63bf30d6ecce9e71f0db0c16632bdfa6b33924b673378eb0c38ab229fa837\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T01:47:06Z\\\",\\\"message\\\":\\\"W1002 01:46:55.406027 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1002 01:46:55.406714 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759369615 cert, and key in /tmp/serving-cert-3795040473/serving-signer.crt, /tmp/serving-cert-3795040473/serving-signer.key\\\\nI1002 01:46:55.614670 1 observer_polling.go:159] Starting file observer\\\\nW1002 01:46:55.617662 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1002 01:46:55.618112 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 01:46:55.620347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3795040473/tls.crt::/tmp/serving-cert-3795040473/tls.key\\\\\\\"\\\\nF1002 01:47:06.224459 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90535f97ab2ed4626d47540ebd7baada837fd332140323bec0f49aa38d3c0009\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://592f89a326c9c1559193a71904b47ca27b1dbe1a7b30e03bc48924bb2f35f921\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://592f89a326c9c1559193a71904b47ca27b1dbe1a7b30e03bc48924bb2f35f921\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:13Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.809742 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f-cni-binary-copy\") pod \"multus-vhgbd\" (UID: \"ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f\") " pod="openshift-multus/multus-vhgbd" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.809776 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f-host-var-lib-cni-multus\") pod \"multus-vhgbd\" (UID: \"ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f\") " pod="openshift-multus/multus-vhgbd" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.809796 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/52b39712-5fa1-4fe0-814b-b170a6e3938b-cni-binary-copy\") pod \"multus-additional-cni-plugins-2ll9b\" (UID: \"52b39712-5fa1-4fe0-814b-b170a6e3938b\") " pod="openshift-multus/multus-additional-cni-plugins-2ll9b" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.809812 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f-host-var-lib-kubelet\") pod \"multus-vhgbd\" (UID: \"ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f\") " pod="openshift-multus/multus-vhgbd" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.809832 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f-hostroot\") pod \"multus-vhgbd\" (UID: \"ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f\") " pod="openshift-multus/multus-vhgbd" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.809847 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b7zwd\" (UniqueName: \"kubernetes.io/projected/ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f-kube-api-access-b7zwd\") pod \"multus-vhgbd\" (UID: \"ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f\") " pod="openshift-multus/multus-vhgbd" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.809879 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d169c8ba-000a-4218-8e3b-4ae53035b110-mcd-auth-proxy-config\") pod \"machine-config-daemon-rttx8\" (UID: \"d169c8ba-000a-4218-8e3b-4ae53035b110\") " pod="openshift-machine-config-operator/machine-config-daemon-rttx8" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.809896 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/52b39712-5fa1-4fe0-814b-b170a6e3938b-cnibin\") pod \"multus-additional-cni-plugins-2ll9b\" (UID: \"52b39712-5fa1-4fe0-814b-b170a6e3938b\") " pod="openshift-multus/multus-additional-cni-plugins-2ll9b" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.809911 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/52b39712-5fa1-4fe0-814b-b170a6e3938b-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-2ll9b\" (UID: \"52b39712-5fa1-4fe0-814b-b170a6e3938b\") " pod="openshift-multus/multus-additional-cni-plugins-2ll9b" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.809926 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-928wt\" (UniqueName: \"kubernetes.io/projected/52b39712-5fa1-4fe0-814b-b170a6e3938b-kube-api-access-928wt\") pod \"multus-additional-cni-plugins-2ll9b\" (UID: \"52b39712-5fa1-4fe0-814b-b170a6e3938b\") " pod="openshift-multus/multus-additional-cni-plugins-2ll9b" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.809940 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f-multus-cni-dir\") pod \"multus-vhgbd\" (UID: \"ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f\") " pod="openshift-multus/multus-vhgbd" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.809956 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f-host-run-multus-certs\") pod \"multus-vhgbd\" (UID: \"ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f\") " pod="openshift-multus/multus-vhgbd" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.810045 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/52b39712-5fa1-4fe0-814b-b170a6e3938b-os-release\") pod \"multus-additional-cni-plugins-2ll9b\" (UID: \"52b39712-5fa1-4fe0-814b-b170a6e3938b\") " pod="openshift-multus/multus-additional-cni-plugins-2ll9b" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.810083 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f-multus-socket-dir-parent\") pod \"multus-vhgbd\" (UID: \"ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f\") " pod="openshift-multus/multus-vhgbd" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.810103 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f-cnibin\") pod \"multus-vhgbd\" (UID: \"ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f\") " pod="openshift-multus/multus-vhgbd" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.810122 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f-host-run-k8s-cni-cncf-io\") pod \"multus-vhgbd\" (UID: \"ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f\") " pod="openshift-multus/multus-vhgbd" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.810137 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f-multus-conf-dir\") pod \"multus-vhgbd\" (UID: \"ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f\") " pod="openshift-multus/multus-vhgbd" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.810155 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f-multus-daemon-config\") pod \"multus-vhgbd\" (UID: \"ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f\") " pod="openshift-multus/multus-vhgbd" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.810177 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f-host-run-netns\") pod \"multus-vhgbd\" (UID: \"ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f\") " pod="openshift-multus/multus-vhgbd" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.810209 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f-system-cni-dir\") pod \"multus-vhgbd\" (UID: \"ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f\") " pod="openshift-multus/multus-vhgbd" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.810226 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-77rzs\" (UniqueName: \"kubernetes.io/projected/d169c8ba-000a-4218-8e3b-4ae53035b110-kube-api-access-77rzs\") pod \"machine-config-daemon-rttx8\" (UID: \"d169c8ba-000a-4218-8e3b-4ae53035b110\") " pod="openshift-machine-config-operator/machine-config-daemon-rttx8" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.810251 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f-os-release\") pod \"multus-vhgbd\" (UID: \"ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f\") " pod="openshift-multus/multus-vhgbd" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.810285 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f-etc-kubernetes\") pod \"multus-vhgbd\" (UID: \"ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f\") " pod="openshift-multus/multus-vhgbd" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.810313 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/d169c8ba-000a-4218-8e3b-4ae53035b110-rootfs\") pod \"machine-config-daemon-rttx8\" (UID: \"d169c8ba-000a-4218-8e3b-4ae53035b110\") " pod="openshift-machine-config-operator/machine-config-daemon-rttx8" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.810341 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/52b39712-5fa1-4fe0-814b-b170a6e3938b-system-cni-dir\") pod \"multus-additional-cni-plugins-2ll9b\" (UID: \"52b39712-5fa1-4fe0-814b-b170a6e3938b\") " pod="openshift-multus/multus-additional-cni-plugins-2ll9b" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.810357 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/52b39712-5fa1-4fe0-814b-b170a6e3938b-tuning-conf-dir\") pod \"multus-additional-cni-plugins-2ll9b\" (UID: \"52b39712-5fa1-4fe0-814b-b170a6e3938b\") " pod="openshift-multus/multus-additional-cni-plugins-2ll9b" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.810372 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f-host-var-lib-cni-bin\") pod \"multus-vhgbd\" (UID: \"ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f\") " pod="openshift-multus/multus-vhgbd" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.810408 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d169c8ba-000a-4218-8e3b-4ae53035b110-proxy-tls\") pod \"machine-config-daemon-rttx8\" (UID: \"d169c8ba-000a-4218-8e3b-4ae53035b110\") " pod="openshift-machine-config-operator/machine-config-daemon-rttx8" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.812791 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b087880e05039816dcab154643ac72c7e4a0a1124817c51dc2c639b62bbee0fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:13Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.823050 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:13Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.833235 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-58lhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac6a19c4-f259-4dbd-a016-d6c2c849c6cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hzc4c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-58lhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:13Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.848822 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vhgbd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7zwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vhgbd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:13Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.864529 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:13Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.882803 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74adcf912ccedb6b5d4f2803fd215e3e2ef62367597aa6da528354c4511ee536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7403ef968d040ac09d722b7557438bef23892d578992826f42b700b370d495d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:13Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.900517 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2ll9b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52b39712-5fa1-4fe0-814b-b170a6e3938b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2ll9b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:13Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.911412 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f-hostroot\") pod \"multus-vhgbd\" (UID: \"ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f\") " pod="openshift-multus/multus-vhgbd" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.911452 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b7zwd\" (UniqueName: \"kubernetes.io/projected/ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f-kube-api-access-b7zwd\") pod \"multus-vhgbd\" (UID: \"ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f\") " pod="openshift-multus/multus-vhgbd" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.911468 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d169c8ba-000a-4218-8e3b-4ae53035b110-mcd-auth-proxy-config\") pod \"machine-config-daemon-rttx8\" (UID: \"d169c8ba-000a-4218-8e3b-4ae53035b110\") " pod="openshift-machine-config-operator/machine-config-daemon-rttx8" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.911484 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f-host-run-multus-certs\") pod \"multus-vhgbd\" (UID: \"ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f\") " pod="openshift-multus/multus-vhgbd" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.911498 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/52b39712-5fa1-4fe0-814b-b170a6e3938b-cnibin\") pod \"multus-additional-cni-plugins-2ll9b\" (UID: \"52b39712-5fa1-4fe0-814b-b170a6e3938b\") " pod="openshift-multus/multus-additional-cni-plugins-2ll9b" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.911517 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/52b39712-5fa1-4fe0-814b-b170a6e3938b-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-2ll9b\" (UID: \"52b39712-5fa1-4fe0-814b-b170a6e3938b\") " pod="openshift-multus/multus-additional-cni-plugins-2ll9b" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.911532 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-928wt\" (UniqueName: \"kubernetes.io/projected/52b39712-5fa1-4fe0-814b-b170a6e3938b-kube-api-access-928wt\") pod \"multus-additional-cni-plugins-2ll9b\" (UID: \"52b39712-5fa1-4fe0-814b-b170a6e3938b\") " pod="openshift-multus/multus-additional-cni-plugins-2ll9b" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.911559 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f-multus-cni-dir\") pod \"multus-vhgbd\" (UID: \"ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f\") " pod="openshift-multus/multus-vhgbd" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.911580 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/52b39712-5fa1-4fe0-814b-b170a6e3938b-os-release\") pod \"multus-additional-cni-plugins-2ll9b\" (UID: \"52b39712-5fa1-4fe0-814b-b170a6e3938b\") " pod="openshift-multus/multus-additional-cni-plugins-2ll9b" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.911595 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f-multus-socket-dir-parent\") pod \"multus-vhgbd\" (UID: \"ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f\") " pod="openshift-multus/multus-vhgbd" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.911621 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f-cnibin\") pod \"multus-vhgbd\" (UID: \"ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f\") " pod="openshift-multus/multus-vhgbd" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.911654 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f-host-run-k8s-cni-cncf-io\") pod \"multus-vhgbd\" (UID: \"ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f\") " pod="openshift-multus/multus-vhgbd" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.911668 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f-multus-conf-dir\") pod \"multus-vhgbd\" (UID: \"ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f\") " pod="openshift-multus/multus-vhgbd" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.911684 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f-multus-daemon-config\") pod \"multus-vhgbd\" (UID: \"ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f\") " pod="openshift-multus/multus-vhgbd" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.911700 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f-host-run-netns\") pod \"multus-vhgbd\" (UID: \"ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f\") " pod="openshift-multus/multus-vhgbd" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.911719 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f-system-cni-dir\") pod \"multus-vhgbd\" (UID: \"ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f\") " pod="openshift-multus/multus-vhgbd" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.911733 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-77rzs\" (UniqueName: \"kubernetes.io/projected/d169c8ba-000a-4218-8e3b-4ae53035b110-kube-api-access-77rzs\") pod \"machine-config-daemon-rttx8\" (UID: \"d169c8ba-000a-4218-8e3b-4ae53035b110\") " pod="openshift-machine-config-operator/machine-config-daemon-rttx8" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.911746 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f-os-release\") pod \"multus-vhgbd\" (UID: \"ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f\") " pod="openshift-multus/multus-vhgbd" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.911762 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f-etc-kubernetes\") pod \"multus-vhgbd\" (UID: \"ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f\") " pod="openshift-multus/multus-vhgbd" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.911776 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/d169c8ba-000a-4218-8e3b-4ae53035b110-rootfs\") pod \"machine-config-daemon-rttx8\" (UID: \"d169c8ba-000a-4218-8e3b-4ae53035b110\") " pod="openshift-machine-config-operator/machine-config-daemon-rttx8" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.911790 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/52b39712-5fa1-4fe0-814b-b170a6e3938b-system-cni-dir\") pod \"multus-additional-cni-plugins-2ll9b\" (UID: \"52b39712-5fa1-4fe0-814b-b170a6e3938b\") " pod="openshift-multus/multus-additional-cni-plugins-2ll9b" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.911805 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/52b39712-5fa1-4fe0-814b-b170a6e3938b-tuning-conf-dir\") pod \"multus-additional-cni-plugins-2ll9b\" (UID: \"52b39712-5fa1-4fe0-814b-b170a6e3938b\") " pod="openshift-multus/multus-additional-cni-plugins-2ll9b" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.911819 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f-host-var-lib-cni-bin\") pod \"multus-vhgbd\" (UID: \"ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f\") " pod="openshift-multus/multus-vhgbd" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.911834 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d169c8ba-000a-4218-8e3b-4ae53035b110-proxy-tls\") pod \"machine-config-daemon-rttx8\" (UID: \"d169c8ba-000a-4218-8e3b-4ae53035b110\") " pod="openshift-machine-config-operator/machine-config-daemon-rttx8" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.911854 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f-cni-binary-copy\") pod \"multus-vhgbd\" (UID: \"ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f\") " pod="openshift-multus/multus-vhgbd" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.911869 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f-host-var-lib-cni-multus\") pod \"multus-vhgbd\" (UID: \"ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f\") " pod="openshift-multus/multus-vhgbd" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.911884 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/52b39712-5fa1-4fe0-814b-b170a6e3938b-cni-binary-copy\") pod \"multus-additional-cni-plugins-2ll9b\" (UID: \"52b39712-5fa1-4fe0-814b-b170a6e3938b\") " pod="openshift-multus/multus-additional-cni-plugins-2ll9b" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.911898 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f-host-var-lib-kubelet\") pod \"multus-vhgbd\" (UID: \"ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f\") " pod="openshift-multus/multus-vhgbd" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.911958 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f-host-var-lib-kubelet\") pod \"multus-vhgbd\" (UID: \"ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f\") " pod="openshift-multus/multus-vhgbd" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.911988 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f-hostroot\") pod \"multus-vhgbd\" (UID: \"ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f\") " pod="openshift-multus/multus-vhgbd" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.913041 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d169c8ba-000a-4218-8e3b-4ae53035b110-mcd-auth-proxy-config\") pod \"machine-config-daemon-rttx8\" (UID: \"d169c8ba-000a-4218-8e3b-4ae53035b110\") " pod="openshift-machine-config-operator/machine-config-daemon-rttx8" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.913082 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f-host-run-multus-certs\") pod \"multus-vhgbd\" (UID: \"ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f\") " pod="openshift-multus/multus-vhgbd" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.913103 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/52b39712-5fa1-4fe0-814b-b170a6e3938b-cnibin\") pod \"multus-additional-cni-plugins-2ll9b\" (UID: \"52b39712-5fa1-4fe0-814b-b170a6e3938b\") " pod="openshift-multus/multus-additional-cni-plugins-2ll9b" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.913551 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/52b39712-5fa1-4fe0-814b-b170a6e3938b-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-2ll9b\" (UID: \"52b39712-5fa1-4fe0-814b-b170a6e3938b\") " pod="openshift-multus/multus-additional-cni-plugins-2ll9b" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.913792 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f-multus-cni-dir\") pod \"multus-vhgbd\" (UID: \"ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f\") " pod="openshift-multus/multus-vhgbd" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.913974 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/52b39712-5fa1-4fe0-814b-b170a6e3938b-os-release\") pod \"multus-additional-cni-plugins-2ll9b\" (UID: \"52b39712-5fa1-4fe0-814b-b170a6e3938b\") " pod="openshift-multus/multus-additional-cni-plugins-2ll9b" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.914026 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f-multus-socket-dir-parent\") pod \"multus-vhgbd\" (UID: \"ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f\") " pod="openshift-multus/multus-vhgbd" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.914056 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f-cnibin\") pod \"multus-vhgbd\" (UID: \"ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f\") " pod="openshift-multus/multus-vhgbd" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.914078 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f-host-run-k8s-cni-cncf-io\") pod \"multus-vhgbd\" (UID: \"ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f\") " pod="openshift-multus/multus-vhgbd" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.914098 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f-multus-conf-dir\") pod \"multus-vhgbd\" (UID: \"ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f\") " pod="openshift-multus/multus-vhgbd" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.914548 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f-multus-daemon-config\") pod \"multus-vhgbd\" (UID: \"ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f\") " pod="openshift-multus/multus-vhgbd" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.914587 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f-host-run-netns\") pod \"multus-vhgbd\" (UID: \"ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f\") " pod="openshift-multus/multus-vhgbd" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.914620 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f-system-cni-dir\") pod \"multus-vhgbd\" (UID: \"ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f\") " pod="openshift-multus/multus-vhgbd" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.914771 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f-os-release\") pod \"multus-vhgbd\" (UID: \"ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f\") " pod="openshift-multus/multus-vhgbd" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.914801 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f-etc-kubernetes\") pod \"multus-vhgbd\" (UID: \"ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f\") " pod="openshift-multus/multus-vhgbd" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.914834 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/d169c8ba-000a-4218-8e3b-4ae53035b110-rootfs\") pod \"machine-config-daemon-rttx8\" (UID: \"d169c8ba-000a-4218-8e3b-4ae53035b110\") " pod="openshift-machine-config-operator/machine-config-daemon-rttx8" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.914856 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/52b39712-5fa1-4fe0-814b-b170a6e3938b-system-cni-dir\") pod \"multus-additional-cni-plugins-2ll9b\" (UID: \"52b39712-5fa1-4fe0-814b-b170a6e3938b\") " pod="openshift-multus/multus-additional-cni-plugins-2ll9b" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.915177 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f-host-var-lib-cni-multus\") pod \"multus-vhgbd\" (UID: \"ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f\") " pod="openshift-multus/multus-vhgbd" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.915682 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/52b39712-5fa1-4fe0-814b-b170a6e3938b-cni-binary-copy\") pod \"multus-additional-cni-plugins-2ll9b\" (UID: \"52b39712-5fa1-4fe0-814b-b170a6e3938b\") " pod="openshift-multus/multus-additional-cni-plugins-2ll9b" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.915741 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f-host-var-lib-cni-bin\") pod \"multus-vhgbd\" (UID: \"ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f\") " pod="openshift-multus/multus-vhgbd" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.915864 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f-cni-binary-copy\") pod \"multus-vhgbd\" (UID: \"ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f\") " pod="openshift-multus/multus-vhgbd" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.915938 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/52b39712-5fa1-4fe0-814b-b170a6e3938b-tuning-conf-dir\") pod \"multus-additional-cni-plugins-2ll9b\" (UID: \"52b39712-5fa1-4fe0-814b-b170a6e3938b\") " pod="openshift-multus/multus-additional-cni-plugins-2ll9b" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.923123 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d169c8ba-000a-4218-8e3b-4ae53035b110\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-77rzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-77rzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rttx8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:13Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.940523 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a0bb885-2eb0-419c-8a7a-ef634cb3c144\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a89b49cbe08d927629564e607de13655ddf3d01ac4eb5d688b325210306c120c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06c5942805a498d9621a906088aa54538b4eed583d1b8d0e3ba2a6718378a36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b9f6f7fe44d5771a838d0b89c5bdea3ebb79023ae2675ee907ec2766b3d35a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://178b4340ee6904c60c814b99d83ee225161aeb23e438d8d02f856579a13ec4e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bb63bf30d6ecce9e71f0db0c16632bdfa6b33924b673378eb0c38ab229fa837\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T01:47:06Z\\\",\\\"message\\\":\\\"W1002 01:46:55.406027 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1002 01:46:55.406714 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759369615 cert, and key in /tmp/serving-cert-3795040473/serving-signer.crt, /tmp/serving-cert-3795040473/serving-signer.key\\\\nI1002 01:46:55.614670 1 observer_polling.go:159] Starting file observer\\\\nW1002 01:46:55.617662 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1002 01:46:55.618112 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 01:46:55.620347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3795040473/tls.crt::/tmp/serving-cert-3795040473/tls.key\\\\\\\"\\\\nF1002 01:47:06.224459 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90535f97ab2ed4626d47540ebd7baada837fd332140323bec0f49aa38d3c0009\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://592f89a326c9c1559193a71904b47ca27b1dbe1a7b30e03bc48924bb2f35f921\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://592f89a326c9c1559193a71904b47ca27b1dbe1a7b30e03bc48924bb2f35f921\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:13Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.957180 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b087880e05039816dcab154643ac72c7e4a0a1124817c51dc2c639b62bbee0fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:13Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.958802 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d169c8ba-000a-4218-8e3b-4ae53035b110-proxy-tls\") pod \"machine-config-daemon-rttx8\" (UID: \"d169c8ba-000a-4218-8e3b-4ae53035b110\") " pod="openshift-machine-config-operator/machine-config-daemon-rttx8" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.958868 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-928wt\" (UniqueName: \"kubernetes.io/projected/52b39712-5fa1-4fe0-814b-b170a6e3938b-kube-api-access-928wt\") pod \"multus-additional-cni-plugins-2ll9b\" (UID: \"52b39712-5fa1-4fe0-814b-b170a6e3938b\") " pod="openshift-multus/multus-additional-cni-plugins-2ll9b" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.959447 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b7zwd\" (UniqueName: \"kubernetes.io/projected/ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f-kube-api-access-b7zwd\") pod \"multus-vhgbd\" (UID: \"ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f\") " pod="openshift-multus/multus-vhgbd" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.959553 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-77rzs\" (UniqueName: \"kubernetes.io/projected/d169c8ba-000a-4218-8e3b-4ae53035b110-kube-api-access-77rzs\") pod \"machine-config-daemon-rttx8\" (UID: \"d169c8ba-000a-4218-8e3b-4ae53035b110\") " pod="openshift-machine-config-operator/machine-config-daemon-rttx8" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.970476 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:13Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.981557 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:13Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:13 crc kubenswrapper[4885]: I1002 01:47:13.994387 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-58lhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac6a19c4-f259-4dbd-a016-d6c2c849c6cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hzc4c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-58lhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:13Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.007085 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"506a599f-fb1c-4d14-be88-0e4833f3d624\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31c2f0193ffef54965b79a3569b0f6f4547e540bac0774155159a1c996c196c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://533e73e5f90b7258acbcdf018ee01323c118f9dc45035b25b065ba01cabecccf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c00db833a593ec40eb63a6b383455e19e960c2e98b9b66f76877ee993057e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21e2f462a0deb0d740185663b2014f4a1e57b4fdbeb82571264b002f928035bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:14Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.015885 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-vhgbd" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.020071 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:14Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.026810 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" Oct 02 01:47:14 crc kubenswrapper[4885]: W1002 01:47:14.030958 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podecfe7ad8_f68d_48fd_bc04_8f7a5f9d447f.slice/crio-d4b69dc4a2b6cb84443ed072dfde3a548ed3a72085267bc75238c211934ae02e WatchSource:0}: Error finding container d4b69dc4a2b6cb84443ed072dfde3a548ed3a72085267bc75238c211934ae02e: Status 404 returned error can't find the container with id d4b69dc4a2b6cb84443ed072dfde3a548ed3a72085267bc75238c211934ae02e Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.038665 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-2ll9b" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.046809 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.046828 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:47:14 crc kubenswrapper[4885]: E1002 01:47:14.046934 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:47:14 crc kubenswrapper[4885]: E1002 01:47:14.047134 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.047180 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:47:14 crc kubenswrapper[4885]: E1002 01:47:14.047233 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:47:14 crc kubenswrapper[4885]: W1002 01:47:14.048786 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd169c8ba_000a_4218_8e3b_4ae53035b110.slice/crio-965c899f5f5b4748577dab8c0c40cc1fe2d0d0c5c8cadc2fd96a282dbce1d34f WatchSource:0}: Error finding container 965c899f5f5b4748577dab8c0c40cc1fe2d0d0c5c8cadc2fd96a282dbce1d34f: Status 404 returned error can't find the container with id 965c899f5f5b4748577dab8c0c40cc1fe2d0d0c5c8cadc2fd96a282dbce1d34f Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.051288 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.051959 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.053290 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.053893 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.054915 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.055565 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.056672 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.057334 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.059659 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.060311 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.060946 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Oct 02 01:47:14 crc kubenswrapper[4885]: W1002 01:47:14.062474 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod52b39712_5fa1_4fe0_814b_b170a6e3938b.slice/crio-284d69006206e5f47d62b340a7dae213999f29021615406ba12c2d2e770af0f4 WatchSource:0}: Error finding container 284d69006206e5f47d62b340a7dae213999f29021615406ba12c2d2e770af0f4: Status 404 returned error can't find the container with id 284d69006206e5f47d62b340a7dae213999f29021615406ba12c2d2e770af0f4 Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.065946 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.066468 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.067409 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.067895 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.068837 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.069408 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.069813 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.070708 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vhgbd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7zwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vhgbd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:14Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.071401 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.072239 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.072868 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.076871 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.079734 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.080551 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.081669 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.082346 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.087404 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.088395 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.089159 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.090078 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.090673 4885 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.090786 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.093063 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.093553 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.093946 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.095769 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.096387 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.096908 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.100929 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.102008 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.102518 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.103108 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.104065 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.105082 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.106780 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.107464 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.108487 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.109402 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.113736 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.114407 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.115316 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.115913 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.116490 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.117400 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.117827 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-dngcm"] Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.118643 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.124420 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.124532 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.124679 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.124699 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.124778 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.124914 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.137654 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.143640 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"506a599f-fb1c-4d14-be88-0e4833f3d624\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31c2f0193ffef54965b79a3569b0f6f4547e540bac0774155159a1c996c196c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://533e73e5f90b7258acbcdf018ee01323c118f9dc45035b25b065ba01cabecccf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c00db833a593ec40eb63a6b383455e19e960c2e98b9b66f76877ee993057e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21e2f462a0deb0d740185663b2014f4a1e57b4fdbeb82571264b002f928035bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:14Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.186903 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e4679dd-f870-41e9-a32b-360f5d2ee81b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dngcm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:14Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.211060 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:14Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.218971 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2ll9b" event={"ID":"52b39712-5fa1-4fe0-814b-b170a6e3938b","Type":"ContainerStarted","Data":"284d69006206e5f47d62b340a7dae213999f29021615406ba12c2d2e770af0f4"} Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.219464 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9e4679dd-f870-41e9-a32b-360f5d2ee81b-host-run-netns\") pod \"ovnkube-node-dngcm\" (UID: \"9e4679dd-f870-41e9-a32b-360f5d2ee81b\") " pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.219502 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9e4679dd-f870-41e9-a32b-360f5d2ee81b-env-overrides\") pod \"ovnkube-node-dngcm\" (UID: \"9e4679dd-f870-41e9-a32b-360f5d2ee81b\") " pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.219518 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4b7jr\" (UniqueName: \"kubernetes.io/projected/9e4679dd-f870-41e9-a32b-360f5d2ee81b-kube-api-access-4b7jr\") pod \"ovnkube-node-dngcm\" (UID: \"9e4679dd-f870-41e9-a32b-360f5d2ee81b\") " pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.219538 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/9e4679dd-f870-41e9-a32b-360f5d2ee81b-systemd-units\") pod \"ovnkube-node-dngcm\" (UID: \"9e4679dd-f870-41e9-a32b-360f5d2ee81b\") " pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.219552 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9e4679dd-f870-41e9-a32b-360f5d2ee81b-ovn-node-metrics-cert\") pod \"ovnkube-node-dngcm\" (UID: \"9e4679dd-f870-41e9-a32b-360f5d2ee81b\") " pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.219578 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9e4679dd-f870-41e9-a32b-360f5d2ee81b-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-dngcm\" (UID: \"9e4679dd-f870-41e9-a32b-360f5d2ee81b\") " pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.219596 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/9e4679dd-f870-41e9-a32b-360f5d2ee81b-log-socket\") pod \"ovnkube-node-dngcm\" (UID: \"9e4679dd-f870-41e9-a32b-360f5d2ee81b\") " pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.219614 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/9e4679dd-f870-41e9-a32b-360f5d2ee81b-host-cni-netd\") pod \"ovnkube-node-dngcm\" (UID: \"9e4679dd-f870-41e9-a32b-360f5d2ee81b\") " pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.219629 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/9e4679dd-f870-41e9-a32b-360f5d2ee81b-host-kubelet\") pod \"ovnkube-node-dngcm\" (UID: \"9e4679dd-f870-41e9-a32b-360f5d2ee81b\") " pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.219642 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9e4679dd-f870-41e9-a32b-360f5d2ee81b-etc-openvswitch\") pod \"ovnkube-node-dngcm\" (UID: \"9e4679dd-f870-41e9-a32b-360f5d2ee81b\") " pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.219657 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/9e4679dd-f870-41e9-a32b-360f5d2ee81b-run-ovn\") pod \"ovnkube-node-dngcm\" (UID: \"9e4679dd-f870-41e9-a32b-360f5d2ee81b\") " pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.219672 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9e4679dd-f870-41e9-a32b-360f5d2ee81b-host-cni-bin\") pod \"ovnkube-node-dngcm\" (UID: \"9e4679dd-f870-41e9-a32b-360f5d2ee81b\") " pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.219692 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9e4679dd-f870-41e9-a32b-360f5d2ee81b-var-lib-openvswitch\") pod \"ovnkube-node-dngcm\" (UID: \"9e4679dd-f870-41e9-a32b-360f5d2ee81b\") " pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.219707 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/9e4679dd-f870-41e9-a32b-360f5d2ee81b-node-log\") pod \"ovnkube-node-dngcm\" (UID: \"9e4679dd-f870-41e9-a32b-360f5d2ee81b\") " pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.219723 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/9e4679dd-f870-41e9-a32b-360f5d2ee81b-ovnkube-script-lib\") pod \"ovnkube-node-dngcm\" (UID: \"9e4679dd-f870-41e9-a32b-360f5d2ee81b\") " pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.219744 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9e4679dd-f870-41e9-a32b-360f5d2ee81b-ovnkube-config\") pod \"ovnkube-node-dngcm\" (UID: \"9e4679dd-f870-41e9-a32b-360f5d2ee81b\") " pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.219759 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/9e4679dd-f870-41e9-a32b-360f5d2ee81b-host-slash\") pod \"ovnkube-node-dngcm\" (UID: \"9e4679dd-f870-41e9-a32b-360f5d2ee81b\") " pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.219772 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/9e4679dd-f870-41e9-a32b-360f5d2ee81b-run-systemd\") pod \"ovnkube-node-dngcm\" (UID: \"9e4679dd-f870-41e9-a32b-360f5d2ee81b\") " pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.219786 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9e4679dd-f870-41e9-a32b-360f5d2ee81b-run-openvswitch\") pod \"ovnkube-node-dngcm\" (UID: \"9e4679dd-f870-41e9-a32b-360f5d2ee81b\") " pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.219800 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9e4679dd-f870-41e9-a32b-360f5d2ee81b-host-run-ovn-kubernetes\") pod \"ovnkube-node-dngcm\" (UID: \"9e4679dd-f870-41e9-a32b-360f5d2ee81b\") " pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.224468 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-58lhj" event={"ID":"ac6a19c4-f259-4dbd-a016-d6c2c849c6cb","Type":"ContainerStarted","Data":"eed76c0f968c063f7795d363af194a222487b647ceda234de0dabb39b25eafe0"} Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.224504 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-58lhj" event={"ID":"ac6a19c4-f259-4dbd-a016-d6c2c849c6cb","Type":"ContainerStarted","Data":"9a977fddf52cca1839c83bff425aeccc6ddbf13e80d607951ab123a5df6e05ce"} Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.234086 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" event={"ID":"d169c8ba-000a-4218-8e3b-4ae53035b110","Type":"ContainerStarted","Data":"965c899f5f5b4748577dab8c0c40cc1fe2d0d0c5c8cadc2fd96a282dbce1d34f"} Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.241667 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-vhgbd" event={"ID":"ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f","Type":"ContainerStarted","Data":"d4b69dc4a2b6cb84443ed072dfde3a548ed3a72085267bc75238c211934ae02e"} Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.244662 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.252695 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vhgbd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7zwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vhgbd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:14Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.276552 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2ll9b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52b39712-5fa1-4fe0-814b-b170a6e3938b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2ll9b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:14Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.278306 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.286381 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.295987 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:14Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.316151 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74adcf912ccedb6b5d4f2803fd215e3e2ef62367597aa6da528354c4511ee536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7403ef968d040ac09d722b7557438bef23892d578992826f42b700b370d495d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:14Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.320140 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9e4679dd-f870-41e9-a32b-360f5d2ee81b-ovnkube-config\") pod \"ovnkube-node-dngcm\" (UID: \"9e4679dd-f870-41e9-a32b-360f5d2ee81b\") " pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.320178 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/9e4679dd-f870-41e9-a32b-360f5d2ee81b-host-slash\") pod \"ovnkube-node-dngcm\" (UID: \"9e4679dd-f870-41e9-a32b-360f5d2ee81b\") " pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.320200 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/9e4679dd-f870-41e9-a32b-360f5d2ee81b-run-systemd\") pod \"ovnkube-node-dngcm\" (UID: \"9e4679dd-f870-41e9-a32b-360f5d2ee81b\") " pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.320215 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9e4679dd-f870-41e9-a32b-360f5d2ee81b-run-openvswitch\") pod \"ovnkube-node-dngcm\" (UID: \"9e4679dd-f870-41e9-a32b-360f5d2ee81b\") " pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.320231 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9e4679dd-f870-41e9-a32b-360f5d2ee81b-host-run-ovn-kubernetes\") pod \"ovnkube-node-dngcm\" (UID: \"9e4679dd-f870-41e9-a32b-360f5d2ee81b\") " pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.320280 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9e4679dd-f870-41e9-a32b-360f5d2ee81b-host-run-netns\") pod \"ovnkube-node-dngcm\" (UID: \"9e4679dd-f870-41e9-a32b-360f5d2ee81b\") " pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.320294 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9e4679dd-f870-41e9-a32b-360f5d2ee81b-env-overrides\") pod \"ovnkube-node-dngcm\" (UID: \"9e4679dd-f870-41e9-a32b-360f5d2ee81b\") " pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.320308 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4b7jr\" (UniqueName: \"kubernetes.io/projected/9e4679dd-f870-41e9-a32b-360f5d2ee81b-kube-api-access-4b7jr\") pod \"ovnkube-node-dngcm\" (UID: \"9e4679dd-f870-41e9-a32b-360f5d2ee81b\") " pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.320324 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/9e4679dd-f870-41e9-a32b-360f5d2ee81b-systemd-units\") pod \"ovnkube-node-dngcm\" (UID: \"9e4679dd-f870-41e9-a32b-360f5d2ee81b\") " pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.320338 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9e4679dd-f870-41e9-a32b-360f5d2ee81b-ovn-node-metrics-cert\") pod \"ovnkube-node-dngcm\" (UID: \"9e4679dd-f870-41e9-a32b-360f5d2ee81b\") " pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.320338 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/9e4679dd-f870-41e9-a32b-360f5d2ee81b-run-systemd\") pod \"ovnkube-node-dngcm\" (UID: \"9e4679dd-f870-41e9-a32b-360f5d2ee81b\") " pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.320362 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9e4679dd-f870-41e9-a32b-360f5d2ee81b-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-dngcm\" (UID: \"9e4679dd-f870-41e9-a32b-360f5d2ee81b\") " pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.320400 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9e4679dd-f870-41e9-a32b-360f5d2ee81b-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-dngcm\" (UID: \"9e4679dd-f870-41e9-a32b-360f5d2ee81b\") " pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.320439 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9e4679dd-f870-41e9-a32b-360f5d2ee81b-run-openvswitch\") pod \"ovnkube-node-dngcm\" (UID: \"9e4679dd-f870-41e9-a32b-360f5d2ee81b\") " pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.320463 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9e4679dd-f870-41e9-a32b-360f5d2ee81b-host-run-ovn-kubernetes\") pod \"ovnkube-node-dngcm\" (UID: \"9e4679dd-f870-41e9-a32b-360f5d2ee81b\") " pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.320469 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/9e4679dd-f870-41e9-a32b-360f5d2ee81b-log-socket\") pod \"ovnkube-node-dngcm\" (UID: \"9e4679dd-f870-41e9-a32b-360f5d2ee81b\") " pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.320474 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/9e4679dd-f870-41e9-a32b-360f5d2ee81b-host-slash\") pod \"ovnkube-node-dngcm\" (UID: \"9e4679dd-f870-41e9-a32b-360f5d2ee81b\") " pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.320491 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/9e4679dd-f870-41e9-a32b-360f5d2ee81b-host-cni-netd\") pod \"ovnkube-node-dngcm\" (UID: \"9e4679dd-f870-41e9-a32b-360f5d2ee81b\") " pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.320512 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/9e4679dd-f870-41e9-a32b-360f5d2ee81b-host-kubelet\") pod \"ovnkube-node-dngcm\" (UID: \"9e4679dd-f870-41e9-a32b-360f5d2ee81b\") " pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.320529 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9e4679dd-f870-41e9-a32b-360f5d2ee81b-etc-openvswitch\") pod \"ovnkube-node-dngcm\" (UID: \"9e4679dd-f870-41e9-a32b-360f5d2ee81b\") " pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.320566 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/9e4679dd-f870-41e9-a32b-360f5d2ee81b-run-ovn\") pod \"ovnkube-node-dngcm\" (UID: \"9e4679dd-f870-41e9-a32b-360f5d2ee81b\") " pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.320581 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9e4679dd-f870-41e9-a32b-360f5d2ee81b-host-cni-bin\") pod \"ovnkube-node-dngcm\" (UID: \"9e4679dd-f870-41e9-a32b-360f5d2ee81b\") " pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.320607 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9e4679dd-f870-41e9-a32b-360f5d2ee81b-var-lib-openvswitch\") pod \"ovnkube-node-dngcm\" (UID: \"9e4679dd-f870-41e9-a32b-360f5d2ee81b\") " pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.320623 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/9e4679dd-f870-41e9-a32b-360f5d2ee81b-node-log\") pod \"ovnkube-node-dngcm\" (UID: \"9e4679dd-f870-41e9-a32b-360f5d2ee81b\") " pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.320643 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/9e4679dd-f870-41e9-a32b-360f5d2ee81b-ovnkube-script-lib\") pod \"ovnkube-node-dngcm\" (UID: \"9e4679dd-f870-41e9-a32b-360f5d2ee81b\") " pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.320877 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9e4679dd-f870-41e9-a32b-360f5d2ee81b-ovnkube-config\") pod \"ovnkube-node-dngcm\" (UID: \"9e4679dd-f870-41e9-a32b-360f5d2ee81b\") " pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.321083 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/9e4679dd-f870-41e9-a32b-360f5d2ee81b-systemd-units\") pod \"ovnkube-node-dngcm\" (UID: \"9e4679dd-f870-41e9-a32b-360f5d2ee81b\") " pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.321106 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9e4679dd-f870-41e9-a32b-360f5d2ee81b-host-cni-bin\") pod \"ovnkube-node-dngcm\" (UID: \"9e4679dd-f870-41e9-a32b-360f5d2ee81b\") " pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.321119 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/9e4679dd-f870-41e9-a32b-360f5d2ee81b-log-socket\") pod \"ovnkube-node-dngcm\" (UID: \"9e4679dd-f870-41e9-a32b-360f5d2ee81b\") " pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.321139 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/9e4679dd-f870-41e9-a32b-360f5d2ee81b-run-ovn\") pod \"ovnkube-node-dngcm\" (UID: \"9e4679dd-f870-41e9-a32b-360f5d2ee81b\") " pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.321146 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/9e4679dd-f870-41e9-a32b-360f5d2ee81b-host-cni-netd\") pod \"ovnkube-node-dngcm\" (UID: \"9e4679dd-f870-41e9-a32b-360f5d2ee81b\") " pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.321165 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9e4679dd-f870-41e9-a32b-360f5d2ee81b-host-run-netns\") pod \"ovnkube-node-dngcm\" (UID: \"9e4679dd-f870-41e9-a32b-360f5d2ee81b\") " pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.321170 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/9e4679dd-f870-41e9-a32b-360f5d2ee81b-host-kubelet\") pod \"ovnkube-node-dngcm\" (UID: \"9e4679dd-f870-41e9-a32b-360f5d2ee81b\") " pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.321192 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9e4679dd-f870-41e9-a32b-360f5d2ee81b-etc-openvswitch\") pod \"ovnkube-node-dngcm\" (UID: \"9e4679dd-f870-41e9-a32b-360f5d2ee81b\") " pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.321194 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9e4679dd-f870-41e9-a32b-360f5d2ee81b-var-lib-openvswitch\") pod \"ovnkube-node-dngcm\" (UID: \"9e4679dd-f870-41e9-a32b-360f5d2ee81b\") " pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.321229 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/9e4679dd-f870-41e9-a32b-360f5d2ee81b-node-log\") pod \"ovnkube-node-dngcm\" (UID: \"9e4679dd-f870-41e9-a32b-360f5d2ee81b\") " pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.321236 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/9e4679dd-f870-41e9-a32b-360f5d2ee81b-ovnkube-script-lib\") pod \"ovnkube-node-dngcm\" (UID: \"9e4679dd-f870-41e9-a32b-360f5d2ee81b\") " pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.321510 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9e4679dd-f870-41e9-a32b-360f5d2ee81b-env-overrides\") pod \"ovnkube-node-dngcm\" (UID: \"9e4679dd-f870-41e9-a32b-360f5d2ee81b\") " pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.326881 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9e4679dd-f870-41e9-a32b-360f5d2ee81b-ovn-node-metrics-cert\") pod \"ovnkube-node-dngcm\" (UID: \"9e4679dd-f870-41e9-a32b-360f5d2ee81b\") " pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.331236 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b087880e05039816dcab154643ac72c7e4a0a1124817c51dc2c639b62bbee0fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:14Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.337751 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4b7jr\" (UniqueName: \"kubernetes.io/projected/9e4679dd-f870-41e9-a32b-360f5d2ee81b-kube-api-access-4b7jr\") pod \"ovnkube-node-dngcm\" (UID: \"9e4679dd-f870-41e9-a32b-360f5d2ee81b\") " pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.345426 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:14Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.357839 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:14Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.366449 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-58lhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac6a19c4-f259-4dbd-a016-d6c2c849c6cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hzc4c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-58lhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:14Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.385251 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d169c8ba-000a-4218-8e3b-4ae53035b110\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-77rzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-77rzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rttx8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:14Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.400844 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a0bb885-2eb0-419c-8a7a-ef634cb3c144\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a89b49cbe08d927629564e607de13655ddf3d01ac4eb5d688b325210306c120c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06c5942805a498d9621a906088aa54538b4eed583d1b8d0e3ba2a6718378a36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b9f6f7fe44d5771a838d0b89c5bdea3ebb79023ae2675ee907ec2766b3d35a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://178b4340ee6904c60c814b99d83ee225161aeb23e438d8d02f856579a13ec4e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bb63bf30d6ecce9e71f0db0c16632bdfa6b33924b673378eb0c38ab229fa837\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T01:47:06Z\\\",\\\"message\\\":\\\"W1002 01:46:55.406027 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1002 01:46:55.406714 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759369615 cert, and key in /tmp/serving-cert-3795040473/serving-signer.crt, /tmp/serving-cert-3795040473/serving-signer.key\\\\nI1002 01:46:55.614670 1 observer_polling.go:159] Starting file observer\\\\nW1002 01:46:55.617662 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1002 01:46:55.618112 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 01:46:55.620347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3795040473/tls.crt::/tmp/serving-cert-3795040473/tls.key\\\\\\\"\\\\nF1002 01:47:06.224459 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90535f97ab2ed4626d47540ebd7baada837fd332140323bec0f49aa38d3c0009\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://592f89a326c9c1559193a71904b47ca27b1dbe1a7b30e03bc48924bb2f35f921\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://592f89a326c9c1559193a71904b47ca27b1dbe1a7b30e03bc48924bb2f35f921\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:14Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.413294 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:14Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.424324 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:14Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.435741 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-58lhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac6a19c4-f259-4dbd-a016-d6c2c849c6cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed76c0f968c063f7795d363af194a222487b647ceda234de0dabb39b25eafe0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hzc4c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-58lhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:14Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.445678 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d169c8ba-000a-4218-8e3b-4ae53035b110\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-77rzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-77rzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rttx8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:14Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.457346 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a0bb885-2eb0-419c-8a7a-ef634cb3c144\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a89b49cbe08d927629564e607de13655ddf3d01ac4eb5d688b325210306c120c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06c5942805a498d9621a906088aa54538b4eed583d1b8d0e3ba2a6718378a36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b9f6f7fe44d5771a838d0b89c5bdea3ebb79023ae2675ee907ec2766b3d35a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://178b4340ee6904c60c814b99d83ee225161aeb23e438d8d02f856579a13ec4e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bb63bf30d6ecce9e71f0db0c16632bdfa6b33924b673378eb0c38ab229fa837\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T01:47:06Z\\\",\\\"message\\\":\\\"W1002 01:46:55.406027 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1002 01:46:55.406714 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759369615 cert, and key in /tmp/serving-cert-3795040473/serving-signer.crt, /tmp/serving-cert-3795040473/serving-signer.key\\\\nI1002 01:46:55.614670 1 observer_polling.go:159] Starting file observer\\\\nW1002 01:46:55.617662 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1002 01:46:55.618112 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 01:46:55.620347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3795040473/tls.crt::/tmp/serving-cert-3795040473/tls.key\\\\\\\"\\\\nF1002 01:47:06.224459 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90535f97ab2ed4626d47540ebd7baada837fd332140323bec0f49aa38d3c0009\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://592f89a326c9c1559193a71904b47ca27b1dbe1a7b30e03bc48924bb2f35f921\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://592f89a326c9c1559193a71904b47ca27b1dbe1a7b30e03bc48924bb2f35f921\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:14Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.467414 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.468438 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b087880e05039816dcab154643ac72c7e4a0a1124817c51dc2c639b62bbee0fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:14Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.499109 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"506a599f-fb1c-4d14-be88-0e4833f3d624\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31c2f0193ffef54965b79a3569b0f6f4547e540bac0774155159a1c996c196c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://533e73e5f90b7258acbcdf018ee01323c118f9dc45035b25b065ba01cabecccf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c00db833a593ec40eb63a6b383455e19e960c2e98b9b66f76877ee993057e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21e2f462a0deb0d740185663b2014f4a1e57b4fdbeb82571264b002f928035bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:14Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:14 crc kubenswrapper[4885]: W1002 01:47:14.500576 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9e4679dd_f870_41e9_a32b_360f5d2ee81b.slice/crio-64efe9cde0ed394d9c75903a23e9b663a113d1fa0e7c74a8fb3c6217708805d4 WatchSource:0}: Error finding container 64efe9cde0ed394d9c75903a23e9b663a113d1fa0e7c74a8fb3c6217708805d4: Status 404 returned error can't find the container with id 64efe9cde0ed394d9c75903a23e9b663a113d1fa0e7c74a8fb3c6217708805d4 Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.519079 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e4679dd-f870-41e9-a32b-360f5d2ee81b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dngcm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:14Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.533391 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vhgbd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c63a7d20d39b65acfb5a9be4c28b21c72111b398bea76c238e15ddb4dd63e139\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7zwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vhgbd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:14Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.562316 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8070bc1c-e3de-4e9a-985e-40d1a9858056\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3e634b3727c517cc48b17c2cfb38f26460f53d1bf25d631d08ec9acb8d780e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4adcc537c23697a324d82af94c31059dd444245efdf7c978db58ede256e112ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edcd1d99644a778dfb6d0a17d2a4abd906400fe8209e611ebf8d7a9abe7ccf28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://97cb59adc2ddfa365ebc9dec649600dc90466ffe59134a6d15a9d3134a7c9b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d2808bc658ea80d68469ed16422d56ef19447f4c90b4af7b2597c64406f8a44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://83acbc20919f7c1958b6d492fadb203855df02761662425df9f91a0757fe2bf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83acbc20919f7c1958b6d492fadb203855df02761662425df9f91a0757fe2bf7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e86ecd03835efc9a253b6804bc61e6ac690e75b8d1f3cc77f5647730e586574\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e86ecd03835efc9a253b6804bc61e6ac690e75b8d1f3cc77f5647730e586574\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://48f95af51feaf014a18496bdfadd59d7f1e11167c5c8b677fa7037e48368ff2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48f95af51feaf014a18496bdfadd59d7f1e11167c5c8b677fa7037e48368ff2a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:14Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.574680 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:14Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.587004 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:14Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.599389 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74adcf912ccedb6b5d4f2803fd215e3e2ef62367597aa6da528354c4511ee536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7403ef968d040ac09d722b7557438bef23892d578992826f42b700b370d495d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:14Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:14 crc kubenswrapper[4885]: I1002 01:47:14.611952 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2ll9b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52b39712-5fa1-4fe0-814b-b170a6e3938b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2ll9b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:14Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:15 crc kubenswrapper[4885]: I1002 01:47:15.246247 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-vhgbd" event={"ID":"ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f","Type":"ContainerStarted","Data":"c63a7d20d39b65acfb5a9be4c28b21c72111b398bea76c238e15ddb4dd63e139"} Oct 02 01:47:15 crc kubenswrapper[4885]: I1002 01:47:15.247927 4885 generic.go:334] "Generic (PLEG): container finished" podID="52b39712-5fa1-4fe0-814b-b170a6e3938b" containerID="8c979b91ba748aa4647826a8926337662ca929c54090fb62285572819ec2fbf9" exitCode=0 Oct 02 01:47:15 crc kubenswrapper[4885]: I1002 01:47:15.247975 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2ll9b" event={"ID":"52b39712-5fa1-4fe0-814b-b170a6e3938b","Type":"ContainerDied","Data":"8c979b91ba748aa4647826a8926337662ca929c54090fb62285572819ec2fbf9"} Oct 02 01:47:15 crc kubenswrapper[4885]: I1002 01:47:15.250013 4885 generic.go:334] "Generic (PLEG): container finished" podID="9e4679dd-f870-41e9-a32b-360f5d2ee81b" containerID="0125e907d061a97286026ee823c4eed303734ed03ab4864f742592b23a1aa2f8" exitCode=0 Oct 02 01:47:15 crc kubenswrapper[4885]: I1002 01:47:15.250064 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" event={"ID":"9e4679dd-f870-41e9-a32b-360f5d2ee81b","Type":"ContainerDied","Data":"0125e907d061a97286026ee823c4eed303734ed03ab4864f742592b23a1aa2f8"} Oct 02 01:47:15 crc kubenswrapper[4885]: I1002 01:47:15.250089 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" event={"ID":"9e4679dd-f870-41e9-a32b-360f5d2ee81b","Type":"ContainerStarted","Data":"64efe9cde0ed394d9c75903a23e9b663a113d1fa0e7c74a8fb3c6217708805d4"} Oct 02 01:47:15 crc kubenswrapper[4885]: I1002 01:47:15.253022 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" event={"ID":"d169c8ba-000a-4218-8e3b-4ae53035b110","Type":"ContainerStarted","Data":"09536e182a46e77318c10812d1777e551f834907c006cab5c726dc66f5bce88a"} Oct 02 01:47:15 crc kubenswrapper[4885]: I1002 01:47:15.253072 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" event={"ID":"d169c8ba-000a-4218-8e3b-4ae53035b110","Type":"ContainerStarted","Data":"61e436b7876107ce5ea9e668c0bbc367be6e663fc804fa3a65b51adf8be8d51f"} Oct 02 01:47:15 crc kubenswrapper[4885]: E1002 01:47:15.261271 4885 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"etcd-crc\" already exists" pod="openshift-etcd/etcd-crc" Oct 02 01:47:15 crc kubenswrapper[4885]: I1002 01:47:15.265125 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:15Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:15 crc kubenswrapper[4885]: I1002 01:47:15.296650 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74adcf912ccedb6b5d4f2803fd215e3e2ef62367597aa6da528354c4511ee536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7403ef968d040ac09d722b7557438bef23892d578992826f42b700b370d495d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:15Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:15 crc kubenswrapper[4885]: I1002 01:47:15.353230 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2ll9b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52b39712-5fa1-4fe0-814b-b170a6e3938b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c979b91ba748aa4647826a8926337662ca929c54090fb62285572819ec2fbf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c979b91ba748aa4647826a8926337662ca929c54090fb62285572819ec2fbf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2ll9b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:15Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:15 crc kubenswrapper[4885]: I1002 01:47:15.364554 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:15Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:15 crc kubenswrapper[4885]: I1002 01:47:15.377390 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:15Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:15 crc kubenswrapper[4885]: I1002 01:47:15.386921 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-58lhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac6a19c4-f259-4dbd-a016-d6c2c849c6cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed76c0f968c063f7795d363af194a222487b647ceda234de0dabb39b25eafe0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hzc4c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-58lhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:15Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:15 crc kubenswrapper[4885]: I1002 01:47:15.404921 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d169c8ba-000a-4218-8e3b-4ae53035b110\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-77rzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-77rzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rttx8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:15Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:15 crc kubenswrapper[4885]: I1002 01:47:15.422252 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a0bb885-2eb0-419c-8a7a-ef634cb3c144\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a89b49cbe08d927629564e607de13655ddf3d01ac4eb5d688b325210306c120c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06c5942805a498d9621a906088aa54538b4eed583d1b8d0e3ba2a6718378a36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b9f6f7fe44d5771a838d0b89c5bdea3ebb79023ae2675ee907ec2766b3d35a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://178b4340ee6904c60c814b99d83ee225161aeb23e438d8d02f856579a13ec4e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bb63bf30d6ecce9e71f0db0c16632bdfa6b33924b673378eb0c38ab229fa837\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T01:47:06Z\\\",\\\"message\\\":\\\"W1002 01:46:55.406027 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1002 01:46:55.406714 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759369615 cert, and key in /tmp/serving-cert-3795040473/serving-signer.crt, /tmp/serving-cert-3795040473/serving-signer.key\\\\nI1002 01:46:55.614670 1 observer_polling.go:159] Starting file observer\\\\nW1002 01:46:55.617662 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1002 01:46:55.618112 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 01:46:55.620347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3795040473/tls.crt::/tmp/serving-cert-3795040473/tls.key\\\\\\\"\\\\nF1002 01:47:06.224459 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90535f97ab2ed4626d47540ebd7baada837fd332140323bec0f49aa38d3c0009\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://592f89a326c9c1559193a71904b47ca27b1dbe1a7b30e03bc48924bb2f35f921\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://592f89a326c9c1559193a71904b47ca27b1dbe1a7b30e03bc48924bb2f35f921\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:15Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:15 crc kubenswrapper[4885]: I1002 01:47:15.433795 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b087880e05039816dcab154643ac72c7e4a0a1124817c51dc2c639b62bbee0fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:15Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:15 crc kubenswrapper[4885]: I1002 01:47:15.447633 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"506a599f-fb1c-4d14-be88-0e4833f3d624\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31c2f0193ffef54965b79a3569b0f6f4547e540bac0774155159a1c996c196c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://533e73e5f90b7258acbcdf018ee01323c118f9dc45035b25b065ba01cabecccf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c00db833a593ec40eb63a6b383455e19e960c2e98b9b66f76877ee993057e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21e2f462a0deb0d740185663b2014f4a1e57b4fdbeb82571264b002f928035bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:15Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:15 crc kubenswrapper[4885]: I1002 01:47:15.464993 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e4679dd-f870-41e9-a32b-360f5d2ee81b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dngcm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:15Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:15 crc kubenswrapper[4885]: I1002 01:47:15.482740 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vhgbd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c63a7d20d39b65acfb5a9be4c28b21c72111b398bea76c238e15ddb4dd63e139\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7zwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vhgbd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:15Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:15 crc kubenswrapper[4885]: I1002 01:47:15.506008 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8070bc1c-e3de-4e9a-985e-40d1a9858056\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3e634b3727c517cc48b17c2cfb38f26460f53d1bf25d631d08ec9acb8d780e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4adcc537c23697a324d82af94c31059dd444245efdf7c978db58ede256e112ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edcd1d99644a778dfb6d0a17d2a4abd906400fe8209e611ebf8d7a9abe7ccf28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://97cb59adc2ddfa365ebc9dec649600dc90466ffe59134a6d15a9d3134a7c9b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d2808bc658ea80d68469ed16422d56ef19447f4c90b4af7b2597c64406f8a44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://83acbc20919f7c1958b6d492fadb203855df02761662425df9f91a0757fe2bf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83acbc20919f7c1958b6d492fadb203855df02761662425df9f91a0757fe2bf7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e86ecd03835efc9a253b6804bc61e6ac690e75b8d1f3cc77f5647730e586574\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e86ecd03835efc9a253b6804bc61e6ac690e75b8d1f3cc77f5647730e586574\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://48f95af51feaf014a18496bdfadd59d7f1e11167c5c8b677fa7037e48368ff2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48f95af51feaf014a18496bdfadd59d7f1e11167c5c8b677fa7037e48368ff2a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:15Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:15 crc kubenswrapper[4885]: I1002 01:47:15.518430 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:15Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:15 crc kubenswrapper[4885]: I1002 01:47:15.529710 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-58lhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac6a19c4-f259-4dbd-a016-d6c2c849c6cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed76c0f968c063f7795d363af194a222487b647ceda234de0dabb39b25eafe0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hzc4c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-58lhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:15Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:15 crc kubenswrapper[4885]: I1002 01:47:15.543764 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d169c8ba-000a-4218-8e3b-4ae53035b110\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61e436b7876107ce5ea9e668c0bbc367be6e663fc804fa3a65b51adf8be8d51f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-77rzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09536e182a46e77318c10812d1777e551f834907c006cab5c726dc66f5bce88a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-77rzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rttx8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:15Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:15 crc kubenswrapper[4885]: I1002 01:47:15.564466 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a0bb885-2eb0-419c-8a7a-ef634cb3c144\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a89b49cbe08d927629564e607de13655ddf3d01ac4eb5d688b325210306c120c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06c5942805a498d9621a906088aa54538b4eed583d1b8d0e3ba2a6718378a36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b9f6f7fe44d5771a838d0b89c5bdea3ebb79023ae2675ee907ec2766b3d35a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://178b4340ee6904c60c814b99d83ee225161aeb23e438d8d02f856579a13ec4e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bb63bf30d6ecce9e71f0db0c16632bdfa6b33924b673378eb0c38ab229fa837\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T01:47:06Z\\\",\\\"message\\\":\\\"W1002 01:46:55.406027 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1002 01:46:55.406714 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759369615 cert, and key in /tmp/serving-cert-3795040473/serving-signer.crt, /tmp/serving-cert-3795040473/serving-signer.key\\\\nI1002 01:46:55.614670 1 observer_polling.go:159] Starting file observer\\\\nW1002 01:46:55.617662 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1002 01:46:55.618112 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 01:46:55.620347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3795040473/tls.crt::/tmp/serving-cert-3795040473/tls.key\\\\\\\"\\\\nF1002 01:47:06.224459 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90535f97ab2ed4626d47540ebd7baada837fd332140323bec0f49aa38d3c0009\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://592f89a326c9c1559193a71904b47ca27b1dbe1a7b30e03bc48924bb2f35f921\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://592f89a326c9c1559193a71904b47ca27b1dbe1a7b30e03bc48924bb2f35f921\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:15Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:15 crc kubenswrapper[4885]: I1002 01:47:15.580079 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b087880e05039816dcab154643ac72c7e4a0a1124817c51dc2c639b62bbee0fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:15Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:15 crc kubenswrapper[4885]: I1002 01:47:15.597082 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:15Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:15 crc kubenswrapper[4885]: I1002 01:47:15.604843 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-jgdvt"] Oct 02 01:47:15 crc kubenswrapper[4885]: I1002 01:47:15.605407 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-jgdvt" Oct 02 01:47:15 crc kubenswrapper[4885]: I1002 01:47:15.607629 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Oct 02 01:47:15 crc kubenswrapper[4885]: I1002 01:47:15.607880 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Oct 02 01:47:15 crc kubenswrapper[4885]: I1002 01:47:15.608018 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Oct 02 01:47:15 crc kubenswrapper[4885]: I1002 01:47:15.608273 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Oct 02 01:47:15 crc kubenswrapper[4885]: I1002 01:47:15.612636 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:15Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:15 crc kubenswrapper[4885]: I1002 01:47:15.633818 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"506a599f-fb1c-4d14-be88-0e4833f3d624\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31c2f0193ffef54965b79a3569b0f6f4547e540bac0774155159a1c996c196c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://533e73e5f90b7258acbcdf018ee01323c118f9dc45035b25b065ba01cabecccf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c00db833a593ec40eb63a6b383455e19e960c2e98b9b66f76877ee993057e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21e2f462a0deb0d740185663b2014f4a1e57b4fdbeb82571264b002f928035bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:15Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:15 crc kubenswrapper[4885]: I1002 01:47:15.633873 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:47:15 crc kubenswrapper[4885]: E1002 01:47:15.633967 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:47:19.633941107 +0000 UTC m=+28.445688546 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:47:15 crc kubenswrapper[4885]: I1002 01:47:15.634222 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:47:15 crc kubenswrapper[4885]: I1002 01:47:15.634287 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/81733e8d-67b0-46fa-8a41-08b40645bd3e-serviceca\") pod \"node-ca-jgdvt\" (UID: \"81733e8d-67b0-46fa-8a41-08b40645bd3e\") " pod="openshift-image-registry/node-ca-jgdvt" Oct 02 01:47:15 crc kubenswrapper[4885]: I1002 01:47:15.634325 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4cwbl\" (UniqueName: \"kubernetes.io/projected/81733e8d-67b0-46fa-8a41-08b40645bd3e-kube-api-access-4cwbl\") pod \"node-ca-jgdvt\" (UID: \"81733e8d-67b0-46fa-8a41-08b40645bd3e\") " pod="openshift-image-registry/node-ca-jgdvt" Oct 02 01:47:15 crc kubenswrapper[4885]: I1002 01:47:15.634360 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/81733e8d-67b0-46fa-8a41-08b40645bd3e-host\") pod \"node-ca-jgdvt\" (UID: \"81733e8d-67b0-46fa-8a41-08b40645bd3e\") " pod="openshift-image-registry/node-ca-jgdvt" Oct 02 01:47:15 crc kubenswrapper[4885]: E1002 01:47:15.634384 4885 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 01:47:15 crc kubenswrapper[4885]: E1002 01:47:15.634465 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 01:47:19.634443651 +0000 UTC m=+28.446191060 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 01:47:15 crc kubenswrapper[4885]: I1002 01:47:15.655000 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e4679dd-f870-41e9-a32b-360f5d2ee81b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0125e907d061a97286026ee823c4eed303734ed03ab4864f742592b23a1aa2f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0125e907d061a97286026ee823c4eed303734ed03ab4864f742592b23a1aa2f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dngcm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:15Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:15 crc kubenswrapper[4885]: I1002 01:47:15.676343 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8070bc1c-e3de-4e9a-985e-40d1a9858056\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3e634b3727c517cc48b17c2cfb38f26460f53d1bf25d631d08ec9acb8d780e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4adcc537c23697a324d82af94c31059dd444245efdf7c978db58ede256e112ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edcd1d99644a778dfb6d0a17d2a4abd906400fe8209e611ebf8d7a9abe7ccf28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://97cb59adc2ddfa365ebc9dec649600dc90466ffe59134a6d15a9d3134a7c9b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d2808bc658ea80d68469ed16422d56ef19447f4c90b4af7b2597c64406f8a44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://83acbc20919f7c1958b6d492fadb203855df02761662425df9f91a0757fe2bf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83acbc20919f7c1958b6d492fadb203855df02761662425df9f91a0757fe2bf7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e86ecd03835efc9a253b6804bc61e6ac690e75b8d1f3cc77f5647730e586574\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e86ecd03835efc9a253b6804bc61e6ac690e75b8d1f3cc77f5647730e586574\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://48f95af51feaf014a18496bdfadd59d7f1e11167c5c8b677fa7037e48368ff2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48f95af51feaf014a18496bdfadd59d7f1e11167c5c8b677fa7037e48368ff2a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:15Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:15 crc kubenswrapper[4885]: I1002 01:47:15.690756 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:15Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:15 crc kubenswrapper[4885]: I1002 01:47:15.709413 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vhgbd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c63a7d20d39b65acfb5a9be4c28b21c72111b398bea76c238e15ddb4dd63e139\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7zwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vhgbd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:15Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:15 crc kubenswrapper[4885]: I1002 01:47:15.769200 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/81733e8d-67b0-46fa-8a41-08b40645bd3e-serviceca\") pod \"node-ca-jgdvt\" (UID: \"81733e8d-67b0-46fa-8a41-08b40645bd3e\") " pod="openshift-image-registry/node-ca-jgdvt" Oct 02 01:47:15 crc kubenswrapper[4885]: I1002 01:47:15.769253 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4cwbl\" (UniqueName: \"kubernetes.io/projected/81733e8d-67b0-46fa-8a41-08b40645bd3e-kube-api-access-4cwbl\") pod \"node-ca-jgdvt\" (UID: \"81733e8d-67b0-46fa-8a41-08b40645bd3e\") " pod="openshift-image-registry/node-ca-jgdvt" Oct 02 01:47:15 crc kubenswrapper[4885]: I1002 01:47:15.769289 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/81733e8d-67b0-46fa-8a41-08b40645bd3e-host\") pod \"node-ca-jgdvt\" (UID: \"81733e8d-67b0-46fa-8a41-08b40645bd3e\") " pod="openshift-image-registry/node-ca-jgdvt" Oct 02 01:47:15 crc kubenswrapper[4885]: I1002 01:47:15.769307 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:47:15 crc kubenswrapper[4885]: I1002 01:47:15.769455 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/81733e8d-67b0-46fa-8a41-08b40645bd3e-host\") pod \"node-ca-jgdvt\" (UID: \"81733e8d-67b0-46fa-8a41-08b40645bd3e\") " pod="openshift-image-registry/node-ca-jgdvt" Oct 02 01:47:15 crc kubenswrapper[4885]: I1002 01:47:15.770176 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/81733e8d-67b0-46fa-8a41-08b40645bd3e-serviceca\") pod \"node-ca-jgdvt\" (UID: \"81733e8d-67b0-46fa-8a41-08b40645bd3e\") " pod="openshift-image-registry/node-ca-jgdvt" Oct 02 01:47:15 crc kubenswrapper[4885]: I1002 01:47:15.770312 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:47:15 crc kubenswrapper[4885]: I1002 01:47:15.770341 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:47:15 crc kubenswrapper[4885]: E1002 01:47:15.770417 4885 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 01:47:15 crc kubenswrapper[4885]: E1002 01:47:15.770423 4885 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 01:47:15 crc kubenswrapper[4885]: E1002 01:47:15.770486 4885 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 01:47:15 crc kubenswrapper[4885]: E1002 01:47:15.770496 4885 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 01:47:15 crc kubenswrapper[4885]: E1002 01:47:15.770432 4885 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 01:47:15 crc kubenswrapper[4885]: E1002 01:47:15.770647 4885 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 01:47:15 crc kubenswrapper[4885]: E1002 01:47:15.770686 4885 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 01:47:15 crc kubenswrapper[4885]: E1002 01:47:15.770489 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 01:47:19.770471065 +0000 UTC m=+28.582218464 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 01:47:15 crc kubenswrapper[4885]: E1002 01:47:15.770721 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-02 01:47:19.770710022 +0000 UTC m=+28.582457421 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 01:47:15 crc kubenswrapper[4885]: E1002 01:47:15.770750 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-02 01:47:19.770726912 +0000 UTC m=+28.582474311 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 01:47:15 crc kubenswrapper[4885]: I1002 01:47:15.774170 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:15Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:15 crc kubenswrapper[4885]: I1002 01:47:15.794613 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4cwbl\" (UniqueName: \"kubernetes.io/projected/81733e8d-67b0-46fa-8a41-08b40645bd3e-kube-api-access-4cwbl\") pod \"node-ca-jgdvt\" (UID: \"81733e8d-67b0-46fa-8a41-08b40645bd3e\") " pod="openshift-image-registry/node-ca-jgdvt" Oct 02 01:47:15 crc kubenswrapper[4885]: I1002 01:47:15.797916 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74adcf912ccedb6b5d4f2803fd215e3e2ef62367597aa6da528354c4511ee536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7403ef968d040ac09d722b7557438bef23892d578992826f42b700b370d495d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:15Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:15 crc kubenswrapper[4885]: I1002 01:47:15.813873 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2ll9b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52b39712-5fa1-4fe0-814b-b170a6e3938b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c979b91ba748aa4647826a8926337662ca929c54090fb62285572819ec2fbf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c979b91ba748aa4647826a8926337662ca929c54090fb62285572819ec2fbf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2ll9b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:15Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:15 crc kubenswrapper[4885]: I1002 01:47:15.825477 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-58lhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac6a19c4-f259-4dbd-a016-d6c2c849c6cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed76c0f968c063f7795d363af194a222487b647ceda234de0dabb39b25eafe0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hzc4c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-58lhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:15Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:15 crc kubenswrapper[4885]: I1002 01:47:15.859231 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d169c8ba-000a-4218-8e3b-4ae53035b110\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61e436b7876107ce5ea9e668c0bbc367be6e663fc804fa3a65b51adf8be8d51f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-77rzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09536e182a46e77318c10812d1777e551f834907c006cab5c726dc66f5bce88a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-77rzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rttx8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:15Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:15 crc kubenswrapper[4885]: I1002 01:47:15.898756 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jgdvt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81733e8d-67b0-46fa-8a41-08b40645bd3e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cwbl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jgdvt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:15Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:15 crc kubenswrapper[4885]: I1002 01:47:15.946949 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a0bb885-2eb0-419c-8a7a-ef634cb3c144\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a89b49cbe08d927629564e607de13655ddf3d01ac4eb5d688b325210306c120c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06c5942805a498d9621a906088aa54538b4eed583d1b8d0e3ba2a6718378a36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b9f6f7fe44d5771a838d0b89c5bdea3ebb79023ae2675ee907ec2766b3d35a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://178b4340ee6904c60c814b99d83ee225161aeb23e438d8d02f856579a13ec4e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bb63bf30d6ecce9e71f0db0c16632bdfa6b33924b673378eb0c38ab229fa837\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T01:47:06Z\\\",\\\"message\\\":\\\"W1002 01:46:55.406027 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1002 01:46:55.406714 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759369615 cert, and key in /tmp/serving-cert-3795040473/serving-signer.crt, /tmp/serving-cert-3795040473/serving-signer.key\\\\nI1002 01:46:55.614670 1 observer_polling.go:159] Starting file observer\\\\nW1002 01:46:55.617662 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1002 01:46:55.618112 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 01:46:55.620347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3795040473/tls.crt::/tmp/serving-cert-3795040473/tls.key\\\\\\\"\\\\nF1002 01:47:06.224459 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90535f97ab2ed4626d47540ebd7baada837fd332140323bec0f49aa38d3c0009\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://592f89a326c9c1559193a71904b47ca27b1dbe1a7b30e03bc48924bb2f35f921\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://592f89a326c9c1559193a71904b47ca27b1dbe1a7b30e03bc48924bb2f35f921\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:15Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:15 crc kubenswrapper[4885]: I1002 01:47:15.954754 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-jgdvt" Oct 02 01:47:15 crc kubenswrapper[4885]: I1002 01:47:15.985491 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b087880e05039816dcab154643ac72c7e4a0a1124817c51dc2c639b62bbee0fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:15Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:16 crc kubenswrapper[4885]: I1002 01:47:16.018455 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:16Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:16 crc kubenswrapper[4885]: I1002 01:47:16.045923 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:47:16 crc kubenswrapper[4885]: I1002 01:47:16.046012 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:47:16 crc kubenswrapper[4885]: E1002 01:47:16.046122 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:47:16 crc kubenswrapper[4885]: I1002 01:47:16.046131 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:47:16 crc kubenswrapper[4885]: E1002 01:47:16.046203 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:47:16 crc kubenswrapper[4885]: E1002 01:47:16.046290 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:47:16 crc kubenswrapper[4885]: I1002 01:47:16.059329 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:16Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:16 crc kubenswrapper[4885]: I1002 01:47:16.096744 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"506a599f-fb1c-4d14-be88-0e4833f3d624\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31c2f0193ffef54965b79a3569b0f6f4547e540bac0774155159a1c996c196c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://533e73e5f90b7258acbcdf018ee01323c118f9dc45035b25b065ba01cabecccf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c00db833a593ec40eb63a6b383455e19e960c2e98b9b66f76877ee993057e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21e2f462a0deb0d740185663b2014f4a1e57b4fdbeb82571264b002f928035bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:16Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:16 crc kubenswrapper[4885]: I1002 01:47:16.143526 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e4679dd-f870-41e9-a32b-360f5d2ee81b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0125e907d061a97286026ee823c4eed303734ed03ab4864f742592b23a1aa2f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0125e907d061a97286026ee823c4eed303734ed03ab4864f742592b23a1aa2f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dngcm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:16Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:16 crc kubenswrapper[4885]: I1002 01:47:16.186630 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8070bc1c-e3de-4e9a-985e-40d1a9858056\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3e634b3727c517cc48b17c2cfb38f26460f53d1bf25d631d08ec9acb8d780e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4adcc537c23697a324d82af94c31059dd444245efdf7c978db58ede256e112ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edcd1d99644a778dfb6d0a17d2a4abd906400fe8209e611ebf8d7a9abe7ccf28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://97cb59adc2ddfa365ebc9dec649600dc90466ffe59134a6d15a9d3134a7c9b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d2808bc658ea80d68469ed16422d56ef19447f4c90b4af7b2597c64406f8a44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://83acbc20919f7c1958b6d492fadb203855df02761662425df9f91a0757fe2bf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83acbc20919f7c1958b6d492fadb203855df02761662425df9f91a0757fe2bf7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e86ecd03835efc9a253b6804bc61e6ac690e75b8d1f3cc77f5647730e586574\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e86ecd03835efc9a253b6804bc61e6ac690e75b8d1f3cc77f5647730e586574\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://48f95af51feaf014a18496bdfadd59d7f1e11167c5c8b677fa7037e48368ff2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48f95af51feaf014a18496bdfadd59d7f1e11167c5c8b677fa7037e48368ff2a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:16Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:16 crc kubenswrapper[4885]: I1002 01:47:16.217733 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:16Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:16 crc kubenswrapper[4885]: I1002 01:47:16.257277 4885 generic.go:334] "Generic (PLEG): container finished" podID="52b39712-5fa1-4fe0-814b-b170a6e3938b" containerID="97d9d693e8e60ee9c183bb44461e698e8b3cfd20a516605aa8754c9205d2bdc2" exitCode=0 Oct 02 01:47:16 crc kubenswrapper[4885]: I1002 01:47:16.257349 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2ll9b" event={"ID":"52b39712-5fa1-4fe0-814b-b170a6e3938b","Type":"ContainerDied","Data":"97d9d693e8e60ee9c183bb44461e698e8b3cfd20a516605aa8754c9205d2bdc2"} Oct 02 01:47:16 crc kubenswrapper[4885]: I1002 01:47:16.263200 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vhgbd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c63a7d20d39b65acfb5a9be4c28b21c72111b398bea76c238e15ddb4dd63e139\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7zwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vhgbd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:16Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:16 crc kubenswrapper[4885]: I1002 01:47:16.263767 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" event={"ID":"9e4679dd-f870-41e9-a32b-360f5d2ee81b","Type":"ContainerStarted","Data":"a6fa72b030d309656b73121e4ac2a3924a4e449aad66683dd9de5ee5f22d5a4e"} Oct 02 01:47:16 crc kubenswrapper[4885]: I1002 01:47:16.263823 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" event={"ID":"9e4679dd-f870-41e9-a32b-360f5d2ee81b","Type":"ContainerStarted","Data":"a7642033b0b5995d332ccf7992d6d7e6a5e9c9369ae2d1071ccb2f536938c8f4"} Oct 02 01:47:16 crc kubenswrapper[4885]: I1002 01:47:16.263857 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" event={"ID":"9e4679dd-f870-41e9-a32b-360f5d2ee81b","Type":"ContainerStarted","Data":"4310909ad9c7b609ed6814e04a525a6afae747775598e7275080413dec8b23a3"} Oct 02 01:47:16 crc kubenswrapper[4885]: I1002 01:47:16.263875 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" event={"ID":"9e4679dd-f870-41e9-a32b-360f5d2ee81b","Type":"ContainerStarted","Data":"ef403e519d197fb37b3e6b76a14a5128ba3f30a7a5e8bf0a420cf6a39b89575f"} Oct 02 01:47:16 crc kubenswrapper[4885]: I1002 01:47:16.263892 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" event={"ID":"9e4679dd-f870-41e9-a32b-360f5d2ee81b","Type":"ContainerStarted","Data":"8482ba457683ffa03100ec1f07ea9d3f918be5e490b5ba05c2ba965d03033191"} Oct 02 01:47:16 crc kubenswrapper[4885]: I1002 01:47:16.265710 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-jgdvt" event={"ID":"81733e8d-67b0-46fa-8a41-08b40645bd3e","Type":"ContainerStarted","Data":"dcef60c816ca1ba9a298da658e84c4d717bea3d352c4c2f807ebd8e56327b384"} Oct 02 01:47:16 crc kubenswrapper[4885]: I1002 01:47:16.265753 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-jgdvt" event={"ID":"81733e8d-67b0-46fa-8a41-08b40645bd3e","Type":"ContainerStarted","Data":"80dcddc7f9ca0fc2e5321728de765d4844266540d1f0f399024e99e77787b7c1"} Oct 02 01:47:16 crc kubenswrapper[4885]: I1002 01:47:16.273776 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"b5c1f95f188569c9aa47e51f2916939fa0b65a5cdf4c16b798612d41f3c8411a"} Oct 02 01:47:16 crc kubenswrapper[4885]: I1002 01:47:16.299483 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:16Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:16 crc kubenswrapper[4885]: I1002 01:47:16.342046 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74adcf912ccedb6b5d4f2803fd215e3e2ef62367597aa6da528354c4511ee536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7403ef968d040ac09d722b7557438bef23892d578992826f42b700b370d495d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:16Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:16 crc kubenswrapper[4885]: I1002 01:47:16.383421 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2ll9b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52b39712-5fa1-4fe0-814b-b170a6e3938b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c979b91ba748aa4647826a8926337662ca929c54090fb62285572819ec2fbf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c979b91ba748aa4647826a8926337662ca929c54090fb62285572819ec2fbf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2ll9b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:16Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:16 crc kubenswrapper[4885]: I1002 01:47:16.419460 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a0bb885-2eb0-419c-8a7a-ef634cb3c144\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a89b49cbe08d927629564e607de13655ddf3d01ac4eb5d688b325210306c120c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06c5942805a498d9621a906088aa54538b4eed583d1b8d0e3ba2a6718378a36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b9f6f7fe44d5771a838d0b89c5bdea3ebb79023ae2675ee907ec2766b3d35a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://178b4340ee6904c60c814b99d83ee225161aeb23e438d8d02f856579a13ec4e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bb63bf30d6ecce9e71f0db0c16632bdfa6b33924b673378eb0c38ab229fa837\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T01:47:06Z\\\",\\\"message\\\":\\\"W1002 01:46:55.406027 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1002 01:46:55.406714 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759369615 cert, and key in /tmp/serving-cert-3795040473/serving-signer.crt, /tmp/serving-cert-3795040473/serving-signer.key\\\\nI1002 01:46:55.614670 1 observer_polling.go:159] Starting file observer\\\\nW1002 01:46:55.617662 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1002 01:46:55.618112 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 01:46:55.620347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3795040473/tls.crt::/tmp/serving-cert-3795040473/tls.key\\\\\\\"\\\\nF1002 01:47:06.224459 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90535f97ab2ed4626d47540ebd7baada837fd332140323bec0f49aa38d3c0009\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://592f89a326c9c1559193a71904b47ca27b1dbe1a7b30e03bc48924bb2f35f921\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://592f89a326c9c1559193a71904b47ca27b1dbe1a7b30e03bc48924bb2f35f921\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:16Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:16 crc kubenswrapper[4885]: I1002 01:47:16.459036 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b087880e05039816dcab154643ac72c7e4a0a1124817c51dc2c639b62bbee0fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:16Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:16 crc kubenswrapper[4885]: I1002 01:47:16.499712 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:16Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:16 crc kubenswrapper[4885]: I1002 01:47:16.540296 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:16Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:16 crc kubenswrapper[4885]: I1002 01:47:16.577796 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-58lhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac6a19c4-f259-4dbd-a016-d6c2c849c6cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed76c0f968c063f7795d363af194a222487b647ceda234de0dabb39b25eafe0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hzc4c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-58lhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:16Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:16 crc kubenswrapper[4885]: I1002 01:47:16.615627 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d169c8ba-000a-4218-8e3b-4ae53035b110\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61e436b7876107ce5ea9e668c0bbc367be6e663fc804fa3a65b51adf8be8d51f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-77rzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09536e182a46e77318c10812d1777e551f834907c006cab5c726dc66f5bce88a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-77rzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rttx8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:16Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:16 crc kubenswrapper[4885]: I1002 01:47:16.685798 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jgdvt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81733e8d-67b0-46fa-8a41-08b40645bd3e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcef60c816ca1ba9a298da658e84c4d717bea3d352c4c2f807ebd8e56327b384\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cwbl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jgdvt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:16Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:16 crc kubenswrapper[4885]: I1002 01:47:16.715176 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"506a599f-fb1c-4d14-be88-0e4833f3d624\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31c2f0193ffef54965b79a3569b0f6f4547e540bac0774155159a1c996c196c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://533e73e5f90b7258acbcdf018ee01323c118f9dc45035b25b065ba01cabecccf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c00db833a593ec40eb63a6b383455e19e960c2e98b9b66f76877ee993057e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21e2f462a0deb0d740185663b2014f4a1e57b4fdbeb82571264b002f928035bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:16Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:16 crc kubenswrapper[4885]: I1002 01:47:16.746621 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e4679dd-f870-41e9-a32b-360f5d2ee81b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0125e907d061a97286026ee823c4eed303734ed03ab4864f742592b23a1aa2f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0125e907d061a97286026ee823c4eed303734ed03ab4864f742592b23a1aa2f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dngcm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:16Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:16 crc kubenswrapper[4885]: I1002 01:47:16.789968 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8070bc1c-e3de-4e9a-985e-40d1a9858056\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3e634b3727c517cc48b17c2cfb38f26460f53d1bf25d631d08ec9acb8d780e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4adcc537c23697a324d82af94c31059dd444245efdf7c978db58ede256e112ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edcd1d99644a778dfb6d0a17d2a4abd906400fe8209e611ebf8d7a9abe7ccf28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://97cb59adc2ddfa365ebc9dec649600dc90466ffe59134a6d15a9d3134a7c9b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d2808bc658ea80d68469ed16422d56ef19447f4c90b4af7b2597c64406f8a44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://83acbc20919f7c1958b6d492fadb203855df02761662425df9f91a0757fe2bf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83acbc20919f7c1958b6d492fadb203855df02761662425df9f91a0757fe2bf7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e86ecd03835efc9a253b6804bc61e6ac690e75b8d1f3cc77f5647730e586574\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e86ecd03835efc9a253b6804bc61e6ac690e75b8d1f3cc77f5647730e586574\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://48f95af51feaf014a18496bdfadd59d7f1e11167c5c8b677fa7037e48368ff2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48f95af51feaf014a18496bdfadd59d7f1e11167c5c8b677fa7037e48368ff2a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:16Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:16 crc kubenswrapper[4885]: I1002 01:47:16.816853 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5c1f95f188569c9aa47e51f2916939fa0b65a5cdf4c16b798612d41f3c8411a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:16Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:16 crc kubenswrapper[4885]: I1002 01:47:16.863042 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vhgbd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c63a7d20d39b65acfb5a9be4c28b21c72111b398bea76c238e15ddb4dd63e139\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7zwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vhgbd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:16Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:16 crc kubenswrapper[4885]: I1002 01:47:16.897109 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:16Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:16 crc kubenswrapper[4885]: I1002 01:47:16.943507 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74adcf912ccedb6b5d4f2803fd215e3e2ef62367597aa6da528354c4511ee536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7403ef968d040ac09d722b7557438bef23892d578992826f42b700b370d495d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:16Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:16 crc kubenswrapper[4885]: I1002 01:47:16.979404 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2ll9b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52b39712-5fa1-4fe0-814b-b170a6e3938b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c979b91ba748aa4647826a8926337662ca929c54090fb62285572819ec2fbf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c979b91ba748aa4647826a8926337662ca929c54090fb62285572819ec2fbf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97d9d693e8e60ee9c183bb44461e698e8b3cfd20a516605aa8754c9205d2bdc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://97d9d693e8e60ee9c183bb44461e698e8b3cfd20a516605aa8754c9205d2bdc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2ll9b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:16Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:17 crc kubenswrapper[4885]: I1002 01:47:17.282684 4885 generic.go:334] "Generic (PLEG): container finished" podID="52b39712-5fa1-4fe0-814b-b170a6e3938b" containerID="5c44252676359c671d65c9588b05572d76c99cf4c90666f0ac089b2f778450a5" exitCode=0 Oct 02 01:47:17 crc kubenswrapper[4885]: I1002 01:47:17.282798 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2ll9b" event={"ID":"52b39712-5fa1-4fe0-814b-b170a6e3938b","Type":"ContainerDied","Data":"5c44252676359c671d65c9588b05572d76c99cf4c90666f0ac089b2f778450a5"} Oct 02 01:47:17 crc kubenswrapper[4885]: I1002 01:47:17.293407 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" event={"ID":"9e4679dd-f870-41e9-a32b-360f5d2ee81b","Type":"ContainerStarted","Data":"c9f99f9e26a04fb42b01a0783a5c4e360514818ce9102edc7fa08626d50b9a34"} Oct 02 01:47:17 crc kubenswrapper[4885]: I1002 01:47:17.308676 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vhgbd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c63a7d20d39b65acfb5a9be4c28b21c72111b398bea76c238e15ddb4dd63e139\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7zwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vhgbd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:17Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:17 crc kubenswrapper[4885]: I1002 01:47:17.335746 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8070bc1c-e3de-4e9a-985e-40d1a9858056\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3e634b3727c517cc48b17c2cfb38f26460f53d1bf25d631d08ec9acb8d780e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4adcc537c23697a324d82af94c31059dd444245efdf7c978db58ede256e112ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edcd1d99644a778dfb6d0a17d2a4abd906400fe8209e611ebf8d7a9abe7ccf28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://97cb59adc2ddfa365ebc9dec649600dc90466ffe59134a6d15a9d3134a7c9b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d2808bc658ea80d68469ed16422d56ef19447f4c90b4af7b2597c64406f8a44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://83acbc20919f7c1958b6d492fadb203855df02761662425df9f91a0757fe2bf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83acbc20919f7c1958b6d492fadb203855df02761662425df9f91a0757fe2bf7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e86ecd03835efc9a253b6804bc61e6ac690e75b8d1f3cc77f5647730e586574\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e86ecd03835efc9a253b6804bc61e6ac690e75b8d1f3cc77f5647730e586574\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://48f95af51feaf014a18496bdfadd59d7f1e11167c5c8b677fa7037e48368ff2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48f95af51feaf014a18496bdfadd59d7f1e11167c5c8b677fa7037e48368ff2a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:17Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:17 crc kubenswrapper[4885]: I1002 01:47:17.352955 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5c1f95f188569c9aa47e51f2916939fa0b65a5cdf4c16b798612d41f3c8411a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:17Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:17 crc kubenswrapper[4885]: I1002 01:47:17.371550 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:17Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:17 crc kubenswrapper[4885]: I1002 01:47:17.389471 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74adcf912ccedb6b5d4f2803fd215e3e2ef62367597aa6da528354c4511ee536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7403ef968d040ac09d722b7557438bef23892d578992826f42b700b370d495d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:17Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:17 crc kubenswrapper[4885]: I1002 01:47:17.410368 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2ll9b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52b39712-5fa1-4fe0-814b-b170a6e3938b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c979b91ba748aa4647826a8926337662ca929c54090fb62285572819ec2fbf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c979b91ba748aa4647826a8926337662ca929c54090fb62285572819ec2fbf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97d9d693e8e60ee9c183bb44461e698e8b3cfd20a516605aa8754c9205d2bdc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://97d9d693e8e60ee9c183bb44461e698e8b3cfd20a516605aa8754c9205d2bdc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c44252676359c671d65c9588b05572d76c99cf4c90666f0ac089b2f778450a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c44252676359c671d65c9588b05572d76c99cf4c90666f0ac089b2f778450a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2ll9b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:17Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:17 crc kubenswrapper[4885]: I1002 01:47:17.425161 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:17Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:17 crc kubenswrapper[4885]: I1002 01:47:17.442099 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:17Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:17 crc kubenswrapper[4885]: I1002 01:47:17.454335 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-58lhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac6a19c4-f259-4dbd-a016-d6c2c849c6cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed76c0f968c063f7795d363af194a222487b647ceda234de0dabb39b25eafe0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hzc4c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-58lhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:17Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:17 crc kubenswrapper[4885]: I1002 01:47:17.466986 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d169c8ba-000a-4218-8e3b-4ae53035b110\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61e436b7876107ce5ea9e668c0bbc367be6e663fc804fa3a65b51adf8be8d51f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-77rzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09536e182a46e77318c10812d1777e551f834907c006cab5c726dc66f5bce88a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-77rzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rttx8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:17Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:17 crc kubenswrapper[4885]: I1002 01:47:17.480118 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jgdvt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81733e8d-67b0-46fa-8a41-08b40645bd3e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcef60c816ca1ba9a298da658e84c4d717bea3d352c4c2f807ebd8e56327b384\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cwbl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jgdvt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:17Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:17 crc kubenswrapper[4885]: I1002 01:47:17.496490 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a0bb885-2eb0-419c-8a7a-ef634cb3c144\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a89b49cbe08d927629564e607de13655ddf3d01ac4eb5d688b325210306c120c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06c5942805a498d9621a906088aa54538b4eed583d1b8d0e3ba2a6718378a36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b9f6f7fe44d5771a838d0b89c5bdea3ebb79023ae2675ee907ec2766b3d35a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://178b4340ee6904c60c814b99d83ee225161aeb23e438d8d02f856579a13ec4e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bb63bf30d6ecce9e71f0db0c16632bdfa6b33924b673378eb0c38ab229fa837\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T01:47:06Z\\\",\\\"message\\\":\\\"W1002 01:46:55.406027 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1002 01:46:55.406714 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759369615 cert, and key in /tmp/serving-cert-3795040473/serving-signer.crt, /tmp/serving-cert-3795040473/serving-signer.key\\\\nI1002 01:46:55.614670 1 observer_polling.go:159] Starting file observer\\\\nW1002 01:46:55.617662 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1002 01:46:55.618112 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 01:46:55.620347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3795040473/tls.crt::/tmp/serving-cert-3795040473/tls.key\\\\\\\"\\\\nF1002 01:47:06.224459 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90535f97ab2ed4626d47540ebd7baada837fd332140323bec0f49aa38d3c0009\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://592f89a326c9c1559193a71904b47ca27b1dbe1a7b30e03bc48924bb2f35f921\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://592f89a326c9c1559193a71904b47ca27b1dbe1a7b30e03bc48924bb2f35f921\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:17Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:17 crc kubenswrapper[4885]: I1002 01:47:17.515662 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b087880e05039816dcab154643ac72c7e4a0a1124817c51dc2c639b62bbee0fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:17Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:17 crc kubenswrapper[4885]: I1002 01:47:17.544179 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"506a599f-fb1c-4d14-be88-0e4833f3d624\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31c2f0193ffef54965b79a3569b0f6f4547e540bac0774155159a1c996c196c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://533e73e5f90b7258acbcdf018ee01323c118f9dc45035b25b065ba01cabecccf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c00db833a593ec40eb63a6b383455e19e960c2e98b9b66f76877ee993057e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21e2f462a0deb0d740185663b2014f4a1e57b4fdbeb82571264b002f928035bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:17Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:17 crc kubenswrapper[4885]: I1002 01:47:17.585862 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e4679dd-f870-41e9-a32b-360f5d2ee81b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0125e907d061a97286026ee823c4eed303734ed03ab4864f742592b23a1aa2f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0125e907d061a97286026ee823c4eed303734ed03ab4864f742592b23a1aa2f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dngcm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:17Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:17 crc kubenswrapper[4885]: I1002 01:47:17.948921 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 01:47:17 crc kubenswrapper[4885]: I1002 01:47:17.956671 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 01:47:17 crc kubenswrapper[4885]: I1002 01:47:17.971767 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:17Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:17 crc kubenswrapper[4885]: I1002 01:47:17.991923 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74adcf912ccedb6b5d4f2803fd215e3e2ef62367597aa6da528354c4511ee536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7403ef968d040ac09d722b7557438bef23892d578992826f42b700b370d495d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:17Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.015026 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2ll9b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52b39712-5fa1-4fe0-814b-b170a6e3938b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c979b91ba748aa4647826a8926337662ca929c54090fb62285572819ec2fbf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c979b91ba748aa4647826a8926337662ca929c54090fb62285572819ec2fbf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97d9d693e8e60ee9c183bb44461e698e8b3cfd20a516605aa8754c9205d2bdc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://97d9d693e8e60ee9c183bb44461e698e8b3cfd20a516605aa8754c9205d2bdc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c44252676359c671d65c9588b05572d76c99cf4c90666f0ac089b2f778450a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c44252676359c671d65c9588b05572d76c99cf4c90666f0ac089b2f778450a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2ll9b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:18Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.030413 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-58lhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac6a19c4-f259-4dbd-a016-d6c2c849c6cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed76c0f968c063f7795d363af194a222487b647ceda234de0dabb39b25eafe0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hzc4c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-58lhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:18Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.045849 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.046003 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.046086 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d169c8ba-000a-4218-8e3b-4ae53035b110\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61e436b7876107ce5ea9e668c0bbc367be6e663fc804fa3a65b51adf8be8d51f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-77rzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09536e182a46e77318c10812d1777e551f834907c006cab5c726dc66f5bce88a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-77rzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rttx8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:18Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:18 crc kubenswrapper[4885]: E1002 01:47:18.046232 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.046283 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:47:18 crc kubenswrapper[4885]: E1002 01:47:18.046412 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:47:18 crc kubenswrapper[4885]: E1002 01:47:18.046578 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.062509 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jgdvt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81733e8d-67b0-46fa-8a41-08b40645bd3e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcef60c816ca1ba9a298da658e84c4d717bea3d352c4c2f807ebd8e56327b384\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cwbl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jgdvt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:18Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.066859 4885 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.070898 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.070988 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.071022 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.071410 4885 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.080675 4885 kubelet_node_status.go:115] "Node was previously registered" node="crc" Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.081147 4885 kubelet_node_status.go:79] "Successfully registered node" node="crc" Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.082710 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.082756 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.082772 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.082797 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.082815 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:18Z","lastTransitionTime":"2025-10-02T01:47:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.088855 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a0bb885-2eb0-419c-8a7a-ef634cb3c144\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a89b49cbe08d927629564e607de13655ddf3d01ac4eb5d688b325210306c120c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06c5942805a498d9621a906088aa54538b4eed583d1b8d0e3ba2a6718378a36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b9f6f7fe44d5771a838d0b89c5bdea3ebb79023ae2675ee907ec2766b3d35a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://178b4340ee6904c60c814b99d83ee225161aeb23e438d8d02f856579a13ec4e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bb63bf30d6ecce9e71f0db0c16632bdfa6b33924b673378eb0c38ab229fa837\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T01:47:06Z\\\",\\\"message\\\":\\\"W1002 01:46:55.406027 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1002 01:46:55.406714 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759369615 cert, and key in /tmp/serving-cert-3795040473/serving-signer.crt, /tmp/serving-cert-3795040473/serving-signer.key\\\\nI1002 01:46:55.614670 1 observer_polling.go:159] Starting file observer\\\\nW1002 01:46:55.617662 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1002 01:46:55.618112 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 01:46:55.620347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3795040473/tls.crt::/tmp/serving-cert-3795040473/tls.key\\\\\\\"\\\\nF1002 01:47:06.224459 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90535f97ab2ed4626d47540ebd7baada837fd332140323bec0f49aa38d3c0009\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://592f89a326c9c1559193a71904b47ca27b1dbe1a7b30e03bc48924bb2f35f921\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://592f89a326c9c1559193a71904b47ca27b1dbe1a7b30e03bc48924bb2f35f921\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:18Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:18 crc kubenswrapper[4885]: E1002 01:47:18.101786 4885 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:47:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:47:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:47:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:47:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"92603ae5-ba38-4201-98e9-979fd56cb50d\\\",\\\"systemUUID\\\":\\\"50ddf2c2-ea95-4960-ae48-d17e9484ee2f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:18Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.106406 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.106460 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.106477 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.106498 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.106516 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:18Z","lastTransitionTime":"2025-10-02T01:47:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.109347 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b087880e05039816dcab154643ac72c7e4a0a1124817c51dc2c639b62bbee0fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:18Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:18 crc kubenswrapper[4885]: E1002 01:47:18.126134 4885 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:47:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:47:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:47:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:47:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"92603ae5-ba38-4201-98e9-979fd56cb50d\\\",\\\"systemUUID\\\":\\\"50ddf2c2-ea95-4960-ae48-d17e9484ee2f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:18Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.131602 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:18Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.131778 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.131825 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.131843 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.131868 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.131885 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:18Z","lastTransitionTime":"2025-10-02T01:47:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.150257 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:18Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:18 crc kubenswrapper[4885]: E1002 01:47:18.154696 4885 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:47:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:47:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:47:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:47:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"92603ae5-ba38-4201-98e9-979fd56cb50d\\\",\\\"systemUUID\\\":\\\"50ddf2c2-ea95-4960-ae48-d17e9484ee2f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:18Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.159767 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.160215 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.160419 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.160625 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.160851 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:18Z","lastTransitionTime":"2025-10-02T01:47:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.168896 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"506a599f-fb1c-4d14-be88-0e4833f3d624\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31c2f0193ffef54965b79a3569b0f6f4547e540bac0774155159a1c996c196c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://533e73e5f90b7258acbcdf018ee01323c118f9dc45035b25b065ba01cabecccf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c00db833a593ec40eb63a6b383455e19e960c2e98b9b66f76877ee993057e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21e2f462a0deb0d740185663b2014f4a1e57b4fdbeb82571264b002f928035bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:18Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:18 crc kubenswrapper[4885]: E1002 01:47:18.179513 4885 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:47:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:47:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:47:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:47:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"92603ae5-ba38-4201-98e9-979fd56cb50d\\\",\\\"systemUUID\\\":\\\"50ddf2c2-ea95-4960-ae48-d17e9484ee2f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:18Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.183996 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.184034 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.184050 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.184071 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.184087 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:18Z","lastTransitionTime":"2025-10-02T01:47:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.197329 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e4679dd-f870-41e9-a32b-360f5d2ee81b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0125e907d061a97286026ee823c4eed303734ed03ab4864f742592b23a1aa2f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0125e907d061a97286026ee823c4eed303734ed03ab4864f742592b23a1aa2f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dngcm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:18Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:18 crc kubenswrapper[4885]: E1002 01:47:18.200424 4885 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:47:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:47:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:47:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:47:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"92603ae5-ba38-4201-98e9-979fd56cb50d\\\",\\\"systemUUID\\\":\\\"50ddf2c2-ea95-4960-ae48-d17e9484ee2f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:18Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:18 crc kubenswrapper[4885]: E1002 01:47:18.200640 4885 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.202813 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.202879 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.202897 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.202923 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.202941 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:18Z","lastTransitionTime":"2025-10-02T01:47:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.228952 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8070bc1c-e3de-4e9a-985e-40d1a9858056\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3e634b3727c517cc48b17c2cfb38f26460f53d1bf25d631d08ec9acb8d780e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4adcc537c23697a324d82af94c31059dd444245efdf7c978db58ede256e112ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edcd1d99644a778dfb6d0a17d2a4abd906400fe8209e611ebf8d7a9abe7ccf28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://97cb59adc2ddfa365ebc9dec649600dc90466ffe59134a6d15a9d3134a7c9b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d2808bc658ea80d68469ed16422d56ef19447f4c90b4af7b2597c64406f8a44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://83acbc20919f7c1958b6d492fadb203855df02761662425df9f91a0757fe2bf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83acbc20919f7c1958b6d492fadb203855df02761662425df9f91a0757fe2bf7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e86ecd03835efc9a253b6804bc61e6ac690e75b8d1f3cc77f5647730e586574\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e86ecd03835efc9a253b6804bc61e6ac690e75b8d1f3cc77f5647730e586574\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://48f95af51feaf014a18496bdfadd59d7f1e11167c5c8b677fa7037e48368ff2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48f95af51feaf014a18496bdfadd59d7f1e11167c5c8b677fa7037e48368ff2a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:18Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.248350 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5c1f95f188569c9aa47e51f2916939fa0b65a5cdf4c16b798612d41f3c8411a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:18Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.266695 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vhgbd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c63a7d20d39b65acfb5a9be4c28b21c72111b398bea76c238e15ddb4dd63e139\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7zwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vhgbd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:18Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.283473 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:18Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.300366 4885 generic.go:334] "Generic (PLEG): container finished" podID="52b39712-5fa1-4fe0-814b-b170a6e3938b" containerID="80c982f07874df686fe7e29cc91043eb6d58f1be5fc93cacd46d33a21dd9b622" exitCode=0 Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.300445 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2ll9b" event={"ID":"52b39712-5fa1-4fe0-814b-b170a6e3938b","Type":"ContainerDied","Data":"80c982f07874df686fe7e29cc91043eb6d58f1be5fc93cacd46d33a21dd9b622"} Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.305129 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.305169 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.305186 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.305207 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.305225 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:18Z","lastTransitionTime":"2025-10-02T01:47:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.307172 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74adcf912ccedb6b5d4f2803fd215e3e2ef62367597aa6da528354c4511ee536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7403ef968d040ac09d722b7557438bef23892d578992826f42b700b370d495d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:18Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.349638 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2ll9b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52b39712-5fa1-4fe0-814b-b170a6e3938b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c979b91ba748aa4647826a8926337662ca929c54090fb62285572819ec2fbf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c979b91ba748aa4647826a8926337662ca929c54090fb62285572819ec2fbf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97d9d693e8e60ee9c183bb44461e698e8b3cfd20a516605aa8754c9205d2bdc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://97d9d693e8e60ee9c183bb44461e698e8b3cfd20a516605aa8754c9205d2bdc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c44252676359c671d65c9588b05572d76c99cf4c90666f0ac089b2f778450a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c44252676359c671d65c9588b05572d76c99cf4c90666f0ac089b2f778450a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2ll9b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:18Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.404383 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jgdvt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81733e8d-67b0-46fa-8a41-08b40645bd3e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcef60c816ca1ba9a298da658e84c4d717bea3d352c4c2f807ebd8e56327b384\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cwbl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jgdvt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:18Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.407370 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.407419 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.407434 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.407455 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.407470 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:18Z","lastTransitionTime":"2025-10-02T01:47:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.424993 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a0bb885-2eb0-419c-8a7a-ef634cb3c144\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a89b49cbe08d927629564e607de13655ddf3d01ac4eb5d688b325210306c120c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06c5942805a498d9621a906088aa54538b4eed583d1b8d0e3ba2a6718378a36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b9f6f7fe44d5771a838d0b89c5bdea3ebb79023ae2675ee907ec2766b3d35a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://178b4340ee6904c60c814b99d83ee225161aeb23e438d8d02f856579a13ec4e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bb63bf30d6ecce9e71f0db0c16632bdfa6b33924b673378eb0c38ab229fa837\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T01:47:06Z\\\",\\\"message\\\":\\\"W1002 01:46:55.406027 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1002 01:46:55.406714 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759369615 cert, and key in /tmp/serving-cert-3795040473/serving-signer.crt, /tmp/serving-cert-3795040473/serving-signer.key\\\\nI1002 01:46:55.614670 1 observer_polling.go:159] Starting file observer\\\\nW1002 01:46:55.617662 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1002 01:46:55.618112 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 01:46:55.620347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3795040473/tls.crt::/tmp/serving-cert-3795040473/tls.key\\\\\\\"\\\\nF1002 01:47:06.224459 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90535f97ab2ed4626d47540ebd7baada837fd332140323bec0f49aa38d3c0009\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://592f89a326c9c1559193a71904b47ca27b1dbe1a7b30e03bc48924bb2f35f921\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://592f89a326c9c1559193a71904b47ca27b1dbe1a7b30e03bc48924bb2f35f921\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:18Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.426621 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.464760 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b087880e05039816dcab154643ac72c7e4a0a1124817c51dc2c639b62bbee0fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:18Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.502309 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:18Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.513490 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.515976 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.516003 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.516040 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.516058 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:18Z","lastTransitionTime":"2025-10-02T01:47:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.543345 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:18Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.578225 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-58lhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac6a19c4-f259-4dbd-a016-d6c2c849c6cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed76c0f968c063f7795d363af194a222487b647ceda234de0dabb39b25eafe0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hzc4c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-58lhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:18Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.621397 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.621452 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.621472 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.621511 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.621529 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:18Z","lastTransitionTime":"2025-10-02T01:47:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.623385 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d169c8ba-000a-4218-8e3b-4ae53035b110\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61e436b7876107ce5ea9e668c0bbc367be6e663fc804fa3a65b51adf8be8d51f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-77rzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09536e182a46e77318c10812d1777e551f834907c006cab5c726dc66f5bce88a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-77rzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rttx8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:18Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.663653 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"506a599f-fb1c-4d14-be88-0e4833f3d624\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31c2f0193ffef54965b79a3569b0f6f4547e540bac0774155159a1c996c196c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://533e73e5f90b7258acbcdf018ee01323c118f9dc45035b25b065ba01cabecccf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c00db833a593ec40eb63a6b383455e19e960c2e98b9b66f76877ee993057e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21e2f462a0deb0d740185663b2014f4a1e57b4fdbeb82571264b002f928035bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:18Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.719425 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e4679dd-f870-41e9-a32b-360f5d2ee81b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0125e907d061a97286026ee823c4eed303734ed03ab4864f742592b23a1aa2f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0125e907d061a97286026ee823c4eed303734ed03ab4864f742592b23a1aa2f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dngcm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:18Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.725099 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.725177 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.725196 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.725222 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.725240 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:18Z","lastTransitionTime":"2025-10-02T01:47:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.752300 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8070bc1c-e3de-4e9a-985e-40d1a9858056\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3e634b3727c517cc48b17c2cfb38f26460f53d1bf25d631d08ec9acb8d780e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4adcc537c23697a324d82af94c31059dd444245efdf7c978db58ede256e112ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edcd1d99644a778dfb6d0a17d2a4abd906400fe8209e611ebf8d7a9abe7ccf28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://97cb59adc2ddfa365ebc9dec649600dc90466ffe59134a6d15a9d3134a7c9b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d2808bc658ea80d68469ed16422d56ef19447f4c90b4af7b2597c64406f8a44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://83acbc20919f7c1958b6d492fadb203855df02761662425df9f91a0757fe2bf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83acbc20919f7c1958b6d492fadb203855df02761662425df9f91a0757fe2bf7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e86ecd03835efc9a253b6804bc61e6ac690e75b8d1f3cc77f5647730e586574\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e86ecd03835efc9a253b6804bc61e6ac690e75b8d1f3cc77f5647730e586574\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://48f95af51feaf014a18496bdfadd59d7f1e11167c5c8b677fa7037e48368ff2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48f95af51feaf014a18496bdfadd59d7f1e11167c5c8b677fa7037e48368ff2a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:18Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.783834 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5c1f95f188569c9aa47e51f2916939fa0b65a5cdf4c16b798612d41f3c8411a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:18Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.826408 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vhgbd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c63a7d20d39b65acfb5a9be4c28b21c72111b398bea76c238e15ddb4dd63e139\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7zwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vhgbd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:18Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.828308 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.828377 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.828400 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.828429 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.828451 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:18Z","lastTransitionTime":"2025-10-02T01:47:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.864303 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:18Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.906502 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74adcf912ccedb6b5d4f2803fd215e3e2ef62367597aa6da528354c4511ee536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7403ef968d040ac09d722b7557438bef23892d578992826f42b700b370d495d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:18Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.931122 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.931189 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.931208 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.931233 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.931251 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:18Z","lastTransitionTime":"2025-10-02T01:47:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.951923 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2ll9b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52b39712-5fa1-4fe0-814b-b170a6e3938b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c979b91ba748aa4647826a8926337662ca929c54090fb62285572819ec2fbf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c979b91ba748aa4647826a8926337662ca929c54090fb62285572819ec2fbf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97d9d693e8e60ee9c183bb44461e698e8b3cfd20a516605aa8754c9205d2bdc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://97d9d693e8e60ee9c183bb44461e698e8b3cfd20a516605aa8754c9205d2bdc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c44252676359c671d65c9588b05572d76c99cf4c90666f0ac089b2f778450a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c44252676359c671d65c9588b05572d76c99cf4c90666f0ac089b2f778450a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80c982f07874df686fe7e29cc91043eb6d58f1be5fc93cacd46d33a21dd9b622\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80c982f07874df686fe7e29cc91043eb6d58f1be5fc93cacd46d33a21dd9b622\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2ll9b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:18Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:18 crc kubenswrapper[4885]: I1002 01:47:18.980018 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-58lhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac6a19c4-f259-4dbd-a016-d6c2c849c6cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed76c0f968c063f7795d363af194a222487b647ceda234de0dabb39b25eafe0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hzc4c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-58lhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:18Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:19 crc kubenswrapper[4885]: I1002 01:47:19.023677 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d169c8ba-000a-4218-8e3b-4ae53035b110\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61e436b7876107ce5ea9e668c0bbc367be6e663fc804fa3a65b51adf8be8d51f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-77rzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09536e182a46e77318c10812d1777e551f834907c006cab5c726dc66f5bce88a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-77rzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rttx8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:19Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:19 crc kubenswrapper[4885]: I1002 01:47:19.034468 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:19 crc kubenswrapper[4885]: I1002 01:47:19.034525 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:19 crc kubenswrapper[4885]: I1002 01:47:19.034544 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:19 crc kubenswrapper[4885]: I1002 01:47:19.034569 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:19 crc kubenswrapper[4885]: I1002 01:47:19.034587 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:19Z","lastTransitionTime":"2025-10-02T01:47:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:19 crc kubenswrapper[4885]: I1002 01:47:19.059565 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jgdvt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81733e8d-67b0-46fa-8a41-08b40645bd3e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcef60c816ca1ba9a298da658e84c4d717bea3d352c4c2f807ebd8e56327b384\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cwbl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jgdvt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:19Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:19 crc kubenswrapper[4885]: I1002 01:47:19.104056 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a0bb885-2eb0-419c-8a7a-ef634cb3c144\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a89b49cbe08d927629564e607de13655ddf3d01ac4eb5d688b325210306c120c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06c5942805a498d9621a906088aa54538b4eed583d1b8d0e3ba2a6718378a36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b9f6f7fe44d5771a838d0b89c5bdea3ebb79023ae2675ee907ec2766b3d35a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://178b4340ee6904c60c814b99d83ee225161aeb23e438d8d02f856579a13ec4e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bb63bf30d6ecce9e71f0db0c16632bdfa6b33924b673378eb0c38ab229fa837\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T01:47:06Z\\\",\\\"message\\\":\\\"W1002 01:46:55.406027 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1002 01:46:55.406714 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759369615 cert, and key in /tmp/serving-cert-3795040473/serving-signer.crt, /tmp/serving-cert-3795040473/serving-signer.key\\\\nI1002 01:46:55.614670 1 observer_polling.go:159] Starting file observer\\\\nW1002 01:46:55.617662 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1002 01:46:55.618112 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 01:46:55.620347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3795040473/tls.crt::/tmp/serving-cert-3795040473/tls.key\\\\\\\"\\\\nF1002 01:47:06.224459 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90535f97ab2ed4626d47540ebd7baada837fd332140323bec0f49aa38d3c0009\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://592f89a326c9c1559193a71904b47ca27b1dbe1a7b30e03bc48924bb2f35f921\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://592f89a326c9c1559193a71904b47ca27b1dbe1a7b30e03bc48924bb2f35f921\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:19Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:19 crc kubenswrapper[4885]: I1002 01:47:19.136687 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:19 crc kubenswrapper[4885]: I1002 01:47:19.136741 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:19 crc kubenswrapper[4885]: I1002 01:47:19.136756 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:19 crc kubenswrapper[4885]: I1002 01:47:19.136776 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:19 crc kubenswrapper[4885]: I1002 01:47:19.136790 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:19Z","lastTransitionTime":"2025-10-02T01:47:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:19 crc kubenswrapper[4885]: I1002 01:47:19.143188 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b087880e05039816dcab154643ac72c7e4a0a1124817c51dc2c639b62bbee0fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:19Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:19 crc kubenswrapper[4885]: I1002 01:47:19.182430 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:19Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:19 crc kubenswrapper[4885]: I1002 01:47:19.225580 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:19Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:19 crc kubenswrapper[4885]: I1002 01:47:19.239524 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:19 crc kubenswrapper[4885]: I1002 01:47:19.239879 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:19 crc kubenswrapper[4885]: I1002 01:47:19.240004 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:19 crc kubenswrapper[4885]: I1002 01:47:19.240170 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:19 crc kubenswrapper[4885]: I1002 01:47:19.240315 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:19Z","lastTransitionTime":"2025-10-02T01:47:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:19 crc kubenswrapper[4885]: I1002 01:47:19.263277 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"506a599f-fb1c-4d14-be88-0e4833f3d624\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31c2f0193ffef54965b79a3569b0f6f4547e540bac0774155159a1c996c196c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://533e73e5f90b7258acbcdf018ee01323c118f9dc45035b25b065ba01cabecccf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c00db833a593ec40eb63a6b383455e19e960c2e98b9b66f76877ee993057e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21e2f462a0deb0d740185663b2014f4a1e57b4fdbeb82571264b002f928035bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:19Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:19 crc kubenswrapper[4885]: I1002 01:47:19.317142 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e4679dd-f870-41e9-a32b-360f5d2ee81b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0125e907d061a97286026ee823c4eed303734ed03ab4864f742592b23a1aa2f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0125e907d061a97286026ee823c4eed303734ed03ab4864f742592b23a1aa2f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dngcm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:19Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:19 crc kubenswrapper[4885]: I1002 01:47:19.322618 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" event={"ID":"9e4679dd-f870-41e9-a32b-360f5d2ee81b","Type":"ContainerStarted","Data":"fafdd1094bd02786d745c984989852b0c690db38655f92b419733a2004f417c0"} Oct 02 01:47:19 crc kubenswrapper[4885]: I1002 01:47:19.340642 4885 generic.go:334] "Generic (PLEG): container finished" podID="52b39712-5fa1-4fe0-814b-b170a6e3938b" containerID="1a354c985652c7ad7fa8944fbfdfe223cfb0f97fb11eecd9bd429020d4892b78" exitCode=0 Oct 02 01:47:19 crc kubenswrapper[4885]: I1002 01:47:19.340703 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2ll9b" event={"ID":"52b39712-5fa1-4fe0-814b-b170a6e3938b","Type":"ContainerDied","Data":"1a354c985652c7ad7fa8944fbfdfe223cfb0f97fb11eecd9bd429020d4892b78"} Oct 02 01:47:19 crc kubenswrapper[4885]: I1002 01:47:19.349852 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:19 crc kubenswrapper[4885]: I1002 01:47:19.349882 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:19 crc kubenswrapper[4885]: I1002 01:47:19.349893 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:19 crc kubenswrapper[4885]: I1002 01:47:19.349908 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:19 crc kubenswrapper[4885]: I1002 01:47:19.349920 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:19Z","lastTransitionTime":"2025-10-02T01:47:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:19 crc kubenswrapper[4885]: I1002 01:47:19.376251 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8070bc1c-e3de-4e9a-985e-40d1a9858056\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3e634b3727c517cc48b17c2cfb38f26460f53d1bf25d631d08ec9acb8d780e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4adcc537c23697a324d82af94c31059dd444245efdf7c978db58ede256e112ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edcd1d99644a778dfb6d0a17d2a4abd906400fe8209e611ebf8d7a9abe7ccf28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://97cb59adc2ddfa365ebc9dec649600dc90466ffe59134a6d15a9d3134a7c9b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d2808bc658ea80d68469ed16422d56ef19447f4c90b4af7b2597c64406f8a44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://83acbc20919f7c1958b6d492fadb203855df02761662425df9f91a0757fe2bf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83acbc20919f7c1958b6d492fadb203855df02761662425df9f91a0757fe2bf7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e86ecd03835efc9a253b6804bc61e6ac690e75b8d1f3cc77f5647730e586574\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e86ecd03835efc9a253b6804bc61e6ac690e75b8d1f3cc77f5647730e586574\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://48f95af51feaf014a18496bdfadd59d7f1e11167c5c8b677fa7037e48368ff2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48f95af51feaf014a18496bdfadd59d7f1e11167c5c8b677fa7037e48368ff2a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:19Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:19 crc kubenswrapper[4885]: I1002 01:47:19.396414 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5c1f95f188569c9aa47e51f2916939fa0b65a5cdf4c16b798612d41f3c8411a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:19Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:19 crc kubenswrapper[4885]: I1002 01:47:19.419181 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vhgbd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c63a7d20d39b65acfb5a9be4c28b21c72111b398bea76c238e15ddb4dd63e139\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7zwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vhgbd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:19Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:19 crc kubenswrapper[4885]: I1002 01:47:19.464924 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a0bb885-2eb0-419c-8a7a-ef634cb3c144\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a89b49cbe08d927629564e607de13655ddf3d01ac4eb5d688b325210306c120c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06c5942805a498d9621a906088aa54538b4eed583d1b8d0e3ba2a6718378a36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b9f6f7fe44d5771a838d0b89c5bdea3ebb79023ae2675ee907ec2766b3d35a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://178b4340ee6904c60c814b99d83ee225161aeb23e438d8d02f856579a13ec4e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bb63bf30d6ecce9e71f0db0c16632bdfa6b33924b673378eb0c38ab229fa837\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T01:47:06Z\\\",\\\"message\\\":\\\"W1002 01:46:55.406027 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1002 01:46:55.406714 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759369615 cert, and key in /tmp/serving-cert-3795040473/serving-signer.crt, /tmp/serving-cert-3795040473/serving-signer.key\\\\nI1002 01:46:55.614670 1 observer_polling.go:159] Starting file observer\\\\nW1002 01:46:55.617662 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1002 01:46:55.618112 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 01:46:55.620347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3795040473/tls.crt::/tmp/serving-cert-3795040473/tls.key\\\\\\\"\\\\nF1002 01:47:06.224459 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90535f97ab2ed4626d47540ebd7baada837fd332140323bec0f49aa38d3c0009\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://592f89a326c9c1559193a71904b47ca27b1dbe1a7b30e03bc48924bb2f35f921\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://592f89a326c9c1559193a71904b47ca27b1dbe1a7b30e03bc48924bb2f35f921\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:19Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:19 crc kubenswrapper[4885]: I1002 01:47:19.467781 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:19 crc kubenswrapper[4885]: I1002 01:47:19.467889 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:19 crc kubenswrapper[4885]: I1002 01:47:19.467934 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:19 crc kubenswrapper[4885]: I1002 01:47:19.467958 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:19 crc kubenswrapper[4885]: I1002 01:47:19.467979 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:19Z","lastTransitionTime":"2025-10-02T01:47:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:19 crc kubenswrapper[4885]: I1002 01:47:19.499576 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b087880e05039816dcab154643ac72c7e4a0a1124817c51dc2c639b62bbee0fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:19Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:19 crc kubenswrapper[4885]: I1002 01:47:19.568021 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:19Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:19 crc kubenswrapper[4885]: I1002 01:47:19.570954 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:19 crc kubenswrapper[4885]: I1002 01:47:19.571020 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:19 crc kubenswrapper[4885]: I1002 01:47:19.571040 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:19 crc kubenswrapper[4885]: I1002 01:47:19.571064 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:19 crc kubenswrapper[4885]: I1002 01:47:19.571083 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:19Z","lastTransitionTime":"2025-10-02T01:47:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:19 crc kubenswrapper[4885]: I1002 01:47:19.585929 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:19Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:19 crc kubenswrapper[4885]: I1002 01:47:19.620462 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-58lhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac6a19c4-f259-4dbd-a016-d6c2c849c6cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed76c0f968c063f7795d363af194a222487b647ceda234de0dabb39b25eafe0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hzc4c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-58lhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:19Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:19 crc kubenswrapper[4885]: I1002 01:47:19.660780 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d169c8ba-000a-4218-8e3b-4ae53035b110\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61e436b7876107ce5ea9e668c0bbc367be6e663fc804fa3a65b51adf8be8d51f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-77rzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09536e182a46e77318c10812d1777e551f834907c006cab5c726dc66f5bce88a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-77rzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rttx8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:19Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:19 crc kubenswrapper[4885]: I1002 01:47:19.673296 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:19 crc kubenswrapper[4885]: I1002 01:47:19.673347 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:19 crc kubenswrapper[4885]: I1002 01:47:19.673366 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:19 crc kubenswrapper[4885]: I1002 01:47:19.673388 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:19 crc kubenswrapper[4885]: I1002 01:47:19.673404 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:19Z","lastTransitionTime":"2025-10-02T01:47:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:19 crc kubenswrapper[4885]: I1002 01:47:19.702074 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jgdvt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81733e8d-67b0-46fa-8a41-08b40645bd3e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcef60c816ca1ba9a298da658e84c4d717bea3d352c4c2f807ebd8e56327b384\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cwbl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jgdvt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:19Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:19 crc kubenswrapper[4885]: I1002 01:47:19.713193 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:47:19 crc kubenswrapper[4885]: E1002 01:47:19.713615 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:47:27.713511236 +0000 UTC m=+36.525258675 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:47:19 crc kubenswrapper[4885]: I1002 01:47:19.713865 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:47:19 crc kubenswrapper[4885]: E1002 01:47:19.714171 4885 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 01:47:19 crc kubenswrapper[4885]: E1002 01:47:19.714278 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 01:47:27.714239358 +0000 UTC m=+36.525986797 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 01:47:19 crc kubenswrapper[4885]: I1002 01:47:19.743439 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"506a599f-fb1c-4d14-be88-0e4833f3d624\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31c2f0193ffef54965b79a3569b0f6f4547e540bac0774155159a1c996c196c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://533e73e5f90b7258acbcdf018ee01323c118f9dc45035b25b065ba01cabecccf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c00db833a593ec40eb63a6b383455e19e960c2e98b9b66f76877ee993057e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21e2f462a0deb0d740185663b2014f4a1e57b4fdbeb82571264b002f928035bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:19Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:19 crc kubenswrapper[4885]: I1002 01:47:19.777220 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:19 crc kubenswrapper[4885]: I1002 01:47:19.777598 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:19 crc kubenswrapper[4885]: I1002 01:47:19.777777 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:19 crc kubenswrapper[4885]: I1002 01:47:19.777904 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:19 crc kubenswrapper[4885]: I1002 01:47:19.778020 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:19Z","lastTransitionTime":"2025-10-02T01:47:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:19 crc kubenswrapper[4885]: I1002 01:47:19.799870 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e4679dd-f870-41e9-a32b-360f5d2ee81b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0125e907d061a97286026ee823c4eed303734ed03ab4864f742592b23a1aa2f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0125e907d061a97286026ee823c4eed303734ed03ab4864f742592b23a1aa2f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dngcm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:19Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:19 crc kubenswrapper[4885]: I1002 01:47:19.814717 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:47:19 crc kubenswrapper[4885]: I1002 01:47:19.814790 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:47:19 crc kubenswrapper[4885]: I1002 01:47:19.814864 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:47:19 crc kubenswrapper[4885]: E1002 01:47:19.815014 4885 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 01:47:19 crc kubenswrapper[4885]: E1002 01:47:19.815075 4885 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 01:47:19 crc kubenswrapper[4885]: E1002 01:47:19.815092 4885 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 01:47:19 crc kubenswrapper[4885]: E1002 01:47:19.815155 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-02 01:47:27.815133746 +0000 UTC m=+36.626881155 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 01:47:19 crc kubenswrapper[4885]: E1002 01:47:19.815033 4885 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 01:47:19 crc kubenswrapper[4885]: E1002 01:47:19.815212 4885 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 01:47:19 crc kubenswrapper[4885]: E1002 01:47:19.815233 4885 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 01:47:19 crc kubenswrapper[4885]: E1002 01:47:19.815324 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-02 01:47:27.815299731 +0000 UTC m=+36.627047170 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 01:47:19 crc kubenswrapper[4885]: E1002 01:47:19.815741 4885 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 01:47:19 crc kubenswrapper[4885]: E1002 01:47:19.815957 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 01:47:27.8159283 +0000 UTC m=+36.627675729 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 01:47:19 crc kubenswrapper[4885]: I1002 01:47:19.838641 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8070bc1c-e3de-4e9a-985e-40d1a9858056\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3e634b3727c517cc48b17c2cfb38f26460f53d1bf25d631d08ec9acb8d780e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4adcc537c23697a324d82af94c31059dd444245efdf7c978db58ede256e112ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edcd1d99644a778dfb6d0a17d2a4abd906400fe8209e611ebf8d7a9abe7ccf28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://97cb59adc2ddfa365ebc9dec649600dc90466ffe59134a6d15a9d3134a7c9b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d2808bc658ea80d68469ed16422d56ef19447f4c90b4af7b2597c64406f8a44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://83acbc20919f7c1958b6d492fadb203855df02761662425df9f91a0757fe2bf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83acbc20919f7c1958b6d492fadb203855df02761662425df9f91a0757fe2bf7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e86ecd03835efc9a253b6804bc61e6ac690e75b8d1f3cc77f5647730e586574\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e86ecd03835efc9a253b6804bc61e6ac690e75b8d1f3cc77f5647730e586574\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://48f95af51feaf014a18496bdfadd59d7f1e11167c5c8b677fa7037e48368ff2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48f95af51feaf014a18496bdfadd59d7f1e11167c5c8b677fa7037e48368ff2a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:19Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:19 crc kubenswrapper[4885]: I1002 01:47:19.859143 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5c1f95f188569c9aa47e51f2916939fa0b65a5cdf4c16b798612d41f3c8411a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:19Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:19 crc kubenswrapper[4885]: I1002 01:47:19.880833 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:19 crc kubenswrapper[4885]: I1002 01:47:19.880886 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:19 crc kubenswrapper[4885]: I1002 01:47:19.880905 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:19 crc kubenswrapper[4885]: I1002 01:47:19.880929 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:19 crc kubenswrapper[4885]: I1002 01:47:19.880949 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:19Z","lastTransitionTime":"2025-10-02T01:47:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:19 crc kubenswrapper[4885]: I1002 01:47:19.906107 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vhgbd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c63a7d20d39b65acfb5a9be4c28b21c72111b398bea76c238e15ddb4dd63e139\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7zwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vhgbd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:19Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:19 crc kubenswrapper[4885]: I1002 01:47:19.940827 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:19Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:19 crc kubenswrapper[4885]: I1002 01:47:19.984399 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74adcf912ccedb6b5d4f2803fd215e3e2ef62367597aa6da528354c4511ee536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7403ef968d040ac09d722b7557438bef23892d578992826f42b700b370d495d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:19Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:19 crc kubenswrapper[4885]: I1002 01:47:19.984644 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:19 crc kubenswrapper[4885]: I1002 01:47:19.984726 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:19 crc kubenswrapper[4885]: I1002 01:47:19.984744 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:19 crc kubenswrapper[4885]: I1002 01:47:19.984771 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:19 crc kubenswrapper[4885]: I1002 01:47:19.984791 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:19Z","lastTransitionTime":"2025-10-02T01:47:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:20 crc kubenswrapper[4885]: I1002 01:47:20.030643 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2ll9b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52b39712-5fa1-4fe0-814b-b170a6e3938b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c979b91ba748aa4647826a8926337662ca929c54090fb62285572819ec2fbf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c979b91ba748aa4647826a8926337662ca929c54090fb62285572819ec2fbf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97d9d693e8e60ee9c183bb44461e698e8b3cfd20a516605aa8754c9205d2bdc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://97d9d693e8e60ee9c183bb44461e698e8b3cfd20a516605aa8754c9205d2bdc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c44252676359c671d65c9588b05572d76c99cf4c90666f0ac089b2f778450a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c44252676359c671d65c9588b05572d76c99cf4c90666f0ac089b2f778450a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80c982f07874df686fe7e29cc91043eb6d58f1be5fc93cacd46d33a21dd9b622\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80c982f07874df686fe7e29cc91043eb6d58f1be5fc93cacd46d33a21dd9b622\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a354c985652c7ad7fa8944fbfdfe223cfb0f97fb11eecd9bd429020d4892b78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a354c985652c7ad7fa8944fbfdfe223cfb0f97fb11eecd9bd429020d4892b78\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2ll9b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:20Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:20 crc kubenswrapper[4885]: I1002 01:47:20.045674 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:47:20 crc kubenswrapper[4885]: I1002 01:47:20.045740 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:47:20 crc kubenswrapper[4885]: I1002 01:47:20.045740 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:47:20 crc kubenswrapper[4885]: E1002 01:47:20.045844 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:47:20 crc kubenswrapper[4885]: E1002 01:47:20.046012 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:47:20 crc kubenswrapper[4885]: E1002 01:47:20.046146 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:47:20 crc kubenswrapper[4885]: I1002 01:47:20.087446 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:20 crc kubenswrapper[4885]: I1002 01:47:20.087508 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:20 crc kubenswrapper[4885]: I1002 01:47:20.087526 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:20 crc kubenswrapper[4885]: I1002 01:47:20.087548 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:20 crc kubenswrapper[4885]: I1002 01:47:20.087566 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:20Z","lastTransitionTime":"2025-10-02T01:47:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:20 crc kubenswrapper[4885]: I1002 01:47:20.190558 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:20 crc kubenswrapper[4885]: I1002 01:47:20.190620 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:20 crc kubenswrapper[4885]: I1002 01:47:20.190637 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:20 crc kubenswrapper[4885]: I1002 01:47:20.190661 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:20 crc kubenswrapper[4885]: I1002 01:47:20.190683 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:20Z","lastTransitionTime":"2025-10-02T01:47:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:20 crc kubenswrapper[4885]: I1002 01:47:20.292946 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:20 crc kubenswrapper[4885]: I1002 01:47:20.293002 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:20 crc kubenswrapper[4885]: I1002 01:47:20.293022 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:20 crc kubenswrapper[4885]: I1002 01:47:20.293047 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:20 crc kubenswrapper[4885]: I1002 01:47:20.293068 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:20Z","lastTransitionTime":"2025-10-02T01:47:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:20 crc kubenswrapper[4885]: I1002 01:47:20.348553 4885 generic.go:334] "Generic (PLEG): container finished" podID="52b39712-5fa1-4fe0-814b-b170a6e3938b" containerID="2efd7283df313d086c1b8b88f289ace28764de256205bee7610d6711002d8515" exitCode=0 Oct 02 01:47:20 crc kubenswrapper[4885]: I1002 01:47:20.348609 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2ll9b" event={"ID":"52b39712-5fa1-4fe0-814b-b170a6e3938b","Type":"ContainerDied","Data":"2efd7283df313d086c1b8b88f289ace28764de256205bee7610d6711002d8515"} Oct 02 01:47:20 crc kubenswrapper[4885]: I1002 01:47:20.367970 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d169c8ba-000a-4218-8e3b-4ae53035b110\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61e436b7876107ce5ea9e668c0bbc367be6e663fc804fa3a65b51adf8be8d51f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-77rzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09536e182a46e77318c10812d1777e551f834907c006cab5c726dc66f5bce88a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-77rzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rttx8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:20Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:20 crc kubenswrapper[4885]: I1002 01:47:20.389667 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jgdvt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81733e8d-67b0-46fa-8a41-08b40645bd3e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcef60c816ca1ba9a298da658e84c4d717bea3d352c4c2f807ebd8e56327b384\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cwbl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jgdvt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:20Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:20 crc kubenswrapper[4885]: I1002 01:47:20.396056 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:20 crc kubenswrapper[4885]: I1002 01:47:20.396170 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:20 crc kubenswrapper[4885]: I1002 01:47:20.396197 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:20 crc kubenswrapper[4885]: I1002 01:47:20.396231 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:20 crc kubenswrapper[4885]: I1002 01:47:20.396258 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:20Z","lastTransitionTime":"2025-10-02T01:47:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:20 crc kubenswrapper[4885]: I1002 01:47:20.414027 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a0bb885-2eb0-419c-8a7a-ef634cb3c144\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a89b49cbe08d927629564e607de13655ddf3d01ac4eb5d688b325210306c120c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06c5942805a498d9621a906088aa54538b4eed583d1b8d0e3ba2a6718378a36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b9f6f7fe44d5771a838d0b89c5bdea3ebb79023ae2675ee907ec2766b3d35a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://178b4340ee6904c60c814b99d83ee225161aeb23e438d8d02f856579a13ec4e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bb63bf30d6ecce9e71f0db0c16632bdfa6b33924b673378eb0c38ab229fa837\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T01:47:06Z\\\",\\\"message\\\":\\\"W1002 01:46:55.406027 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1002 01:46:55.406714 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759369615 cert, and key in /tmp/serving-cert-3795040473/serving-signer.crt, /tmp/serving-cert-3795040473/serving-signer.key\\\\nI1002 01:46:55.614670 1 observer_polling.go:159] Starting file observer\\\\nW1002 01:46:55.617662 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1002 01:46:55.618112 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 01:46:55.620347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3795040473/tls.crt::/tmp/serving-cert-3795040473/tls.key\\\\\\\"\\\\nF1002 01:47:06.224459 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90535f97ab2ed4626d47540ebd7baada837fd332140323bec0f49aa38d3c0009\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://592f89a326c9c1559193a71904b47ca27b1dbe1a7b30e03bc48924bb2f35f921\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://592f89a326c9c1559193a71904b47ca27b1dbe1a7b30e03bc48924bb2f35f921\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:20Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:20 crc kubenswrapper[4885]: I1002 01:47:20.438571 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b087880e05039816dcab154643ac72c7e4a0a1124817c51dc2c639b62bbee0fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:20Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:20 crc kubenswrapper[4885]: I1002 01:47:20.458296 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:20Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:20 crc kubenswrapper[4885]: I1002 01:47:20.477373 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:20Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:20 crc kubenswrapper[4885]: I1002 01:47:20.492189 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-58lhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac6a19c4-f259-4dbd-a016-d6c2c849c6cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed76c0f968c063f7795d363af194a222487b647ceda234de0dabb39b25eafe0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hzc4c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-58lhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:20Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:20 crc kubenswrapper[4885]: I1002 01:47:20.501293 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:20 crc kubenswrapper[4885]: I1002 01:47:20.501354 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:20 crc kubenswrapper[4885]: I1002 01:47:20.501376 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:20 crc kubenswrapper[4885]: I1002 01:47:20.501402 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:20 crc kubenswrapper[4885]: I1002 01:47:20.501419 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:20Z","lastTransitionTime":"2025-10-02T01:47:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:20 crc kubenswrapper[4885]: I1002 01:47:20.517465 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"506a599f-fb1c-4d14-be88-0e4833f3d624\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31c2f0193ffef54965b79a3569b0f6f4547e540bac0774155159a1c996c196c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://533e73e5f90b7258acbcdf018ee01323c118f9dc45035b25b065ba01cabecccf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c00db833a593ec40eb63a6b383455e19e960c2e98b9b66f76877ee993057e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21e2f462a0deb0d740185663b2014f4a1e57b4fdbeb82571264b002f928035bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:20Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:20 crc kubenswrapper[4885]: I1002 01:47:20.547570 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e4679dd-f870-41e9-a32b-360f5d2ee81b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0125e907d061a97286026ee823c4eed303734ed03ab4864f742592b23a1aa2f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0125e907d061a97286026ee823c4eed303734ed03ab4864f742592b23a1aa2f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dngcm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:20Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:20 crc kubenswrapper[4885]: I1002 01:47:20.579793 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8070bc1c-e3de-4e9a-985e-40d1a9858056\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3e634b3727c517cc48b17c2cfb38f26460f53d1bf25d631d08ec9acb8d780e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4adcc537c23697a324d82af94c31059dd444245efdf7c978db58ede256e112ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edcd1d99644a778dfb6d0a17d2a4abd906400fe8209e611ebf8d7a9abe7ccf28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://97cb59adc2ddfa365ebc9dec649600dc90466ffe59134a6d15a9d3134a7c9b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d2808bc658ea80d68469ed16422d56ef19447f4c90b4af7b2597c64406f8a44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://83acbc20919f7c1958b6d492fadb203855df02761662425df9f91a0757fe2bf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83acbc20919f7c1958b6d492fadb203855df02761662425df9f91a0757fe2bf7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e86ecd03835efc9a253b6804bc61e6ac690e75b8d1f3cc77f5647730e586574\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e86ecd03835efc9a253b6804bc61e6ac690e75b8d1f3cc77f5647730e586574\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://48f95af51feaf014a18496bdfadd59d7f1e11167c5c8b677fa7037e48368ff2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48f95af51feaf014a18496bdfadd59d7f1e11167c5c8b677fa7037e48368ff2a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:20Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:20 crc kubenswrapper[4885]: I1002 01:47:20.600332 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5c1f95f188569c9aa47e51f2916939fa0b65a5cdf4c16b798612d41f3c8411a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:20Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:20 crc kubenswrapper[4885]: I1002 01:47:20.605299 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:20 crc kubenswrapper[4885]: I1002 01:47:20.605342 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:20 crc kubenswrapper[4885]: I1002 01:47:20.605359 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:20 crc kubenswrapper[4885]: I1002 01:47:20.605382 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:20 crc kubenswrapper[4885]: I1002 01:47:20.605399 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:20Z","lastTransitionTime":"2025-10-02T01:47:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:20 crc kubenswrapper[4885]: I1002 01:47:20.619883 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vhgbd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c63a7d20d39b65acfb5a9be4c28b21c72111b398bea76c238e15ddb4dd63e139\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7zwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vhgbd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:20Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:20 crc kubenswrapper[4885]: I1002 01:47:20.638572 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:20Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:20 crc kubenswrapper[4885]: I1002 01:47:20.662399 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74adcf912ccedb6b5d4f2803fd215e3e2ef62367597aa6da528354c4511ee536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7403ef968d040ac09d722b7557438bef23892d578992826f42b700b370d495d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:20Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:20 crc kubenswrapper[4885]: I1002 01:47:20.685677 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2ll9b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52b39712-5fa1-4fe0-814b-b170a6e3938b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c979b91ba748aa4647826a8926337662ca929c54090fb62285572819ec2fbf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c979b91ba748aa4647826a8926337662ca929c54090fb62285572819ec2fbf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97d9d693e8e60ee9c183bb44461e698e8b3cfd20a516605aa8754c9205d2bdc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://97d9d693e8e60ee9c183bb44461e698e8b3cfd20a516605aa8754c9205d2bdc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c44252676359c671d65c9588b05572d76c99cf4c90666f0ac089b2f778450a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c44252676359c671d65c9588b05572d76c99cf4c90666f0ac089b2f778450a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80c982f07874df686fe7e29cc91043eb6d58f1be5fc93cacd46d33a21dd9b622\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80c982f07874df686fe7e29cc91043eb6d58f1be5fc93cacd46d33a21dd9b622\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a354c985652c7ad7fa8944fbfdfe223cfb0f97fb11eecd9bd429020d4892b78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a354c985652c7ad7fa8944fbfdfe223cfb0f97fb11eecd9bd429020d4892b78\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2efd7283df313d086c1b8b88f289ace28764de256205bee7610d6711002d8515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2efd7283df313d086c1b8b88f289ace28764de256205bee7610d6711002d8515\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2ll9b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:20Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:20 crc kubenswrapper[4885]: I1002 01:47:20.708242 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:20 crc kubenswrapper[4885]: I1002 01:47:20.708330 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:20 crc kubenswrapper[4885]: I1002 01:47:20.708348 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:20 crc kubenswrapper[4885]: I1002 01:47:20.708395 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:20 crc kubenswrapper[4885]: I1002 01:47:20.708415 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:20Z","lastTransitionTime":"2025-10-02T01:47:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:20 crc kubenswrapper[4885]: I1002 01:47:20.811222 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:20 crc kubenswrapper[4885]: I1002 01:47:20.811307 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:20 crc kubenswrapper[4885]: I1002 01:47:20.811326 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:20 crc kubenswrapper[4885]: I1002 01:47:20.811352 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:20 crc kubenswrapper[4885]: I1002 01:47:20.811369 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:20Z","lastTransitionTime":"2025-10-02T01:47:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:20 crc kubenswrapper[4885]: I1002 01:47:20.914779 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:20 crc kubenswrapper[4885]: I1002 01:47:20.914836 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:20 crc kubenswrapper[4885]: I1002 01:47:20.914853 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:20 crc kubenswrapper[4885]: I1002 01:47:20.914878 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:20 crc kubenswrapper[4885]: I1002 01:47:20.914896 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:20Z","lastTransitionTime":"2025-10-02T01:47:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:21 crc kubenswrapper[4885]: I1002 01:47:21.017297 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:21 crc kubenswrapper[4885]: I1002 01:47:21.017359 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:21 crc kubenswrapper[4885]: I1002 01:47:21.017378 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:21 crc kubenswrapper[4885]: I1002 01:47:21.017403 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:21 crc kubenswrapper[4885]: I1002 01:47:21.017422 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:21Z","lastTransitionTime":"2025-10-02T01:47:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:21 crc kubenswrapper[4885]: I1002 01:47:21.120460 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:21 crc kubenswrapper[4885]: I1002 01:47:21.120529 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:21 crc kubenswrapper[4885]: I1002 01:47:21.120547 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:21 crc kubenswrapper[4885]: I1002 01:47:21.120572 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:21 crc kubenswrapper[4885]: I1002 01:47:21.120589 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:21Z","lastTransitionTime":"2025-10-02T01:47:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:21 crc kubenswrapper[4885]: I1002 01:47:21.224248 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:21 crc kubenswrapper[4885]: I1002 01:47:21.224348 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:21 crc kubenswrapper[4885]: I1002 01:47:21.224367 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:21 crc kubenswrapper[4885]: I1002 01:47:21.224392 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:21 crc kubenswrapper[4885]: I1002 01:47:21.224410 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:21Z","lastTransitionTime":"2025-10-02T01:47:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:21 crc kubenswrapper[4885]: I1002 01:47:21.327614 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:21 crc kubenswrapper[4885]: I1002 01:47:21.327688 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:21 crc kubenswrapper[4885]: I1002 01:47:21.327714 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:21 crc kubenswrapper[4885]: I1002 01:47:21.327744 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:21 crc kubenswrapper[4885]: I1002 01:47:21.327768 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:21Z","lastTransitionTime":"2025-10-02T01:47:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:21 crc kubenswrapper[4885]: I1002 01:47:21.358624 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" event={"ID":"9e4679dd-f870-41e9-a32b-360f5d2ee81b","Type":"ContainerStarted","Data":"5082a806f0c20edb9171182c100dd73895a0e51869f8252c191811e678795463"} Oct 02 01:47:21 crc kubenswrapper[4885]: I1002 01:47:21.359053 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" Oct 02 01:47:21 crc kubenswrapper[4885]: I1002 01:47:21.359104 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" Oct 02 01:47:21 crc kubenswrapper[4885]: I1002 01:47:21.368127 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2ll9b" event={"ID":"52b39712-5fa1-4fe0-814b-b170a6e3938b","Type":"ContainerStarted","Data":"b9a96383d0fbfd116c419054aaf1a7adba1937e8b0bad294cefd68b4ab0d2298"} Oct 02 01:47:21 crc kubenswrapper[4885]: I1002 01:47:21.380312 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"506a599f-fb1c-4d14-be88-0e4833f3d624\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31c2f0193ffef54965b79a3569b0f6f4547e540bac0774155159a1c996c196c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://533e73e5f90b7258acbcdf018ee01323c118f9dc45035b25b065ba01cabecccf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c00db833a593ec40eb63a6b383455e19e960c2e98b9b66f76877ee993057e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21e2f462a0deb0d740185663b2014f4a1e57b4fdbeb82571264b002f928035bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:21Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:21 crc kubenswrapper[4885]: I1002 01:47:21.398593 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" Oct 02 01:47:21 crc kubenswrapper[4885]: I1002 01:47:21.399930 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" Oct 02 01:47:21 crc kubenswrapper[4885]: I1002 01:47:21.413937 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e4679dd-f870-41e9-a32b-360f5d2ee81b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4310909ad9c7b609ed6814e04a525a6afae747775598e7275080413dec8b23a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7642033b0b5995d332ccf7992d6d7e6a5e9c9369ae2d1071ccb2f536938c8f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9f99f9e26a04fb42b01a0783a5c4e360514818ce9102edc7fa08626d50b9a34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6fa72b030d309656b73121e4ac2a3924a4e449aad66683dd9de5ee5f22d5a4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef403e519d197fb37b3e6b76a14a5128ba3f30a7a5e8bf0a420cf6a39b89575f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8482ba457683ffa03100ec1f07ea9d3f918be5e490b5ba05c2ba965d03033191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5082a806f0c20edb9171182c100dd73895a0e51869f8252c191811e678795463\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fafdd1094bd02786d745c984989852b0c690db38655f92b419733a2004f417c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0125e907d061a97286026ee823c4eed303734ed03ab4864f742592b23a1aa2f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0125e907d061a97286026ee823c4eed303734ed03ab4864f742592b23a1aa2f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dngcm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:21Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:21 crc kubenswrapper[4885]: I1002 01:47:21.431643 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:21 crc kubenswrapper[4885]: I1002 01:47:21.431703 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:21 crc kubenswrapper[4885]: I1002 01:47:21.431721 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:21 crc kubenswrapper[4885]: I1002 01:47:21.431747 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:21 crc kubenswrapper[4885]: I1002 01:47:21.431764 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:21Z","lastTransitionTime":"2025-10-02T01:47:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:21 crc kubenswrapper[4885]: I1002 01:47:21.450560 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8070bc1c-e3de-4e9a-985e-40d1a9858056\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3e634b3727c517cc48b17c2cfb38f26460f53d1bf25d631d08ec9acb8d780e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4adcc537c23697a324d82af94c31059dd444245efdf7c978db58ede256e112ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edcd1d99644a778dfb6d0a17d2a4abd906400fe8209e611ebf8d7a9abe7ccf28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://97cb59adc2ddfa365ebc9dec649600dc90466ffe59134a6d15a9d3134a7c9b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d2808bc658ea80d68469ed16422d56ef19447f4c90b4af7b2597c64406f8a44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://83acbc20919f7c1958b6d492fadb203855df02761662425df9f91a0757fe2bf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83acbc20919f7c1958b6d492fadb203855df02761662425df9f91a0757fe2bf7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e86ecd03835efc9a253b6804bc61e6ac690e75b8d1f3cc77f5647730e586574\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e86ecd03835efc9a253b6804bc61e6ac690e75b8d1f3cc77f5647730e586574\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://48f95af51feaf014a18496bdfadd59d7f1e11167c5c8b677fa7037e48368ff2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48f95af51feaf014a18496bdfadd59d7f1e11167c5c8b677fa7037e48368ff2a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:21Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:21 crc kubenswrapper[4885]: I1002 01:47:21.468383 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5c1f95f188569c9aa47e51f2916939fa0b65a5cdf4c16b798612d41f3c8411a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:21Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:21 crc kubenswrapper[4885]: I1002 01:47:21.487688 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vhgbd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c63a7d20d39b65acfb5a9be4c28b21c72111b398bea76c238e15ddb4dd63e139\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7zwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vhgbd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:21Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:21 crc kubenswrapper[4885]: I1002 01:47:21.507764 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:21Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:21 crc kubenswrapper[4885]: I1002 01:47:21.528080 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74adcf912ccedb6b5d4f2803fd215e3e2ef62367597aa6da528354c4511ee536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7403ef968d040ac09d722b7557438bef23892d578992826f42b700b370d495d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:21Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:21 crc kubenswrapper[4885]: I1002 01:47:21.535047 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:21 crc kubenswrapper[4885]: I1002 01:47:21.535117 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:21 crc kubenswrapper[4885]: I1002 01:47:21.535139 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:21 crc kubenswrapper[4885]: I1002 01:47:21.535168 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:21 crc kubenswrapper[4885]: I1002 01:47:21.535185 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:21Z","lastTransitionTime":"2025-10-02T01:47:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:21 crc kubenswrapper[4885]: I1002 01:47:21.550345 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2ll9b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52b39712-5fa1-4fe0-814b-b170a6e3938b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c979b91ba748aa4647826a8926337662ca929c54090fb62285572819ec2fbf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c979b91ba748aa4647826a8926337662ca929c54090fb62285572819ec2fbf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97d9d693e8e60ee9c183bb44461e698e8b3cfd20a516605aa8754c9205d2bdc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://97d9d693e8e60ee9c183bb44461e698e8b3cfd20a516605aa8754c9205d2bdc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c44252676359c671d65c9588b05572d76c99cf4c90666f0ac089b2f778450a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c44252676359c671d65c9588b05572d76c99cf4c90666f0ac089b2f778450a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80c982f07874df686fe7e29cc91043eb6d58f1be5fc93cacd46d33a21dd9b622\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80c982f07874df686fe7e29cc91043eb6d58f1be5fc93cacd46d33a21dd9b622\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a354c985652c7ad7fa8944fbfdfe223cfb0f97fb11eecd9bd429020d4892b78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a354c985652c7ad7fa8944fbfdfe223cfb0f97fb11eecd9bd429020d4892b78\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2efd7283df313d086c1b8b88f289ace28764de256205bee7610d6711002d8515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2efd7283df313d086c1b8b88f289ace28764de256205bee7610d6711002d8515\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2ll9b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:21Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:21 crc kubenswrapper[4885]: I1002 01:47:21.566639 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-58lhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac6a19c4-f259-4dbd-a016-d6c2c849c6cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed76c0f968c063f7795d363af194a222487b647ceda234de0dabb39b25eafe0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hzc4c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-58lhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:21Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:21 crc kubenswrapper[4885]: I1002 01:47:21.583367 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d169c8ba-000a-4218-8e3b-4ae53035b110\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61e436b7876107ce5ea9e668c0bbc367be6e663fc804fa3a65b51adf8be8d51f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-77rzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09536e182a46e77318c10812d1777e551f834907c006cab5c726dc66f5bce88a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-77rzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rttx8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:21Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:21 crc kubenswrapper[4885]: I1002 01:47:21.598712 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jgdvt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81733e8d-67b0-46fa-8a41-08b40645bd3e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcef60c816ca1ba9a298da658e84c4d717bea3d352c4c2f807ebd8e56327b384\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cwbl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jgdvt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:21Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:21 crc kubenswrapper[4885]: I1002 01:47:21.619936 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a0bb885-2eb0-419c-8a7a-ef634cb3c144\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a89b49cbe08d927629564e607de13655ddf3d01ac4eb5d688b325210306c120c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06c5942805a498d9621a906088aa54538b4eed583d1b8d0e3ba2a6718378a36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b9f6f7fe44d5771a838d0b89c5bdea3ebb79023ae2675ee907ec2766b3d35a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://178b4340ee6904c60c814b99d83ee225161aeb23e438d8d02f856579a13ec4e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bb63bf30d6ecce9e71f0db0c16632bdfa6b33924b673378eb0c38ab229fa837\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T01:47:06Z\\\",\\\"message\\\":\\\"W1002 01:46:55.406027 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1002 01:46:55.406714 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759369615 cert, and key in /tmp/serving-cert-3795040473/serving-signer.crt, /tmp/serving-cert-3795040473/serving-signer.key\\\\nI1002 01:46:55.614670 1 observer_polling.go:159] Starting file observer\\\\nW1002 01:46:55.617662 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1002 01:46:55.618112 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 01:46:55.620347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3795040473/tls.crt::/tmp/serving-cert-3795040473/tls.key\\\\\\\"\\\\nF1002 01:47:06.224459 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90535f97ab2ed4626d47540ebd7baada837fd332140323bec0f49aa38d3c0009\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://592f89a326c9c1559193a71904b47ca27b1dbe1a7b30e03bc48924bb2f35f921\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://592f89a326c9c1559193a71904b47ca27b1dbe1a7b30e03bc48924bb2f35f921\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:21Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:21 crc kubenswrapper[4885]: I1002 01:47:21.637586 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:21 crc kubenswrapper[4885]: I1002 01:47:21.637651 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:21 crc kubenswrapper[4885]: I1002 01:47:21.637757 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:21 crc kubenswrapper[4885]: I1002 01:47:21.637852 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:21 crc kubenswrapper[4885]: I1002 01:47:21.637876 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:21Z","lastTransitionTime":"2025-10-02T01:47:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:21 crc kubenswrapper[4885]: I1002 01:47:21.639936 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b087880e05039816dcab154643ac72c7e4a0a1124817c51dc2c639b62bbee0fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:21Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:21 crc kubenswrapper[4885]: I1002 01:47:21.660327 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:21Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:21 crc kubenswrapper[4885]: I1002 01:47:21.678954 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:21Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:21 crc kubenswrapper[4885]: I1002 01:47:21.711140 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8070bc1c-e3de-4e9a-985e-40d1a9858056\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3e634b3727c517cc48b17c2cfb38f26460f53d1bf25d631d08ec9acb8d780e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4adcc537c23697a324d82af94c31059dd444245efdf7c978db58ede256e112ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edcd1d99644a778dfb6d0a17d2a4abd906400fe8209e611ebf8d7a9abe7ccf28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://97cb59adc2ddfa365ebc9dec649600dc90466ffe59134a6d15a9d3134a7c9b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d2808bc658ea80d68469ed16422d56ef19447f4c90b4af7b2597c64406f8a44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://83acbc20919f7c1958b6d492fadb203855df02761662425df9f91a0757fe2bf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83acbc20919f7c1958b6d492fadb203855df02761662425df9f91a0757fe2bf7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e86ecd03835efc9a253b6804bc61e6ac690e75b8d1f3cc77f5647730e586574\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e86ecd03835efc9a253b6804bc61e6ac690e75b8d1f3cc77f5647730e586574\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://48f95af51feaf014a18496bdfadd59d7f1e11167c5c8b677fa7037e48368ff2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48f95af51feaf014a18496bdfadd59d7f1e11167c5c8b677fa7037e48368ff2a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:21Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:21 crc kubenswrapper[4885]: I1002 01:47:21.730052 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5c1f95f188569c9aa47e51f2916939fa0b65a5cdf4c16b798612d41f3c8411a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:21Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:21 crc kubenswrapper[4885]: I1002 01:47:21.741414 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:21 crc kubenswrapper[4885]: I1002 01:47:21.741478 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:21 crc kubenswrapper[4885]: I1002 01:47:21.741501 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:21 crc kubenswrapper[4885]: I1002 01:47:21.741527 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:21 crc kubenswrapper[4885]: I1002 01:47:21.741545 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:21Z","lastTransitionTime":"2025-10-02T01:47:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:21 crc kubenswrapper[4885]: I1002 01:47:21.753385 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vhgbd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c63a7d20d39b65acfb5a9be4c28b21c72111b398bea76c238e15ddb4dd63e139\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7zwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vhgbd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:21Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:21 crc kubenswrapper[4885]: I1002 01:47:21.773071 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:21Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:21 crc kubenswrapper[4885]: I1002 01:47:21.791785 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74adcf912ccedb6b5d4f2803fd215e3e2ef62367597aa6da528354c4511ee536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7403ef968d040ac09d722b7557438bef23892d578992826f42b700b370d495d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:21Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:21 crc kubenswrapper[4885]: I1002 01:47:21.813523 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2ll9b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52b39712-5fa1-4fe0-814b-b170a6e3938b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9a96383d0fbfd116c419054aaf1a7adba1937e8b0bad294cefd68b4ab0d2298\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c979b91ba748aa4647826a8926337662ca929c54090fb62285572819ec2fbf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c979b91ba748aa4647826a8926337662ca929c54090fb62285572819ec2fbf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97d9d693e8e60ee9c183bb44461e698e8b3cfd20a516605aa8754c9205d2bdc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://97d9d693e8e60ee9c183bb44461e698e8b3cfd20a516605aa8754c9205d2bdc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c44252676359c671d65c9588b05572d76c99cf4c90666f0ac089b2f778450a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c44252676359c671d65c9588b05572d76c99cf4c90666f0ac089b2f778450a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80c982f07874df686fe7e29cc91043eb6d58f1be5fc93cacd46d33a21dd9b622\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80c982f07874df686fe7e29cc91043eb6d58f1be5fc93cacd46d33a21dd9b622\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a354c985652c7ad7fa8944fbfdfe223cfb0f97fb11eecd9bd429020d4892b78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a354c985652c7ad7fa8944fbfdfe223cfb0f97fb11eecd9bd429020d4892b78\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2efd7283df313d086c1b8b88f289ace28764de256205bee7610d6711002d8515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2efd7283df313d086c1b8b88f289ace28764de256205bee7610d6711002d8515\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2ll9b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:21Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:21 crc kubenswrapper[4885]: I1002 01:47:21.828828 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jgdvt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81733e8d-67b0-46fa-8a41-08b40645bd3e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcef60c816ca1ba9a298da658e84c4d717bea3d352c4c2f807ebd8e56327b384\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cwbl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jgdvt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:21Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:21 crc kubenswrapper[4885]: I1002 01:47:21.844752 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:21 crc kubenswrapper[4885]: I1002 01:47:21.844825 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:21 crc kubenswrapper[4885]: I1002 01:47:21.844851 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:21 crc kubenswrapper[4885]: I1002 01:47:21.844880 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:21 crc kubenswrapper[4885]: I1002 01:47:21.844902 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:21Z","lastTransitionTime":"2025-10-02T01:47:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:21 crc kubenswrapper[4885]: I1002 01:47:21.850239 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a0bb885-2eb0-419c-8a7a-ef634cb3c144\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a89b49cbe08d927629564e607de13655ddf3d01ac4eb5d688b325210306c120c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06c5942805a498d9621a906088aa54538b4eed583d1b8d0e3ba2a6718378a36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b9f6f7fe44d5771a838d0b89c5bdea3ebb79023ae2675ee907ec2766b3d35a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://178b4340ee6904c60c814b99d83ee225161aeb23e438d8d02f856579a13ec4e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bb63bf30d6ecce9e71f0db0c16632bdfa6b33924b673378eb0c38ab229fa837\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T01:47:06Z\\\",\\\"message\\\":\\\"W1002 01:46:55.406027 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1002 01:46:55.406714 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759369615 cert, and key in /tmp/serving-cert-3795040473/serving-signer.crt, /tmp/serving-cert-3795040473/serving-signer.key\\\\nI1002 01:46:55.614670 1 observer_polling.go:159] Starting file observer\\\\nW1002 01:46:55.617662 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1002 01:46:55.618112 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 01:46:55.620347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3795040473/tls.crt::/tmp/serving-cert-3795040473/tls.key\\\\\\\"\\\\nF1002 01:47:06.224459 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90535f97ab2ed4626d47540ebd7baada837fd332140323bec0f49aa38d3c0009\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://592f89a326c9c1559193a71904b47ca27b1dbe1a7b30e03bc48924bb2f35f921\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://592f89a326c9c1559193a71904b47ca27b1dbe1a7b30e03bc48924bb2f35f921\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:21Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:21 crc kubenswrapper[4885]: I1002 01:47:21.869964 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b087880e05039816dcab154643ac72c7e4a0a1124817c51dc2c639b62bbee0fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:21Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:21 crc kubenswrapper[4885]: I1002 01:47:21.889022 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:21Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:21 crc kubenswrapper[4885]: I1002 01:47:21.909426 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:21Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:21 crc kubenswrapper[4885]: I1002 01:47:21.927325 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-58lhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac6a19c4-f259-4dbd-a016-d6c2c849c6cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed76c0f968c063f7795d363af194a222487b647ceda234de0dabb39b25eafe0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hzc4c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-58lhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:21Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:21 crc kubenswrapper[4885]: I1002 01:47:21.947759 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d169c8ba-000a-4218-8e3b-4ae53035b110\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61e436b7876107ce5ea9e668c0bbc367be6e663fc804fa3a65b51adf8be8d51f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-77rzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09536e182a46e77318c10812d1777e551f834907c006cab5c726dc66f5bce88a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-77rzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rttx8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:21Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:21 crc kubenswrapper[4885]: I1002 01:47:21.948939 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:21 crc kubenswrapper[4885]: I1002 01:47:21.949145 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:21 crc kubenswrapper[4885]: I1002 01:47:21.949328 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:21 crc kubenswrapper[4885]: I1002 01:47:21.949474 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:21 crc kubenswrapper[4885]: I1002 01:47:21.949620 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:21Z","lastTransitionTime":"2025-10-02T01:47:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:21 crc kubenswrapper[4885]: I1002 01:47:21.972531 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"506a599f-fb1c-4d14-be88-0e4833f3d624\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31c2f0193ffef54965b79a3569b0f6f4547e540bac0774155159a1c996c196c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://533e73e5f90b7258acbcdf018ee01323c118f9dc45035b25b065ba01cabecccf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c00db833a593ec40eb63a6b383455e19e960c2e98b9b66f76877ee993057e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21e2f462a0deb0d740185663b2014f4a1e57b4fdbeb82571264b002f928035bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:21Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:22 crc kubenswrapper[4885]: I1002 01:47:22.005393 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e4679dd-f870-41e9-a32b-360f5d2ee81b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4310909ad9c7b609ed6814e04a525a6afae747775598e7275080413dec8b23a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7642033b0b5995d332ccf7992d6d7e6a5e9c9369ae2d1071ccb2f536938c8f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9f99f9e26a04fb42b01a0783a5c4e360514818ce9102edc7fa08626d50b9a34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6fa72b030d309656b73121e4ac2a3924a4e449aad66683dd9de5ee5f22d5a4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef403e519d197fb37b3e6b76a14a5128ba3f30a7a5e8bf0a420cf6a39b89575f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8482ba457683ffa03100ec1f07ea9d3f918be5e490b5ba05c2ba965d03033191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5082a806f0c20edb9171182c100dd73895a0e51869f8252c191811e678795463\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fafdd1094bd02786d745c984989852b0c690db38655f92b419733a2004f417c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0125e907d061a97286026ee823c4eed303734ed03ab4864f742592b23a1aa2f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0125e907d061a97286026ee823c4eed303734ed03ab4864f742592b23a1aa2f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dngcm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:22Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:22 crc kubenswrapper[4885]: I1002 01:47:22.045830 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:47:22 crc kubenswrapper[4885]: I1002 01:47:22.045927 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:47:22 crc kubenswrapper[4885]: I1002 01:47:22.045830 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:47:22 crc kubenswrapper[4885]: E1002 01:47:22.046083 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:47:22 crc kubenswrapper[4885]: E1002 01:47:22.046224 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:47:22 crc kubenswrapper[4885]: E1002 01:47:22.046437 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:47:22 crc kubenswrapper[4885]: I1002 01:47:22.053451 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:22 crc kubenswrapper[4885]: I1002 01:47:22.053684 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:22 crc kubenswrapper[4885]: I1002 01:47:22.053829 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:22 crc kubenswrapper[4885]: I1002 01:47:22.053968 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:22 crc kubenswrapper[4885]: I1002 01:47:22.054115 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:22Z","lastTransitionTime":"2025-10-02T01:47:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:22 crc kubenswrapper[4885]: I1002 01:47:22.072657 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:22Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:22 crc kubenswrapper[4885]: I1002 01:47:22.093500 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74adcf912ccedb6b5d4f2803fd215e3e2ef62367597aa6da528354c4511ee536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7403ef968d040ac09d722b7557438bef23892d578992826f42b700b370d495d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:22Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:22 crc kubenswrapper[4885]: I1002 01:47:22.117191 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2ll9b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52b39712-5fa1-4fe0-814b-b170a6e3938b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9a96383d0fbfd116c419054aaf1a7adba1937e8b0bad294cefd68b4ab0d2298\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c979b91ba748aa4647826a8926337662ca929c54090fb62285572819ec2fbf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c979b91ba748aa4647826a8926337662ca929c54090fb62285572819ec2fbf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97d9d693e8e60ee9c183bb44461e698e8b3cfd20a516605aa8754c9205d2bdc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://97d9d693e8e60ee9c183bb44461e698e8b3cfd20a516605aa8754c9205d2bdc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c44252676359c671d65c9588b05572d76c99cf4c90666f0ac089b2f778450a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c44252676359c671d65c9588b05572d76c99cf4c90666f0ac089b2f778450a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80c982f07874df686fe7e29cc91043eb6d58f1be5fc93cacd46d33a21dd9b622\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80c982f07874df686fe7e29cc91043eb6d58f1be5fc93cacd46d33a21dd9b622\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a354c985652c7ad7fa8944fbfdfe223cfb0f97fb11eecd9bd429020d4892b78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a354c985652c7ad7fa8944fbfdfe223cfb0f97fb11eecd9bd429020d4892b78\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2efd7283df313d086c1b8b88f289ace28764de256205bee7610d6711002d8515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2efd7283df313d086c1b8b88f289ace28764de256205bee7610d6711002d8515\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2ll9b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:22Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:22 crc kubenswrapper[4885]: I1002 01:47:22.132694 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-58lhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac6a19c4-f259-4dbd-a016-d6c2c849c6cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed76c0f968c063f7795d363af194a222487b647ceda234de0dabb39b25eafe0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hzc4c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-58lhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:22Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:22 crc kubenswrapper[4885]: I1002 01:47:22.156186 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d169c8ba-000a-4218-8e3b-4ae53035b110\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61e436b7876107ce5ea9e668c0bbc367be6e663fc804fa3a65b51adf8be8d51f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-77rzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09536e182a46e77318c10812d1777e551f834907c006cab5c726dc66f5bce88a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-77rzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rttx8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:22Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:22 crc kubenswrapper[4885]: I1002 01:47:22.156690 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:22 crc kubenswrapper[4885]: I1002 01:47:22.156744 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:22 crc kubenswrapper[4885]: I1002 01:47:22.156759 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:22 crc kubenswrapper[4885]: I1002 01:47:22.156780 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:22 crc kubenswrapper[4885]: I1002 01:47:22.156793 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:22Z","lastTransitionTime":"2025-10-02T01:47:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:22 crc kubenswrapper[4885]: I1002 01:47:22.173375 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jgdvt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81733e8d-67b0-46fa-8a41-08b40645bd3e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcef60c816ca1ba9a298da658e84c4d717bea3d352c4c2f807ebd8e56327b384\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cwbl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jgdvt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:22Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:22 crc kubenswrapper[4885]: I1002 01:47:22.193954 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a0bb885-2eb0-419c-8a7a-ef634cb3c144\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a89b49cbe08d927629564e607de13655ddf3d01ac4eb5d688b325210306c120c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06c5942805a498d9621a906088aa54538b4eed583d1b8d0e3ba2a6718378a36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b9f6f7fe44d5771a838d0b89c5bdea3ebb79023ae2675ee907ec2766b3d35a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://178b4340ee6904c60c814b99d83ee225161aeb23e438d8d02f856579a13ec4e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bb63bf30d6ecce9e71f0db0c16632bdfa6b33924b673378eb0c38ab229fa837\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T01:47:06Z\\\",\\\"message\\\":\\\"W1002 01:46:55.406027 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1002 01:46:55.406714 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759369615 cert, and key in /tmp/serving-cert-3795040473/serving-signer.crt, /tmp/serving-cert-3795040473/serving-signer.key\\\\nI1002 01:46:55.614670 1 observer_polling.go:159] Starting file observer\\\\nW1002 01:46:55.617662 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1002 01:46:55.618112 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 01:46:55.620347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3795040473/tls.crt::/tmp/serving-cert-3795040473/tls.key\\\\\\\"\\\\nF1002 01:47:06.224459 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90535f97ab2ed4626d47540ebd7baada837fd332140323bec0f49aa38d3c0009\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://592f89a326c9c1559193a71904b47ca27b1dbe1a7b30e03bc48924bb2f35f921\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://592f89a326c9c1559193a71904b47ca27b1dbe1a7b30e03bc48924bb2f35f921\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:22Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:22 crc kubenswrapper[4885]: I1002 01:47:22.217541 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b087880e05039816dcab154643ac72c7e4a0a1124817c51dc2c639b62bbee0fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:22Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:22 crc kubenswrapper[4885]: I1002 01:47:22.236972 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:22Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:22 crc kubenswrapper[4885]: I1002 01:47:22.255359 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:22Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:22 crc kubenswrapper[4885]: I1002 01:47:22.259722 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:22 crc kubenswrapper[4885]: I1002 01:47:22.259762 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:22 crc kubenswrapper[4885]: I1002 01:47:22.259774 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:22 crc kubenswrapper[4885]: I1002 01:47:22.259792 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:22 crc kubenswrapper[4885]: I1002 01:47:22.259805 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:22Z","lastTransitionTime":"2025-10-02T01:47:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:22 crc kubenswrapper[4885]: I1002 01:47:22.273703 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"506a599f-fb1c-4d14-be88-0e4833f3d624\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31c2f0193ffef54965b79a3569b0f6f4547e540bac0774155159a1c996c196c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://533e73e5f90b7258acbcdf018ee01323c118f9dc45035b25b065ba01cabecccf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c00db833a593ec40eb63a6b383455e19e960c2e98b9b66f76877ee993057e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21e2f462a0deb0d740185663b2014f4a1e57b4fdbeb82571264b002f928035bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:22Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:22 crc kubenswrapper[4885]: I1002 01:47:22.316007 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e4679dd-f870-41e9-a32b-360f5d2ee81b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4310909ad9c7b609ed6814e04a525a6afae747775598e7275080413dec8b23a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7642033b0b5995d332ccf7992d6d7e6a5e9c9369ae2d1071ccb2f536938c8f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9f99f9e26a04fb42b01a0783a5c4e360514818ce9102edc7fa08626d50b9a34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6fa72b030d309656b73121e4ac2a3924a4e449aad66683dd9de5ee5f22d5a4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef403e519d197fb37b3e6b76a14a5128ba3f30a7a5e8bf0a420cf6a39b89575f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8482ba457683ffa03100ec1f07ea9d3f918be5e490b5ba05c2ba965d03033191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5082a806f0c20edb9171182c100dd73895a0e51869f8252c191811e678795463\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fafdd1094bd02786d745c984989852b0c690db38655f92b419733a2004f417c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0125e907d061a97286026ee823c4eed303734ed03ab4864f742592b23a1aa2f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0125e907d061a97286026ee823c4eed303734ed03ab4864f742592b23a1aa2f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dngcm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:22Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:22 crc kubenswrapper[4885]: I1002 01:47:22.356167 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8070bc1c-e3de-4e9a-985e-40d1a9858056\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3e634b3727c517cc48b17c2cfb38f26460f53d1bf25d631d08ec9acb8d780e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4adcc537c23697a324d82af94c31059dd444245efdf7c978db58ede256e112ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edcd1d99644a778dfb6d0a17d2a4abd906400fe8209e611ebf8d7a9abe7ccf28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://97cb59adc2ddfa365ebc9dec649600dc90466ffe59134a6d15a9d3134a7c9b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d2808bc658ea80d68469ed16422d56ef19447f4c90b4af7b2597c64406f8a44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://83acbc20919f7c1958b6d492fadb203855df02761662425df9f91a0757fe2bf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83acbc20919f7c1958b6d492fadb203855df02761662425df9f91a0757fe2bf7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e86ecd03835efc9a253b6804bc61e6ac690e75b8d1f3cc77f5647730e586574\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e86ecd03835efc9a253b6804bc61e6ac690e75b8d1f3cc77f5647730e586574\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://48f95af51feaf014a18496bdfadd59d7f1e11167c5c8b677fa7037e48368ff2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48f95af51feaf014a18496bdfadd59d7f1e11167c5c8b677fa7037e48368ff2a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:22Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:22 crc kubenswrapper[4885]: I1002 01:47:22.361890 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:22 crc kubenswrapper[4885]: I1002 01:47:22.361947 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:22 crc kubenswrapper[4885]: I1002 01:47:22.361965 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:22 crc kubenswrapper[4885]: I1002 01:47:22.361991 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:22 crc kubenswrapper[4885]: I1002 01:47:22.362009 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:22Z","lastTransitionTime":"2025-10-02T01:47:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:22 crc kubenswrapper[4885]: I1002 01:47:22.370955 4885 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 02 01:47:22 crc kubenswrapper[4885]: I1002 01:47:22.383670 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5c1f95f188569c9aa47e51f2916939fa0b65a5cdf4c16b798612d41f3c8411a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:22Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:22 crc kubenswrapper[4885]: I1002 01:47:22.424716 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vhgbd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c63a7d20d39b65acfb5a9be4c28b21c72111b398bea76c238e15ddb4dd63e139\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7zwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vhgbd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:22Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:22 crc kubenswrapper[4885]: I1002 01:47:22.464532 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:22 crc kubenswrapper[4885]: I1002 01:47:22.464587 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:22 crc kubenswrapper[4885]: I1002 01:47:22.464606 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:22 crc kubenswrapper[4885]: I1002 01:47:22.464629 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:22 crc kubenswrapper[4885]: I1002 01:47:22.464650 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:22Z","lastTransitionTime":"2025-10-02T01:47:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:22 crc kubenswrapper[4885]: I1002 01:47:22.571846 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:22 crc kubenswrapper[4885]: I1002 01:47:22.571917 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:22 crc kubenswrapper[4885]: I1002 01:47:22.571945 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:22 crc kubenswrapper[4885]: I1002 01:47:22.571972 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:22 crc kubenswrapper[4885]: I1002 01:47:22.571991 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:22Z","lastTransitionTime":"2025-10-02T01:47:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:22 crc kubenswrapper[4885]: I1002 01:47:22.674525 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:22 crc kubenswrapper[4885]: I1002 01:47:22.674586 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:22 crc kubenswrapper[4885]: I1002 01:47:22.674606 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:22 crc kubenswrapper[4885]: I1002 01:47:22.674628 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:22 crc kubenswrapper[4885]: I1002 01:47:22.674645 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:22Z","lastTransitionTime":"2025-10-02T01:47:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:22 crc kubenswrapper[4885]: I1002 01:47:22.781695 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:22 crc kubenswrapper[4885]: I1002 01:47:22.781753 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:22 crc kubenswrapper[4885]: I1002 01:47:22.781771 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:22 crc kubenswrapper[4885]: I1002 01:47:22.781796 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:22 crc kubenswrapper[4885]: I1002 01:47:22.781812 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:22Z","lastTransitionTime":"2025-10-02T01:47:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:22 crc kubenswrapper[4885]: I1002 01:47:22.885345 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:22 crc kubenswrapper[4885]: I1002 01:47:22.885444 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:22 crc kubenswrapper[4885]: I1002 01:47:22.885464 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:22 crc kubenswrapper[4885]: I1002 01:47:22.885490 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:22 crc kubenswrapper[4885]: I1002 01:47:22.885508 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:22Z","lastTransitionTime":"2025-10-02T01:47:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:22 crc kubenswrapper[4885]: I1002 01:47:22.988252 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:22 crc kubenswrapper[4885]: I1002 01:47:22.988330 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:22 crc kubenswrapper[4885]: I1002 01:47:22.988349 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:22 crc kubenswrapper[4885]: I1002 01:47:22.988375 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:22 crc kubenswrapper[4885]: I1002 01:47:22.988394 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:22Z","lastTransitionTime":"2025-10-02T01:47:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:23 crc kubenswrapper[4885]: I1002 01:47:23.091608 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:23 crc kubenswrapper[4885]: I1002 01:47:23.091680 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:23 crc kubenswrapper[4885]: I1002 01:47:23.091699 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:23 crc kubenswrapper[4885]: I1002 01:47:23.091723 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:23 crc kubenswrapper[4885]: I1002 01:47:23.091743 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:23Z","lastTransitionTime":"2025-10-02T01:47:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:23 crc kubenswrapper[4885]: I1002 01:47:23.194695 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:23 crc kubenswrapper[4885]: I1002 01:47:23.195068 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:23 crc kubenswrapper[4885]: I1002 01:47:23.195229 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:23 crc kubenswrapper[4885]: I1002 01:47:23.195428 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:23 crc kubenswrapper[4885]: I1002 01:47:23.195583 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:23Z","lastTransitionTime":"2025-10-02T01:47:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:23 crc kubenswrapper[4885]: I1002 01:47:23.299129 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:23 crc kubenswrapper[4885]: I1002 01:47:23.299186 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:23 crc kubenswrapper[4885]: I1002 01:47:23.299204 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:23 crc kubenswrapper[4885]: I1002 01:47:23.299228 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:23 crc kubenswrapper[4885]: I1002 01:47:23.299244 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:23Z","lastTransitionTime":"2025-10-02T01:47:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:23 crc kubenswrapper[4885]: I1002 01:47:23.374001 4885 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 02 01:47:23 crc kubenswrapper[4885]: I1002 01:47:23.401338 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:23 crc kubenswrapper[4885]: I1002 01:47:23.401370 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:23 crc kubenswrapper[4885]: I1002 01:47:23.401379 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:23 crc kubenswrapper[4885]: I1002 01:47:23.401392 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:23 crc kubenswrapper[4885]: I1002 01:47:23.401402 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:23Z","lastTransitionTime":"2025-10-02T01:47:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:23 crc kubenswrapper[4885]: I1002 01:47:23.504946 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:23 crc kubenswrapper[4885]: I1002 01:47:23.505370 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:23 crc kubenswrapper[4885]: I1002 01:47:23.505391 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:23 crc kubenswrapper[4885]: I1002 01:47:23.505415 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:23 crc kubenswrapper[4885]: I1002 01:47:23.505433 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:23Z","lastTransitionTime":"2025-10-02T01:47:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:23 crc kubenswrapper[4885]: I1002 01:47:23.608544 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:23 crc kubenswrapper[4885]: I1002 01:47:23.608612 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:23 crc kubenswrapper[4885]: I1002 01:47:23.608637 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:23 crc kubenswrapper[4885]: I1002 01:47:23.608669 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:23 crc kubenswrapper[4885]: I1002 01:47:23.608687 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:23Z","lastTransitionTime":"2025-10-02T01:47:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:23 crc kubenswrapper[4885]: I1002 01:47:23.711804 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:23 crc kubenswrapper[4885]: I1002 01:47:23.711872 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:23 crc kubenswrapper[4885]: I1002 01:47:23.711891 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:23 crc kubenswrapper[4885]: I1002 01:47:23.711918 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:23 crc kubenswrapper[4885]: I1002 01:47:23.711938 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:23Z","lastTransitionTime":"2025-10-02T01:47:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:23 crc kubenswrapper[4885]: I1002 01:47:23.816551 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:23 crc kubenswrapper[4885]: I1002 01:47:23.816615 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:23 crc kubenswrapper[4885]: I1002 01:47:23.816636 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:23 crc kubenswrapper[4885]: I1002 01:47:23.816661 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:23 crc kubenswrapper[4885]: I1002 01:47:23.816679 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:23Z","lastTransitionTime":"2025-10-02T01:47:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:23 crc kubenswrapper[4885]: I1002 01:47:23.919315 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:23 crc kubenswrapper[4885]: I1002 01:47:23.919388 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:23 crc kubenswrapper[4885]: I1002 01:47:23.919406 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:23 crc kubenswrapper[4885]: I1002 01:47:23.919431 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:23 crc kubenswrapper[4885]: I1002 01:47:23.919449 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:23Z","lastTransitionTime":"2025-10-02T01:47:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:24 crc kubenswrapper[4885]: I1002 01:47:24.021960 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:24 crc kubenswrapper[4885]: I1002 01:47:24.022017 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:24 crc kubenswrapper[4885]: I1002 01:47:24.022040 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:24 crc kubenswrapper[4885]: I1002 01:47:24.022066 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:24 crc kubenswrapper[4885]: I1002 01:47:24.022083 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:24Z","lastTransitionTime":"2025-10-02T01:47:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:24 crc kubenswrapper[4885]: I1002 01:47:24.046535 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:47:24 crc kubenswrapper[4885]: I1002 01:47:24.046580 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:47:24 crc kubenswrapper[4885]: I1002 01:47:24.046711 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:47:24 crc kubenswrapper[4885]: E1002 01:47:24.046920 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:47:24 crc kubenswrapper[4885]: E1002 01:47:24.047140 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:47:24 crc kubenswrapper[4885]: E1002 01:47:24.047416 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:47:24 crc kubenswrapper[4885]: I1002 01:47:24.124873 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:24 crc kubenswrapper[4885]: I1002 01:47:24.124920 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:24 crc kubenswrapper[4885]: I1002 01:47:24.124936 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:24 crc kubenswrapper[4885]: I1002 01:47:24.124960 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:24 crc kubenswrapper[4885]: I1002 01:47:24.124977 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:24Z","lastTransitionTime":"2025-10-02T01:47:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:24 crc kubenswrapper[4885]: I1002 01:47:24.227547 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:24 crc kubenswrapper[4885]: I1002 01:47:24.227600 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:24 crc kubenswrapper[4885]: I1002 01:47:24.227617 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:24 crc kubenswrapper[4885]: I1002 01:47:24.227640 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:24 crc kubenswrapper[4885]: I1002 01:47:24.227657 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:24Z","lastTransitionTime":"2025-10-02T01:47:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:24 crc kubenswrapper[4885]: I1002 01:47:24.330189 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:24 crc kubenswrapper[4885]: I1002 01:47:24.330245 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:24 crc kubenswrapper[4885]: I1002 01:47:24.330287 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:24 crc kubenswrapper[4885]: I1002 01:47:24.330309 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:24 crc kubenswrapper[4885]: I1002 01:47:24.330326 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:24Z","lastTransitionTime":"2025-10-02T01:47:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:24 crc kubenswrapper[4885]: I1002 01:47:24.379615 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dngcm_9e4679dd-f870-41e9-a32b-360f5d2ee81b/ovnkube-controller/0.log" Oct 02 01:47:24 crc kubenswrapper[4885]: I1002 01:47:24.383766 4885 generic.go:334] "Generic (PLEG): container finished" podID="9e4679dd-f870-41e9-a32b-360f5d2ee81b" containerID="5082a806f0c20edb9171182c100dd73895a0e51869f8252c191811e678795463" exitCode=1 Oct 02 01:47:24 crc kubenswrapper[4885]: I1002 01:47:24.383814 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" event={"ID":"9e4679dd-f870-41e9-a32b-360f5d2ee81b","Type":"ContainerDied","Data":"5082a806f0c20edb9171182c100dd73895a0e51869f8252c191811e678795463"} Oct 02 01:47:24 crc kubenswrapper[4885]: I1002 01:47:24.384863 4885 scope.go:117] "RemoveContainer" containerID="5082a806f0c20edb9171182c100dd73895a0e51869f8252c191811e678795463" Oct 02 01:47:24 crc kubenswrapper[4885]: I1002 01:47:24.402148 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:24Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:24 crc kubenswrapper[4885]: I1002 01:47:24.420778 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74adcf912ccedb6b5d4f2803fd215e3e2ef62367597aa6da528354c4511ee536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7403ef968d040ac09d722b7557438bef23892d578992826f42b700b370d495d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:24Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:24 crc kubenswrapper[4885]: I1002 01:47:24.437378 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:24 crc kubenswrapper[4885]: I1002 01:47:24.437468 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:24 crc kubenswrapper[4885]: I1002 01:47:24.437489 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:24 crc kubenswrapper[4885]: I1002 01:47:24.437521 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:24 crc kubenswrapper[4885]: I1002 01:47:24.437544 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:24Z","lastTransitionTime":"2025-10-02T01:47:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:24 crc kubenswrapper[4885]: I1002 01:47:24.443810 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2ll9b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52b39712-5fa1-4fe0-814b-b170a6e3938b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9a96383d0fbfd116c419054aaf1a7adba1937e8b0bad294cefd68b4ab0d2298\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c979b91ba748aa4647826a8926337662ca929c54090fb62285572819ec2fbf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c979b91ba748aa4647826a8926337662ca929c54090fb62285572819ec2fbf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97d9d693e8e60ee9c183bb44461e698e8b3cfd20a516605aa8754c9205d2bdc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://97d9d693e8e60ee9c183bb44461e698e8b3cfd20a516605aa8754c9205d2bdc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c44252676359c671d65c9588b05572d76c99cf4c90666f0ac089b2f778450a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c44252676359c671d65c9588b05572d76c99cf4c90666f0ac089b2f778450a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80c982f07874df686fe7e29cc91043eb6d58f1be5fc93cacd46d33a21dd9b622\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80c982f07874df686fe7e29cc91043eb6d58f1be5fc93cacd46d33a21dd9b622\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a354c985652c7ad7fa8944fbfdfe223cfb0f97fb11eecd9bd429020d4892b78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a354c985652c7ad7fa8944fbfdfe223cfb0f97fb11eecd9bd429020d4892b78\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2efd7283df313d086c1b8b88f289ace28764de256205bee7610d6711002d8515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2efd7283df313d086c1b8b88f289ace28764de256205bee7610d6711002d8515\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2ll9b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:24Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:24 crc kubenswrapper[4885]: I1002 01:47:24.466537 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:24Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:24 crc kubenswrapper[4885]: I1002 01:47:24.485549 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:24Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:24 crc kubenswrapper[4885]: I1002 01:47:24.504911 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-58lhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac6a19c4-f259-4dbd-a016-d6c2c849c6cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed76c0f968c063f7795d363af194a222487b647ceda234de0dabb39b25eafe0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hzc4c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-58lhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:24Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:24 crc kubenswrapper[4885]: I1002 01:47:24.521294 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d169c8ba-000a-4218-8e3b-4ae53035b110\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61e436b7876107ce5ea9e668c0bbc367be6e663fc804fa3a65b51adf8be8d51f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-77rzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09536e182a46e77318c10812d1777e551f834907c006cab5c726dc66f5bce88a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-77rzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rttx8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:24Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:24 crc kubenswrapper[4885]: I1002 01:47:24.536062 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jgdvt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81733e8d-67b0-46fa-8a41-08b40645bd3e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcef60c816ca1ba9a298da658e84c4d717bea3d352c4c2f807ebd8e56327b384\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cwbl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jgdvt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:24Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:24 crc kubenswrapper[4885]: I1002 01:47:24.540881 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:24 crc kubenswrapper[4885]: I1002 01:47:24.540930 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:24 crc kubenswrapper[4885]: I1002 01:47:24.540948 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:24 crc kubenswrapper[4885]: I1002 01:47:24.540973 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:24 crc kubenswrapper[4885]: I1002 01:47:24.541031 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:24Z","lastTransitionTime":"2025-10-02T01:47:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:24 crc kubenswrapper[4885]: I1002 01:47:24.557388 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a0bb885-2eb0-419c-8a7a-ef634cb3c144\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a89b49cbe08d927629564e607de13655ddf3d01ac4eb5d688b325210306c120c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06c5942805a498d9621a906088aa54538b4eed583d1b8d0e3ba2a6718378a36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b9f6f7fe44d5771a838d0b89c5bdea3ebb79023ae2675ee907ec2766b3d35a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://178b4340ee6904c60c814b99d83ee225161aeb23e438d8d02f856579a13ec4e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bb63bf30d6ecce9e71f0db0c16632bdfa6b33924b673378eb0c38ab229fa837\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T01:47:06Z\\\",\\\"message\\\":\\\"W1002 01:46:55.406027 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1002 01:46:55.406714 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759369615 cert, and key in /tmp/serving-cert-3795040473/serving-signer.crt, /tmp/serving-cert-3795040473/serving-signer.key\\\\nI1002 01:46:55.614670 1 observer_polling.go:159] Starting file observer\\\\nW1002 01:46:55.617662 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1002 01:46:55.618112 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 01:46:55.620347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3795040473/tls.crt::/tmp/serving-cert-3795040473/tls.key\\\\\\\"\\\\nF1002 01:47:06.224459 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90535f97ab2ed4626d47540ebd7baada837fd332140323bec0f49aa38d3c0009\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://592f89a326c9c1559193a71904b47ca27b1dbe1a7b30e03bc48924bb2f35f921\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://592f89a326c9c1559193a71904b47ca27b1dbe1a7b30e03bc48924bb2f35f921\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:24Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:24 crc kubenswrapper[4885]: I1002 01:47:24.577973 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b087880e05039816dcab154643ac72c7e4a0a1124817c51dc2c639b62bbee0fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:24Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:24 crc kubenswrapper[4885]: I1002 01:47:24.597738 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"506a599f-fb1c-4d14-be88-0e4833f3d624\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31c2f0193ffef54965b79a3569b0f6f4547e540bac0774155159a1c996c196c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://533e73e5f90b7258acbcdf018ee01323c118f9dc45035b25b065ba01cabecccf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c00db833a593ec40eb63a6b383455e19e960c2e98b9b66f76877ee993057e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21e2f462a0deb0d740185663b2014f4a1e57b4fdbeb82571264b002f928035bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:24Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:24 crc kubenswrapper[4885]: I1002 01:47:24.630296 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e4679dd-f870-41e9-a32b-360f5d2ee81b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4310909ad9c7b609ed6814e04a525a6afae747775598e7275080413dec8b23a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7642033b0b5995d332ccf7992d6d7e6a5e9c9369ae2d1071ccb2f536938c8f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9f99f9e26a04fb42b01a0783a5c4e360514818ce9102edc7fa08626d50b9a34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6fa72b030d309656b73121e4ac2a3924a4e449aad66683dd9de5ee5f22d5a4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef403e519d197fb37b3e6b76a14a5128ba3f30a7a5e8bf0a420cf6a39b89575f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8482ba457683ffa03100ec1f07ea9d3f918be5e490b5ba05c2ba965d03033191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5082a806f0c20edb9171182c100dd73895a0e51869f8252c191811e678795463\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5082a806f0c20edb9171182c100dd73895a0e51869f8252c191811e678795463\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T01:47:23Z\\\",\\\"message\\\":\\\" from k8s.io/client-go/informers/factory.go:160\\\\nI1002 01:47:23.739948 6233 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 01:47:23.740066 6233 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 01:47:23.740783 6233 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1002 01:47:23.740803 6233 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1002 01:47:23.740832 6233 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1002 01:47:23.740854 6233 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1002 01:47:23.740862 6233 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1002 01:47:23.740890 6233 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1002 01:47:23.740899 6233 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1002 01:47:23.740929 6233 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1002 01:47:23.741528 6233 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1002 01:47:23.741591 6233 factory.go:656] Stopping watch factory\\\\nI1002 01:47:23.741615 6233 ovnkube.go:599] Stopped ovnkube\\\\nI1002 0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fafdd1094bd02786d745c984989852b0c690db38655f92b419733a2004f417c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0125e907d061a97286026ee823c4eed303734ed03ab4864f742592b23a1aa2f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0125e907d061a97286026ee823c4eed303734ed03ab4864f742592b23a1aa2f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dngcm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:24Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:24 crc kubenswrapper[4885]: I1002 01:47:24.643568 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:24 crc kubenswrapper[4885]: I1002 01:47:24.643617 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:24 crc kubenswrapper[4885]: I1002 01:47:24.643634 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:24 crc kubenswrapper[4885]: I1002 01:47:24.643658 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:24 crc kubenswrapper[4885]: I1002 01:47:24.643676 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:24Z","lastTransitionTime":"2025-10-02T01:47:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:24 crc kubenswrapper[4885]: I1002 01:47:24.650909 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vhgbd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c63a7d20d39b65acfb5a9be4c28b21c72111b398bea76c238e15ddb4dd63e139\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7zwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vhgbd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:24Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:24 crc kubenswrapper[4885]: I1002 01:47:24.682147 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8070bc1c-e3de-4e9a-985e-40d1a9858056\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3e634b3727c517cc48b17c2cfb38f26460f53d1bf25d631d08ec9acb8d780e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4adcc537c23697a324d82af94c31059dd444245efdf7c978db58ede256e112ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edcd1d99644a778dfb6d0a17d2a4abd906400fe8209e611ebf8d7a9abe7ccf28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://97cb59adc2ddfa365ebc9dec649600dc90466ffe59134a6d15a9d3134a7c9b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d2808bc658ea80d68469ed16422d56ef19447f4c90b4af7b2597c64406f8a44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://83acbc20919f7c1958b6d492fadb203855df02761662425df9f91a0757fe2bf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83acbc20919f7c1958b6d492fadb203855df02761662425df9f91a0757fe2bf7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e86ecd03835efc9a253b6804bc61e6ac690e75b8d1f3cc77f5647730e586574\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e86ecd03835efc9a253b6804bc61e6ac690e75b8d1f3cc77f5647730e586574\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://48f95af51feaf014a18496bdfadd59d7f1e11167c5c8b677fa7037e48368ff2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48f95af51feaf014a18496bdfadd59d7f1e11167c5c8b677fa7037e48368ff2a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:24Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:24 crc kubenswrapper[4885]: I1002 01:47:24.696813 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5c1f95f188569c9aa47e51f2916939fa0b65a5cdf4c16b798612d41f3c8411a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:24Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:24 crc kubenswrapper[4885]: I1002 01:47:24.745663 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:24 crc kubenswrapper[4885]: I1002 01:47:24.745697 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:24 crc kubenswrapper[4885]: I1002 01:47:24.745707 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:24 crc kubenswrapper[4885]: I1002 01:47:24.745723 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:24 crc kubenswrapper[4885]: I1002 01:47:24.745734 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:24Z","lastTransitionTime":"2025-10-02T01:47:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:24 crc kubenswrapper[4885]: I1002 01:47:24.848988 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:24 crc kubenswrapper[4885]: I1002 01:47:24.849039 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:24 crc kubenswrapper[4885]: I1002 01:47:24.849056 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:24 crc kubenswrapper[4885]: I1002 01:47:24.849079 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:24 crc kubenswrapper[4885]: I1002 01:47:24.849096 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:24Z","lastTransitionTime":"2025-10-02T01:47:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:24 crc kubenswrapper[4885]: I1002 01:47:24.953065 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:24 crc kubenswrapper[4885]: I1002 01:47:24.953134 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:24 crc kubenswrapper[4885]: I1002 01:47:24.953151 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:24 crc kubenswrapper[4885]: I1002 01:47:24.953174 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:24 crc kubenswrapper[4885]: I1002 01:47:24.953193 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:24Z","lastTransitionTime":"2025-10-02T01:47:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:25 crc kubenswrapper[4885]: I1002 01:47:25.055542 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:25 crc kubenswrapper[4885]: I1002 01:47:25.055597 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:25 crc kubenswrapper[4885]: I1002 01:47:25.055608 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:25 crc kubenswrapper[4885]: I1002 01:47:25.055625 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:25 crc kubenswrapper[4885]: I1002 01:47:25.055638 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:25Z","lastTransitionTime":"2025-10-02T01:47:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:25 crc kubenswrapper[4885]: I1002 01:47:25.158442 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:25 crc kubenswrapper[4885]: I1002 01:47:25.158504 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:25 crc kubenswrapper[4885]: I1002 01:47:25.158521 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:25 crc kubenswrapper[4885]: I1002 01:47:25.158545 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:25 crc kubenswrapper[4885]: I1002 01:47:25.158562 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:25Z","lastTransitionTime":"2025-10-02T01:47:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:25 crc kubenswrapper[4885]: I1002 01:47:25.261044 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:25 crc kubenswrapper[4885]: I1002 01:47:25.261090 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:25 crc kubenswrapper[4885]: I1002 01:47:25.261102 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:25 crc kubenswrapper[4885]: I1002 01:47:25.261119 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:25 crc kubenswrapper[4885]: I1002 01:47:25.261134 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:25Z","lastTransitionTime":"2025-10-02T01:47:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:25 crc kubenswrapper[4885]: I1002 01:47:25.363872 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:25 crc kubenswrapper[4885]: I1002 01:47:25.363933 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:25 crc kubenswrapper[4885]: I1002 01:47:25.363949 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:25 crc kubenswrapper[4885]: I1002 01:47:25.363973 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:25 crc kubenswrapper[4885]: I1002 01:47:25.363998 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:25Z","lastTransitionTime":"2025-10-02T01:47:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:25 crc kubenswrapper[4885]: I1002 01:47:25.391088 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dngcm_9e4679dd-f870-41e9-a32b-360f5d2ee81b/ovnkube-controller/0.log" Oct 02 01:47:25 crc kubenswrapper[4885]: I1002 01:47:25.395444 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" event={"ID":"9e4679dd-f870-41e9-a32b-360f5d2ee81b","Type":"ContainerStarted","Data":"0f85438a1b57b0a590e4d005978fe70ff8c891980badb20d65e1f0e14080fae6"} Oct 02 01:47:25 crc kubenswrapper[4885]: I1002 01:47:25.395591 4885 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 02 01:47:25 crc kubenswrapper[4885]: I1002 01:47:25.413495 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:25Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:25 crc kubenswrapper[4885]: I1002 01:47:25.428442 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:25Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:25 crc kubenswrapper[4885]: I1002 01:47:25.442189 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-58lhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac6a19c4-f259-4dbd-a016-d6c2c849c6cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed76c0f968c063f7795d363af194a222487b647ceda234de0dabb39b25eafe0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hzc4c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-58lhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:25Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:25 crc kubenswrapper[4885]: I1002 01:47:25.458962 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d169c8ba-000a-4218-8e3b-4ae53035b110\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61e436b7876107ce5ea9e668c0bbc367be6e663fc804fa3a65b51adf8be8d51f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-77rzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09536e182a46e77318c10812d1777e551f834907c006cab5c726dc66f5bce88a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-77rzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rttx8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:25Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:25 crc kubenswrapper[4885]: I1002 01:47:25.467392 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:25 crc kubenswrapper[4885]: I1002 01:47:25.467427 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:25 crc kubenswrapper[4885]: I1002 01:47:25.467444 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:25 crc kubenswrapper[4885]: I1002 01:47:25.467470 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:25 crc kubenswrapper[4885]: I1002 01:47:25.467488 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:25Z","lastTransitionTime":"2025-10-02T01:47:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:25 crc kubenswrapper[4885]: I1002 01:47:25.479968 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jgdvt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81733e8d-67b0-46fa-8a41-08b40645bd3e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcef60c816ca1ba9a298da658e84c4d717bea3d352c4c2f807ebd8e56327b384\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cwbl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jgdvt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:25Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:25 crc kubenswrapper[4885]: I1002 01:47:25.502703 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a0bb885-2eb0-419c-8a7a-ef634cb3c144\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a89b49cbe08d927629564e607de13655ddf3d01ac4eb5d688b325210306c120c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06c5942805a498d9621a906088aa54538b4eed583d1b8d0e3ba2a6718378a36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b9f6f7fe44d5771a838d0b89c5bdea3ebb79023ae2675ee907ec2766b3d35a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://178b4340ee6904c60c814b99d83ee225161aeb23e438d8d02f856579a13ec4e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bb63bf30d6ecce9e71f0db0c16632bdfa6b33924b673378eb0c38ab229fa837\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T01:47:06Z\\\",\\\"message\\\":\\\"W1002 01:46:55.406027 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1002 01:46:55.406714 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759369615 cert, and key in /tmp/serving-cert-3795040473/serving-signer.crt, /tmp/serving-cert-3795040473/serving-signer.key\\\\nI1002 01:46:55.614670 1 observer_polling.go:159] Starting file observer\\\\nW1002 01:46:55.617662 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1002 01:46:55.618112 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 01:46:55.620347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3795040473/tls.crt::/tmp/serving-cert-3795040473/tls.key\\\\\\\"\\\\nF1002 01:47:06.224459 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90535f97ab2ed4626d47540ebd7baada837fd332140323bec0f49aa38d3c0009\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://592f89a326c9c1559193a71904b47ca27b1dbe1a7b30e03bc48924bb2f35f921\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://592f89a326c9c1559193a71904b47ca27b1dbe1a7b30e03bc48924bb2f35f921\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:25Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:25 crc kubenswrapper[4885]: I1002 01:47:25.527725 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b087880e05039816dcab154643ac72c7e4a0a1124817c51dc2c639b62bbee0fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:25Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:25 crc kubenswrapper[4885]: I1002 01:47:25.547088 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"506a599f-fb1c-4d14-be88-0e4833f3d624\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31c2f0193ffef54965b79a3569b0f6f4547e540bac0774155159a1c996c196c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://533e73e5f90b7258acbcdf018ee01323c118f9dc45035b25b065ba01cabecccf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c00db833a593ec40eb63a6b383455e19e960c2e98b9b66f76877ee993057e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21e2f462a0deb0d740185663b2014f4a1e57b4fdbeb82571264b002f928035bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:25Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:25 crc kubenswrapper[4885]: I1002 01:47:25.571561 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:25 crc kubenswrapper[4885]: I1002 01:47:25.571605 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:25 crc kubenswrapper[4885]: I1002 01:47:25.571617 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:25 crc kubenswrapper[4885]: I1002 01:47:25.571633 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:25 crc kubenswrapper[4885]: I1002 01:47:25.571643 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:25Z","lastTransitionTime":"2025-10-02T01:47:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:25 crc kubenswrapper[4885]: I1002 01:47:25.575171 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e4679dd-f870-41e9-a32b-360f5d2ee81b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4310909ad9c7b609ed6814e04a525a6afae747775598e7275080413dec8b23a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7642033b0b5995d332ccf7992d6d7e6a5e9c9369ae2d1071ccb2f536938c8f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9f99f9e26a04fb42b01a0783a5c4e360514818ce9102edc7fa08626d50b9a34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6fa72b030d309656b73121e4ac2a3924a4e449aad66683dd9de5ee5f22d5a4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef403e519d197fb37b3e6b76a14a5128ba3f30a7a5e8bf0a420cf6a39b89575f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8482ba457683ffa03100ec1f07ea9d3f918be5e490b5ba05c2ba965d03033191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f85438a1b57b0a590e4d005978fe70ff8c891980badb20d65e1f0e14080fae6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5082a806f0c20edb9171182c100dd73895a0e51869f8252c191811e678795463\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T01:47:23Z\\\",\\\"message\\\":\\\" from k8s.io/client-go/informers/factory.go:160\\\\nI1002 01:47:23.739948 6233 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 01:47:23.740066 6233 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 01:47:23.740783 6233 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1002 01:47:23.740803 6233 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1002 01:47:23.740832 6233 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1002 01:47:23.740854 6233 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1002 01:47:23.740862 6233 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1002 01:47:23.740890 6233 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1002 01:47:23.740899 6233 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1002 01:47:23.740929 6233 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1002 01:47:23.741528 6233 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1002 01:47:23.741591 6233 factory.go:656] Stopping watch factory\\\\nI1002 01:47:23.741615 6233 ovnkube.go:599] Stopped ovnkube\\\\nI1002 0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:20Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fafdd1094bd02786d745c984989852b0c690db38655f92b419733a2004f417c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0125e907d061a97286026ee823c4eed303734ed03ab4864f742592b23a1aa2f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0125e907d061a97286026ee823c4eed303734ed03ab4864f742592b23a1aa2f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dngcm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:25Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:25 crc kubenswrapper[4885]: I1002 01:47:25.595138 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vhgbd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c63a7d20d39b65acfb5a9be4c28b21c72111b398bea76c238e15ddb4dd63e139\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7zwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vhgbd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:25Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:25 crc kubenswrapper[4885]: I1002 01:47:25.621844 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8070bc1c-e3de-4e9a-985e-40d1a9858056\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3e634b3727c517cc48b17c2cfb38f26460f53d1bf25d631d08ec9acb8d780e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4adcc537c23697a324d82af94c31059dd444245efdf7c978db58ede256e112ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edcd1d99644a778dfb6d0a17d2a4abd906400fe8209e611ebf8d7a9abe7ccf28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://97cb59adc2ddfa365ebc9dec649600dc90466ffe59134a6d15a9d3134a7c9b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d2808bc658ea80d68469ed16422d56ef19447f4c90b4af7b2597c64406f8a44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://83acbc20919f7c1958b6d492fadb203855df02761662425df9f91a0757fe2bf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83acbc20919f7c1958b6d492fadb203855df02761662425df9f91a0757fe2bf7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e86ecd03835efc9a253b6804bc61e6ac690e75b8d1f3cc77f5647730e586574\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e86ecd03835efc9a253b6804bc61e6ac690e75b8d1f3cc77f5647730e586574\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://48f95af51feaf014a18496bdfadd59d7f1e11167c5c8b677fa7037e48368ff2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48f95af51feaf014a18496bdfadd59d7f1e11167c5c8b677fa7037e48368ff2a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:25Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:25 crc kubenswrapper[4885]: I1002 01:47:25.639864 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5c1f95f188569c9aa47e51f2916939fa0b65a5cdf4c16b798612d41f3c8411a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:25Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:25 crc kubenswrapper[4885]: I1002 01:47:25.659721 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:25Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:25 crc kubenswrapper[4885]: I1002 01:47:25.673744 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:25 crc kubenswrapper[4885]: I1002 01:47:25.673800 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:25 crc kubenswrapper[4885]: I1002 01:47:25.673819 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:25 crc kubenswrapper[4885]: I1002 01:47:25.673843 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:25 crc kubenswrapper[4885]: I1002 01:47:25.673861 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:25Z","lastTransitionTime":"2025-10-02T01:47:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:25 crc kubenswrapper[4885]: I1002 01:47:25.681216 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74adcf912ccedb6b5d4f2803fd215e3e2ef62367597aa6da528354c4511ee536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7403ef968d040ac09d722b7557438bef23892d578992826f42b700b370d495d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:25Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:25 crc kubenswrapper[4885]: I1002 01:47:25.701099 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2ll9b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52b39712-5fa1-4fe0-814b-b170a6e3938b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9a96383d0fbfd116c419054aaf1a7adba1937e8b0bad294cefd68b4ab0d2298\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c979b91ba748aa4647826a8926337662ca929c54090fb62285572819ec2fbf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c979b91ba748aa4647826a8926337662ca929c54090fb62285572819ec2fbf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97d9d693e8e60ee9c183bb44461e698e8b3cfd20a516605aa8754c9205d2bdc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://97d9d693e8e60ee9c183bb44461e698e8b3cfd20a516605aa8754c9205d2bdc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c44252676359c671d65c9588b05572d76c99cf4c90666f0ac089b2f778450a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c44252676359c671d65c9588b05572d76c99cf4c90666f0ac089b2f778450a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80c982f07874df686fe7e29cc91043eb6d58f1be5fc93cacd46d33a21dd9b622\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80c982f07874df686fe7e29cc91043eb6d58f1be5fc93cacd46d33a21dd9b622\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a354c985652c7ad7fa8944fbfdfe223cfb0f97fb11eecd9bd429020d4892b78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a354c985652c7ad7fa8944fbfdfe223cfb0f97fb11eecd9bd429020d4892b78\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2efd7283df313d086c1b8b88f289ace28764de256205bee7610d6711002d8515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2efd7283df313d086c1b8b88f289ace28764de256205bee7610d6711002d8515\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2ll9b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:25Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:25 crc kubenswrapper[4885]: I1002 01:47:25.776938 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:25 crc kubenswrapper[4885]: I1002 01:47:25.776996 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:25 crc kubenswrapper[4885]: I1002 01:47:25.777020 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:25 crc kubenswrapper[4885]: I1002 01:47:25.777050 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:25 crc kubenswrapper[4885]: I1002 01:47:25.777071 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:25Z","lastTransitionTime":"2025-10-02T01:47:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:25 crc kubenswrapper[4885]: I1002 01:47:25.880504 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:25 crc kubenswrapper[4885]: I1002 01:47:25.880560 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:25 crc kubenswrapper[4885]: I1002 01:47:25.880577 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:25 crc kubenswrapper[4885]: I1002 01:47:25.880600 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:25 crc kubenswrapper[4885]: I1002 01:47:25.880617 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:25Z","lastTransitionTime":"2025-10-02T01:47:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:25 crc kubenswrapper[4885]: I1002 01:47:25.983881 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:25 crc kubenswrapper[4885]: I1002 01:47:25.983994 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:25 crc kubenswrapper[4885]: I1002 01:47:25.984019 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:25 crc kubenswrapper[4885]: I1002 01:47:25.984049 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:25 crc kubenswrapper[4885]: I1002 01:47:25.984070 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:25Z","lastTransitionTime":"2025-10-02T01:47:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.046491 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.046536 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.046616 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:47:26 crc kubenswrapper[4885]: E1002 01:47:26.046774 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:47:26 crc kubenswrapper[4885]: E1002 01:47:26.046915 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:47:26 crc kubenswrapper[4885]: E1002 01:47:26.047043 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.086916 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.087008 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.087027 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.087051 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.087069 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:26Z","lastTransitionTime":"2025-10-02T01:47:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.189864 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.189944 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.189960 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.189984 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.190001 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:26Z","lastTransitionTime":"2025-10-02T01:47:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.277732 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rrn8b"] Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.278331 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rrn8b" Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.281067 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.286845 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.293632 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.293693 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.293715 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.293744 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.293767 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:26Z","lastTransitionTime":"2025-10-02T01:47:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.301051 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"506a599f-fb1c-4d14-be88-0e4833f3d624\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31c2f0193ffef54965b79a3569b0f6f4547e540bac0774155159a1c996c196c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://533e73e5f90b7258acbcdf018ee01323c118f9dc45035b25b065ba01cabecccf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c00db833a593ec40eb63a6b383455e19e960c2e98b9b66f76877ee993057e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21e2f462a0deb0d740185663b2014f4a1e57b4fdbeb82571264b002f928035bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:26Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.331844 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e4679dd-f870-41e9-a32b-360f5d2ee81b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4310909ad9c7b609ed6814e04a525a6afae747775598e7275080413dec8b23a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7642033b0b5995d332ccf7992d6d7e6a5e9c9369ae2d1071ccb2f536938c8f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9f99f9e26a04fb42b01a0783a5c4e360514818ce9102edc7fa08626d50b9a34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6fa72b030d309656b73121e4ac2a3924a4e449aad66683dd9de5ee5f22d5a4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef403e519d197fb37b3e6b76a14a5128ba3f30a7a5e8bf0a420cf6a39b89575f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8482ba457683ffa03100ec1f07ea9d3f918be5e490b5ba05c2ba965d03033191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f85438a1b57b0a590e4d005978fe70ff8c891980badb20d65e1f0e14080fae6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5082a806f0c20edb9171182c100dd73895a0e51869f8252c191811e678795463\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T01:47:23Z\\\",\\\"message\\\":\\\" from k8s.io/client-go/informers/factory.go:160\\\\nI1002 01:47:23.739948 6233 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 01:47:23.740066 6233 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 01:47:23.740783 6233 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1002 01:47:23.740803 6233 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1002 01:47:23.740832 6233 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1002 01:47:23.740854 6233 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1002 01:47:23.740862 6233 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1002 01:47:23.740890 6233 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1002 01:47:23.740899 6233 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1002 01:47:23.740929 6233 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1002 01:47:23.741528 6233 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1002 01:47:23.741591 6233 factory.go:656] Stopping watch factory\\\\nI1002 01:47:23.741615 6233 ovnkube.go:599] Stopped ovnkube\\\\nI1002 0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:20Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fafdd1094bd02786d745c984989852b0c690db38655f92b419733a2004f417c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0125e907d061a97286026ee823c4eed303734ed03ab4864f742592b23a1aa2f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0125e907d061a97286026ee823c4eed303734ed03ab4864f742592b23a1aa2f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dngcm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:26Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.350621 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rrn8b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52967a77-525c-4537-9b1f-d7bb754a8494\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtr74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtr74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:26Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rrn8b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:26Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.383547 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8070bc1c-e3de-4e9a-985e-40d1a9858056\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3e634b3727c517cc48b17c2cfb38f26460f53d1bf25d631d08ec9acb8d780e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4adcc537c23697a324d82af94c31059dd444245efdf7c978db58ede256e112ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edcd1d99644a778dfb6d0a17d2a4abd906400fe8209e611ebf8d7a9abe7ccf28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://97cb59adc2ddfa365ebc9dec649600dc90466ffe59134a6d15a9d3134a7c9b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d2808bc658ea80d68469ed16422d56ef19447f4c90b4af7b2597c64406f8a44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://83acbc20919f7c1958b6d492fadb203855df02761662425df9f91a0757fe2bf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83acbc20919f7c1958b6d492fadb203855df02761662425df9f91a0757fe2bf7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e86ecd03835efc9a253b6804bc61e6ac690e75b8d1f3cc77f5647730e586574\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e86ecd03835efc9a253b6804bc61e6ac690e75b8d1f3cc77f5647730e586574\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://48f95af51feaf014a18496bdfadd59d7f1e11167c5c8b677fa7037e48368ff2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48f95af51feaf014a18496bdfadd59d7f1e11167c5c8b677fa7037e48368ff2a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:26Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.386006 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/52967a77-525c-4537-9b1f-d7bb754a8494-env-overrides\") pod \"ovnkube-control-plane-749d76644c-rrn8b\" (UID: \"52967a77-525c-4537-9b1f-d7bb754a8494\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rrn8b" Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.386086 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/52967a77-525c-4537-9b1f-d7bb754a8494-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-rrn8b\" (UID: \"52967a77-525c-4537-9b1f-d7bb754a8494\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rrn8b" Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.386146 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/52967a77-525c-4537-9b1f-d7bb754a8494-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-rrn8b\" (UID: \"52967a77-525c-4537-9b1f-d7bb754a8494\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rrn8b" Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.386193 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jtr74\" (UniqueName: \"kubernetes.io/projected/52967a77-525c-4537-9b1f-d7bb754a8494-kube-api-access-jtr74\") pod \"ovnkube-control-plane-749d76644c-rrn8b\" (UID: \"52967a77-525c-4537-9b1f-d7bb754a8494\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rrn8b" Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.396727 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.396804 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.396821 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.396883 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.396902 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:26Z","lastTransitionTime":"2025-10-02T01:47:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.402393 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dngcm_9e4679dd-f870-41e9-a32b-360f5d2ee81b/ovnkube-controller/1.log" Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.403360 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dngcm_9e4679dd-f870-41e9-a32b-360f5d2ee81b/ovnkube-controller/0.log" Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.403545 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5c1f95f188569c9aa47e51f2916939fa0b65a5cdf4c16b798612d41f3c8411a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:26Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.411676 4885 generic.go:334] "Generic (PLEG): container finished" podID="9e4679dd-f870-41e9-a32b-360f5d2ee81b" containerID="0f85438a1b57b0a590e4d005978fe70ff8c891980badb20d65e1f0e14080fae6" exitCode=1 Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.411751 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" event={"ID":"9e4679dd-f870-41e9-a32b-360f5d2ee81b","Type":"ContainerDied","Data":"0f85438a1b57b0a590e4d005978fe70ff8c891980badb20d65e1f0e14080fae6"} Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.411800 4885 scope.go:117] "RemoveContainer" containerID="5082a806f0c20edb9171182c100dd73895a0e51869f8252c191811e678795463" Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.413715 4885 scope.go:117] "RemoveContainer" containerID="0f85438a1b57b0a590e4d005978fe70ff8c891980badb20d65e1f0e14080fae6" Oct 02 01:47:26 crc kubenswrapper[4885]: E1002 01:47:26.414095 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-dngcm_openshift-ovn-kubernetes(9e4679dd-f870-41e9-a32b-360f5d2ee81b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" podUID="9e4679dd-f870-41e9-a32b-360f5d2ee81b" Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.420557 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vhgbd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c63a7d20d39b65acfb5a9be4c28b21c72111b398bea76c238e15ddb4dd63e139\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7zwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vhgbd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:26Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.440966 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:26Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.457741 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74adcf912ccedb6b5d4f2803fd215e3e2ef62367597aa6da528354c4511ee536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7403ef968d040ac09d722b7557438bef23892d578992826f42b700b370d495d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:26Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.478628 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2ll9b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52b39712-5fa1-4fe0-814b-b170a6e3938b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9a96383d0fbfd116c419054aaf1a7adba1937e8b0bad294cefd68b4ab0d2298\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c979b91ba748aa4647826a8926337662ca929c54090fb62285572819ec2fbf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c979b91ba748aa4647826a8926337662ca929c54090fb62285572819ec2fbf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97d9d693e8e60ee9c183bb44461e698e8b3cfd20a516605aa8754c9205d2bdc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://97d9d693e8e60ee9c183bb44461e698e8b3cfd20a516605aa8754c9205d2bdc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c44252676359c671d65c9588b05572d76c99cf4c90666f0ac089b2f778450a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c44252676359c671d65c9588b05572d76c99cf4c90666f0ac089b2f778450a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80c982f07874df686fe7e29cc91043eb6d58f1be5fc93cacd46d33a21dd9b622\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80c982f07874df686fe7e29cc91043eb6d58f1be5fc93cacd46d33a21dd9b622\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a354c985652c7ad7fa8944fbfdfe223cfb0f97fb11eecd9bd429020d4892b78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a354c985652c7ad7fa8944fbfdfe223cfb0f97fb11eecd9bd429020d4892b78\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2efd7283df313d086c1b8b88f289ace28764de256205bee7610d6711002d8515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2efd7283df313d086c1b8b88f289ace28764de256205bee7610d6711002d8515\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2ll9b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:26Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.487191 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/52967a77-525c-4537-9b1f-d7bb754a8494-env-overrides\") pod \"ovnkube-control-plane-749d76644c-rrn8b\" (UID: \"52967a77-525c-4537-9b1f-d7bb754a8494\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rrn8b" Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.487480 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/52967a77-525c-4537-9b1f-d7bb754a8494-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-rrn8b\" (UID: \"52967a77-525c-4537-9b1f-d7bb754a8494\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rrn8b" Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.487806 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/52967a77-525c-4537-9b1f-d7bb754a8494-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-rrn8b\" (UID: \"52967a77-525c-4537-9b1f-d7bb754a8494\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rrn8b" Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.487932 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jtr74\" (UniqueName: \"kubernetes.io/projected/52967a77-525c-4537-9b1f-d7bb754a8494-kube-api-access-jtr74\") pod \"ovnkube-control-plane-749d76644c-rrn8b\" (UID: \"52967a77-525c-4537-9b1f-d7bb754a8494\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rrn8b" Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.488492 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/52967a77-525c-4537-9b1f-d7bb754a8494-env-overrides\") pod \"ovnkube-control-plane-749d76644c-rrn8b\" (UID: \"52967a77-525c-4537-9b1f-d7bb754a8494\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rrn8b" Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.489084 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/52967a77-525c-4537-9b1f-d7bb754a8494-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-rrn8b\" (UID: \"52967a77-525c-4537-9b1f-d7bb754a8494\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rrn8b" Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.492737 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:26Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.493431 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/52967a77-525c-4537-9b1f-d7bb754a8494-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-rrn8b\" (UID: \"52967a77-525c-4537-9b1f-d7bb754a8494\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rrn8b" Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.499419 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.499552 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.499639 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.499724 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.499803 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:26Z","lastTransitionTime":"2025-10-02T01:47:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.504506 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-58lhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac6a19c4-f259-4dbd-a016-d6c2c849c6cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed76c0f968c063f7795d363af194a222487b647ceda234de0dabb39b25eafe0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hzc4c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-58lhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:26Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.515681 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d169c8ba-000a-4218-8e3b-4ae53035b110\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61e436b7876107ce5ea9e668c0bbc367be6e663fc804fa3a65b51adf8be8d51f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-77rzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09536e182a46e77318c10812d1777e551f834907c006cab5c726dc66f5bce88a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-77rzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rttx8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:26Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.521331 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jtr74\" (UniqueName: \"kubernetes.io/projected/52967a77-525c-4537-9b1f-d7bb754a8494-kube-api-access-jtr74\") pod \"ovnkube-control-plane-749d76644c-rrn8b\" (UID: \"52967a77-525c-4537-9b1f-d7bb754a8494\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rrn8b" Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.528541 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jgdvt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81733e8d-67b0-46fa-8a41-08b40645bd3e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcef60c816ca1ba9a298da658e84c4d717bea3d352c4c2f807ebd8e56327b384\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cwbl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jgdvt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:26Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.548373 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a0bb885-2eb0-419c-8a7a-ef634cb3c144\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a89b49cbe08d927629564e607de13655ddf3d01ac4eb5d688b325210306c120c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06c5942805a498d9621a906088aa54538b4eed583d1b8d0e3ba2a6718378a36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b9f6f7fe44d5771a838d0b89c5bdea3ebb79023ae2675ee907ec2766b3d35a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://178b4340ee6904c60c814b99d83ee225161aeb23e438d8d02f856579a13ec4e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bb63bf30d6ecce9e71f0db0c16632bdfa6b33924b673378eb0c38ab229fa837\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T01:47:06Z\\\",\\\"message\\\":\\\"W1002 01:46:55.406027 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1002 01:46:55.406714 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759369615 cert, and key in /tmp/serving-cert-3795040473/serving-signer.crt, /tmp/serving-cert-3795040473/serving-signer.key\\\\nI1002 01:46:55.614670 1 observer_polling.go:159] Starting file observer\\\\nW1002 01:46:55.617662 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1002 01:46:55.618112 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 01:46:55.620347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3795040473/tls.crt::/tmp/serving-cert-3795040473/tls.key\\\\\\\"\\\\nF1002 01:47:06.224459 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90535f97ab2ed4626d47540ebd7baada837fd332140323bec0f49aa38d3c0009\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://592f89a326c9c1559193a71904b47ca27b1dbe1a7b30e03bc48924bb2f35f921\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://592f89a326c9c1559193a71904b47ca27b1dbe1a7b30e03bc48924bb2f35f921\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:26Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.566139 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b087880e05039816dcab154643ac72c7e4a0a1124817c51dc2c639b62bbee0fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:26Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.582208 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:26Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.599228 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74adcf912ccedb6b5d4f2803fd215e3e2ef62367597aa6da528354c4511ee536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7403ef968d040ac09d722b7557438bef23892d578992826f42b700b370d495d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:26Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.602282 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.602469 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.602543 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.602607 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.602674 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:26Z","lastTransitionTime":"2025-10-02T01:47:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.604898 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rrn8b" Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.619004 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2ll9b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52b39712-5fa1-4fe0-814b-b170a6e3938b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9a96383d0fbfd116c419054aaf1a7adba1937e8b0bad294cefd68b4ab0d2298\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c979b91ba748aa4647826a8926337662ca929c54090fb62285572819ec2fbf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c979b91ba748aa4647826a8926337662ca929c54090fb62285572819ec2fbf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97d9d693e8e60ee9c183bb44461e698e8b3cfd20a516605aa8754c9205d2bdc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://97d9d693e8e60ee9c183bb44461e698e8b3cfd20a516605aa8754c9205d2bdc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c44252676359c671d65c9588b05572d76c99cf4c90666f0ac089b2f778450a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c44252676359c671d65c9588b05572d76c99cf4c90666f0ac089b2f778450a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80c982f07874df686fe7e29cc91043eb6d58f1be5fc93cacd46d33a21dd9b622\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80c982f07874df686fe7e29cc91043eb6d58f1be5fc93cacd46d33a21dd9b622\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a354c985652c7ad7fa8944fbfdfe223cfb0f97fb11eecd9bd429020d4892b78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a354c985652c7ad7fa8944fbfdfe223cfb0f97fb11eecd9bd429020d4892b78\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2efd7283df313d086c1b8b88f289ace28764de256205bee7610d6711002d8515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2efd7283df313d086c1b8b88f289ace28764de256205bee7610d6711002d8515\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2ll9b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:26Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:26 crc kubenswrapper[4885]: W1002 01:47:26.627474 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod52967a77_525c_4537_9b1f_d7bb754a8494.slice/crio-4174ff264873a635345158837ee122d98d17f01b394f21439d485047a7129570 WatchSource:0}: Error finding container 4174ff264873a635345158837ee122d98d17f01b394f21439d485047a7129570: Status 404 returned error can't find the container with id 4174ff264873a635345158837ee122d98d17f01b394f21439d485047a7129570 Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.638860 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:26Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.664918 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b087880e05039816dcab154643ac72c7e4a0a1124817c51dc2c639b62bbee0fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:26Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.683561 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:26Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.698280 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:26Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.705423 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.705458 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.705470 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.705491 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.705508 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:26Z","lastTransitionTime":"2025-10-02T01:47:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.711565 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-58lhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac6a19c4-f259-4dbd-a016-d6c2c849c6cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed76c0f968c063f7795d363af194a222487b647ceda234de0dabb39b25eafe0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hzc4c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-58lhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:26Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.725613 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d169c8ba-000a-4218-8e3b-4ae53035b110\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61e436b7876107ce5ea9e668c0bbc367be6e663fc804fa3a65b51adf8be8d51f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-77rzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09536e182a46e77318c10812d1777e551f834907c006cab5c726dc66f5bce88a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-77rzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rttx8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:26Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.736986 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jgdvt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81733e8d-67b0-46fa-8a41-08b40645bd3e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcef60c816ca1ba9a298da658e84c4d717bea3d352c4c2f807ebd8e56327b384\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cwbl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jgdvt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:26Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.752199 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a0bb885-2eb0-419c-8a7a-ef634cb3c144\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a89b49cbe08d927629564e607de13655ddf3d01ac4eb5d688b325210306c120c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06c5942805a498d9621a906088aa54538b4eed583d1b8d0e3ba2a6718378a36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b9f6f7fe44d5771a838d0b89c5bdea3ebb79023ae2675ee907ec2766b3d35a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://178b4340ee6904c60c814b99d83ee225161aeb23e438d8d02f856579a13ec4e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bb63bf30d6ecce9e71f0db0c16632bdfa6b33924b673378eb0c38ab229fa837\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T01:47:06Z\\\",\\\"message\\\":\\\"W1002 01:46:55.406027 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1002 01:46:55.406714 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759369615 cert, and key in /tmp/serving-cert-3795040473/serving-signer.crt, /tmp/serving-cert-3795040473/serving-signer.key\\\\nI1002 01:46:55.614670 1 observer_polling.go:159] Starting file observer\\\\nW1002 01:46:55.617662 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1002 01:46:55.618112 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 01:46:55.620347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3795040473/tls.crt::/tmp/serving-cert-3795040473/tls.key\\\\\\\"\\\\nF1002 01:47:06.224459 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90535f97ab2ed4626d47540ebd7baada837fd332140323bec0f49aa38d3c0009\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://592f89a326c9c1559193a71904b47ca27b1dbe1a7b30e03bc48924bb2f35f921\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://592f89a326c9c1559193a71904b47ca27b1dbe1a7b30e03bc48924bb2f35f921\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:26Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.771047 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e4679dd-f870-41e9-a32b-360f5d2ee81b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4310909ad9c7b609ed6814e04a525a6afae747775598e7275080413dec8b23a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7642033b0b5995d332ccf7992d6d7e6a5e9c9369ae2d1071ccb2f536938c8f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9f99f9e26a04fb42b01a0783a5c4e360514818ce9102edc7fa08626d50b9a34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6fa72b030d309656b73121e4ac2a3924a4e449aad66683dd9de5ee5f22d5a4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef403e519d197fb37b3e6b76a14a5128ba3f30a7a5e8bf0a420cf6a39b89575f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8482ba457683ffa03100ec1f07ea9d3f918be5e490b5ba05c2ba965d03033191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f85438a1b57b0a590e4d005978fe70ff8c891980badb20d65e1f0e14080fae6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5082a806f0c20edb9171182c100dd73895a0e51869f8252c191811e678795463\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T01:47:23Z\\\",\\\"message\\\":\\\" from k8s.io/client-go/informers/factory.go:160\\\\nI1002 01:47:23.739948 6233 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 01:47:23.740066 6233 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 01:47:23.740783 6233 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1002 01:47:23.740803 6233 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1002 01:47:23.740832 6233 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1002 01:47:23.740854 6233 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1002 01:47:23.740862 6233 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1002 01:47:23.740890 6233 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1002 01:47:23.740899 6233 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1002 01:47:23.740929 6233 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1002 01:47:23.741528 6233 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1002 01:47:23.741591 6233 factory.go:656] Stopping watch factory\\\\nI1002 01:47:23.741615 6233 ovnkube.go:599] Stopped ovnkube\\\\nI1002 0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:20Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f85438a1b57b0a590e4d005978fe70ff8c891980badb20d65e1f0e14080fae6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T01:47:25Z\\\",\\\"message\\\":\\\"dressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-api/machine-api-operator-machine-webhook_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-operator-machine-webhook\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.250\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1002 01:47:25.452713 6353 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1002 01:47:25.452745 6353 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-source-55646444c4-trplf] creating logical port openshift-network-diagnostics_network-check-source-55646444c4-trplf for pod on switch crc\\\\nF1002 01:47:25.452781 6353 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fafdd1094bd02786d745c984989852b0c690db38655f92b419733a2004f417c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0125e907d061a97286026ee823c4eed303734ed03ab4864f742592b23a1aa2f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0125e907d061a97286026ee823c4eed303734ed03ab4864f742592b23a1aa2f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dngcm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:26Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.784870 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"506a599f-fb1c-4d14-be88-0e4833f3d624\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31c2f0193ffef54965b79a3569b0f6f4547e540bac0774155159a1c996c196c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://533e73e5f90b7258acbcdf018ee01323c118f9dc45035b25b065ba01cabecccf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c00db833a593ec40eb63a6b383455e19e960c2e98b9b66f76877ee993057e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21e2f462a0deb0d740185663b2014f4a1e57b4fdbeb82571264b002f928035bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:26Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.805201 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8070bc1c-e3de-4e9a-985e-40d1a9858056\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3e634b3727c517cc48b17c2cfb38f26460f53d1bf25d631d08ec9acb8d780e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4adcc537c23697a324d82af94c31059dd444245efdf7c978db58ede256e112ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edcd1d99644a778dfb6d0a17d2a4abd906400fe8209e611ebf8d7a9abe7ccf28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://97cb59adc2ddfa365ebc9dec649600dc90466ffe59134a6d15a9d3134a7c9b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d2808bc658ea80d68469ed16422d56ef19447f4c90b4af7b2597c64406f8a44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://83acbc20919f7c1958b6d492fadb203855df02761662425df9f91a0757fe2bf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83acbc20919f7c1958b6d492fadb203855df02761662425df9f91a0757fe2bf7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e86ecd03835efc9a253b6804bc61e6ac690e75b8d1f3cc77f5647730e586574\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e86ecd03835efc9a253b6804bc61e6ac690e75b8d1f3cc77f5647730e586574\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://48f95af51feaf014a18496bdfadd59d7f1e11167c5c8b677fa7037e48368ff2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48f95af51feaf014a18496bdfadd59d7f1e11167c5c8b677fa7037e48368ff2a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:26Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.808126 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.808200 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.808218 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.808238 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.808252 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:26Z","lastTransitionTime":"2025-10-02T01:47:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.817662 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5c1f95f188569c9aa47e51f2916939fa0b65a5cdf4c16b798612d41f3c8411a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:26Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.836908 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vhgbd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c63a7d20d39b65acfb5a9be4c28b21c72111b398bea76c238e15ddb4dd63e139\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7zwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vhgbd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:26Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.850022 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rrn8b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52967a77-525c-4537-9b1f-d7bb754a8494\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtr74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtr74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:26Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rrn8b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:26Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.910722 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.910757 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.910767 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.910784 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:26 crc kubenswrapper[4885]: I1002 01:47:26.910795 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:26Z","lastTransitionTime":"2025-10-02T01:47:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:27 crc kubenswrapper[4885]: I1002 01:47:27.013480 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:27 crc kubenswrapper[4885]: I1002 01:47:27.013901 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:27 crc kubenswrapper[4885]: I1002 01:47:27.014137 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:27 crc kubenswrapper[4885]: I1002 01:47:27.014381 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:27 crc kubenswrapper[4885]: I1002 01:47:27.014566 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:27Z","lastTransitionTime":"2025-10-02T01:47:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:27 crc kubenswrapper[4885]: I1002 01:47:27.117808 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:27 crc kubenswrapper[4885]: I1002 01:47:27.117866 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:27 crc kubenswrapper[4885]: I1002 01:47:27.117889 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:27 crc kubenswrapper[4885]: I1002 01:47:27.117909 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:27 crc kubenswrapper[4885]: I1002 01:47:27.117929 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:27Z","lastTransitionTime":"2025-10-02T01:47:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:27 crc kubenswrapper[4885]: I1002 01:47:27.221496 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:27 crc kubenswrapper[4885]: I1002 01:47:27.221559 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:27 crc kubenswrapper[4885]: I1002 01:47:27.221577 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:27 crc kubenswrapper[4885]: I1002 01:47:27.221607 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:27 crc kubenswrapper[4885]: I1002 01:47:27.221624 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:27Z","lastTransitionTime":"2025-10-02T01:47:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:27 crc kubenswrapper[4885]: I1002 01:47:27.325399 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:27 crc kubenswrapper[4885]: I1002 01:47:27.325452 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:27 crc kubenswrapper[4885]: I1002 01:47:27.325461 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:27 crc kubenswrapper[4885]: I1002 01:47:27.325480 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:27 crc kubenswrapper[4885]: I1002 01:47:27.325492 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:27Z","lastTransitionTime":"2025-10-02T01:47:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:27 crc kubenswrapper[4885]: I1002 01:47:27.417642 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dngcm_9e4679dd-f870-41e9-a32b-360f5d2ee81b/ovnkube-controller/1.log" Oct 02 01:47:27 crc kubenswrapper[4885]: I1002 01:47:27.429451 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:27 crc kubenswrapper[4885]: I1002 01:47:27.429350 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rrn8b" event={"ID":"52967a77-525c-4537-9b1f-d7bb754a8494","Type":"ContainerStarted","Data":"3e9412788d6df7332592beac48281822b26e1aa19b39addce8e1be556af9d982"} Oct 02 01:47:27 crc kubenswrapper[4885]: I1002 01:47:27.429534 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rrn8b" event={"ID":"52967a77-525c-4537-9b1f-d7bb754a8494","Type":"ContainerStarted","Data":"a3dee451ebe9006bf1c519fed4c65267570b6139d76841858352e087ae78d26a"} Oct 02 01:47:27 crc kubenswrapper[4885]: I1002 01:47:27.429556 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rrn8b" event={"ID":"52967a77-525c-4537-9b1f-d7bb754a8494","Type":"ContainerStarted","Data":"4174ff264873a635345158837ee122d98d17f01b394f21439d485047a7129570"} Oct 02 01:47:27 crc kubenswrapper[4885]: I1002 01:47:27.429505 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:27 crc kubenswrapper[4885]: I1002 01:47:27.429728 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:27 crc kubenswrapper[4885]: I1002 01:47:27.429756 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:27 crc kubenswrapper[4885]: I1002 01:47:27.429773 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:27Z","lastTransitionTime":"2025-10-02T01:47:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:27 crc kubenswrapper[4885]: I1002 01:47:27.450716 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"506a599f-fb1c-4d14-be88-0e4833f3d624\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31c2f0193ffef54965b79a3569b0f6f4547e540bac0774155159a1c996c196c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://533e73e5f90b7258acbcdf018ee01323c118f9dc45035b25b065ba01cabecccf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c00db833a593ec40eb63a6b383455e19e960c2e98b9b66f76877ee993057e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21e2f462a0deb0d740185663b2014f4a1e57b4fdbeb82571264b002f928035bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:27Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:27 crc kubenswrapper[4885]: I1002 01:47:27.479223 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e4679dd-f870-41e9-a32b-360f5d2ee81b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4310909ad9c7b609ed6814e04a525a6afae747775598e7275080413dec8b23a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7642033b0b5995d332ccf7992d6d7e6a5e9c9369ae2d1071ccb2f536938c8f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9f99f9e26a04fb42b01a0783a5c4e360514818ce9102edc7fa08626d50b9a34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6fa72b030d309656b73121e4ac2a3924a4e449aad66683dd9de5ee5f22d5a4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef403e519d197fb37b3e6b76a14a5128ba3f30a7a5e8bf0a420cf6a39b89575f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8482ba457683ffa03100ec1f07ea9d3f918be5e490b5ba05c2ba965d03033191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f85438a1b57b0a590e4d005978fe70ff8c891980badb20d65e1f0e14080fae6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5082a806f0c20edb9171182c100dd73895a0e51869f8252c191811e678795463\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T01:47:23Z\\\",\\\"message\\\":\\\" from k8s.io/client-go/informers/factory.go:160\\\\nI1002 01:47:23.739948 6233 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 01:47:23.740066 6233 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 01:47:23.740783 6233 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1002 01:47:23.740803 6233 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1002 01:47:23.740832 6233 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1002 01:47:23.740854 6233 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1002 01:47:23.740862 6233 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1002 01:47:23.740890 6233 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1002 01:47:23.740899 6233 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1002 01:47:23.740929 6233 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1002 01:47:23.741528 6233 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1002 01:47:23.741591 6233 factory.go:656] Stopping watch factory\\\\nI1002 01:47:23.741615 6233 ovnkube.go:599] Stopped ovnkube\\\\nI1002 0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:20Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f85438a1b57b0a590e4d005978fe70ff8c891980badb20d65e1f0e14080fae6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T01:47:25Z\\\",\\\"message\\\":\\\"dressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-api/machine-api-operator-machine-webhook_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-operator-machine-webhook\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.250\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1002 01:47:25.452713 6353 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1002 01:47:25.452745 6353 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-source-55646444c4-trplf] creating logical port openshift-network-diagnostics_network-check-source-55646444c4-trplf for pod on switch crc\\\\nF1002 01:47:25.452781 6353 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fafdd1094bd02786d745c984989852b0c690db38655f92b419733a2004f417c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0125e907d061a97286026ee823c4eed303734ed03ab4864f742592b23a1aa2f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0125e907d061a97286026ee823c4eed303734ed03ab4864f742592b23a1aa2f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dngcm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:27Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:27 crc kubenswrapper[4885]: I1002 01:47:27.509917 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8070bc1c-e3de-4e9a-985e-40d1a9858056\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3e634b3727c517cc48b17c2cfb38f26460f53d1bf25d631d08ec9acb8d780e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4adcc537c23697a324d82af94c31059dd444245efdf7c978db58ede256e112ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edcd1d99644a778dfb6d0a17d2a4abd906400fe8209e611ebf8d7a9abe7ccf28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://97cb59adc2ddfa365ebc9dec649600dc90466ffe59134a6d15a9d3134a7c9b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d2808bc658ea80d68469ed16422d56ef19447f4c90b4af7b2597c64406f8a44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://83acbc20919f7c1958b6d492fadb203855df02761662425df9f91a0757fe2bf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83acbc20919f7c1958b6d492fadb203855df02761662425df9f91a0757fe2bf7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e86ecd03835efc9a253b6804bc61e6ac690e75b8d1f3cc77f5647730e586574\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e86ecd03835efc9a253b6804bc61e6ac690e75b8d1f3cc77f5647730e586574\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://48f95af51feaf014a18496bdfadd59d7f1e11167c5c8b677fa7037e48368ff2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48f95af51feaf014a18496bdfadd59d7f1e11167c5c8b677fa7037e48368ff2a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:27Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:27 crc kubenswrapper[4885]: I1002 01:47:27.526190 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5c1f95f188569c9aa47e51f2916939fa0b65a5cdf4c16b798612d41f3c8411a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:27Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:27 crc kubenswrapper[4885]: I1002 01:47:27.532666 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:27 crc kubenswrapper[4885]: I1002 01:47:27.532729 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:27 crc kubenswrapper[4885]: I1002 01:47:27.532747 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:27 crc kubenswrapper[4885]: I1002 01:47:27.532773 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:27 crc kubenswrapper[4885]: I1002 01:47:27.532791 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:27Z","lastTransitionTime":"2025-10-02T01:47:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:27 crc kubenswrapper[4885]: I1002 01:47:27.544162 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vhgbd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c63a7d20d39b65acfb5a9be4c28b21c72111b398bea76c238e15ddb4dd63e139\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7zwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vhgbd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:27Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:27 crc kubenswrapper[4885]: I1002 01:47:27.558180 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rrn8b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52967a77-525c-4537-9b1f-d7bb754a8494\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3dee451ebe9006bf1c519fed4c65267570b6139d76841858352e087ae78d26a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtr74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e9412788d6df7332592beac48281822b26e1aa19b39addce8e1be556af9d982\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtr74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:26Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rrn8b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:27Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:27 crc kubenswrapper[4885]: I1002 01:47:27.573489 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:27Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:27 crc kubenswrapper[4885]: I1002 01:47:27.589203 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74adcf912ccedb6b5d4f2803fd215e3e2ef62367597aa6da528354c4511ee536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7403ef968d040ac09d722b7557438bef23892d578992826f42b700b370d495d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:27Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:27 crc kubenswrapper[4885]: I1002 01:47:27.607609 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2ll9b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52b39712-5fa1-4fe0-814b-b170a6e3938b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9a96383d0fbfd116c419054aaf1a7adba1937e8b0bad294cefd68b4ab0d2298\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c979b91ba748aa4647826a8926337662ca929c54090fb62285572819ec2fbf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c979b91ba748aa4647826a8926337662ca929c54090fb62285572819ec2fbf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97d9d693e8e60ee9c183bb44461e698e8b3cfd20a516605aa8754c9205d2bdc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://97d9d693e8e60ee9c183bb44461e698e8b3cfd20a516605aa8754c9205d2bdc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c44252676359c671d65c9588b05572d76c99cf4c90666f0ac089b2f778450a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c44252676359c671d65c9588b05572d76c99cf4c90666f0ac089b2f778450a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80c982f07874df686fe7e29cc91043eb6d58f1be5fc93cacd46d33a21dd9b622\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80c982f07874df686fe7e29cc91043eb6d58f1be5fc93cacd46d33a21dd9b622\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a354c985652c7ad7fa8944fbfdfe223cfb0f97fb11eecd9bd429020d4892b78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a354c985652c7ad7fa8944fbfdfe223cfb0f97fb11eecd9bd429020d4892b78\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2efd7283df313d086c1b8b88f289ace28764de256205bee7610d6711002d8515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2efd7283df313d086c1b8b88f289ace28764de256205bee7610d6711002d8515\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2ll9b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:27Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:27 crc kubenswrapper[4885]: I1002 01:47:27.621918 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-58lhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac6a19c4-f259-4dbd-a016-d6c2c849c6cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed76c0f968c063f7795d363af194a222487b647ceda234de0dabb39b25eafe0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hzc4c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-58lhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:27Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:27 crc kubenswrapper[4885]: I1002 01:47:27.634683 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d169c8ba-000a-4218-8e3b-4ae53035b110\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61e436b7876107ce5ea9e668c0bbc367be6e663fc804fa3a65b51adf8be8d51f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-77rzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09536e182a46e77318c10812d1777e551f834907c006cab5c726dc66f5bce88a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-77rzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rttx8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:27Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:27 crc kubenswrapper[4885]: I1002 01:47:27.635397 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:27 crc kubenswrapper[4885]: I1002 01:47:27.635581 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:27 crc kubenswrapper[4885]: I1002 01:47:27.635717 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:27 crc kubenswrapper[4885]: I1002 01:47:27.635879 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:27 crc kubenswrapper[4885]: I1002 01:47:27.636021 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:27Z","lastTransitionTime":"2025-10-02T01:47:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:27 crc kubenswrapper[4885]: I1002 01:47:27.646342 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jgdvt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81733e8d-67b0-46fa-8a41-08b40645bd3e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcef60c816ca1ba9a298da658e84c4d717bea3d352c4c2f807ebd8e56327b384\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cwbl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jgdvt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:27Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:27 crc kubenswrapper[4885]: I1002 01:47:27.664865 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a0bb885-2eb0-419c-8a7a-ef634cb3c144\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a89b49cbe08d927629564e607de13655ddf3d01ac4eb5d688b325210306c120c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06c5942805a498d9621a906088aa54538b4eed583d1b8d0e3ba2a6718378a36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b9f6f7fe44d5771a838d0b89c5bdea3ebb79023ae2675ee907ec2766b3d35a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://178b4340ee6904c60c814b99d83ee225161aeb23e438d8d02f856579a13ec4e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bb63bf30d6ecce9e71f0db0c16632bdfa6b33924b673378eb0c38ab229fa837\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T01:47:06Z\\\",\\\"message\\\":\\\"W1002 01:46:55.406027 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1002 01:46:55.406714 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759369615 cert, and key in /tmp/serving-cert-3795040473/serving-signer.crt, /tmp/serving-cert-3795040473/serving-signer.key\\\\nI1002 01:46:55.614670 1 observer_polling.go:159] Starting file observer\\\\nW1002 01:46:55.617662 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1002 01:46:55.618112 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 01:46:55.620347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3795040473/tls.crt::/tmp/serving-cert-3795040473/tls.key\\\\\\\"\\\\nF1002 01:47:06.224459 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90535f97ab2ed4626d47540ebd7baada837fd332140323bec0f49aa38d3c0009\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://592f89a326c9c1559193a71904b47ca27b1dbe1a7b30e03bc48924bb2f35f921\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://592f89a326c9c1559193a71904b47ca27b1dbe1a7b30e03bc48924bb2f35f921\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:27Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:27 crc kubenswrapper[4885]: I1002 01:47:27.681309 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b087880e05039816dcab154643ac72c7e4a0a1124817c51dc2c639b62bbee0fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:27Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:27 crc kubenswrapper[4885]: I1002 01:47:27.697952 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:27Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:27 crc kubenswrapper[4885]: I1002 01:47:27.713450 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:27Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:27 crc kubenswrapper[4885]: I1002 01:47:27.739139 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:27 crc kubenswrapper[4885]: I1002 01:47:27.739222 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:27 crc kubenswrapper[4885]: I1002 01:47:27.739245 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:27 crc kubenswrapper[4885]: I1002 01:47:27.739311 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:27 crc kubenswrapper[4885]: I1002 01:47:27.739343 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:27Z","lastTransitionTime":"2025-10-02T01:47:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:27 crc kubenswrapper[4885]: I1002 01:47:27.801753 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:47:27 crc kubenswrapper[4885]: E1002 01:47:27.802024 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:47:43.801964423 +0000 UTC m=+52.613711852 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:47:27 crc kubenswrapper[4885]: I1002 01:47:27.802539 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:47:27 crc kubenswrapper[4885]: E1002 01:47:27.802731 4885 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 01:47:27 crc kubenswrapper[4885]: E1002 01:47:27.802821 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 01:47:43.802803808 +0000 UTC m=+52.614551247 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 01:47:27 crc kubenswrapper[4885]: I1002 01:47:27.842723 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:27 crc kubenswrapper[4885]: I1002 01:47:27.842784 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:27 crc kubenswrapper[4885]: I1002 01:47:27.842802 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:27 crc kubenswrapper[4885]: I1002 01:47:27.842827 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:27 crc kubenswrapper[4885]: I1002 01:47:27.842845 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:27Z","lastTransitionTime":"2025-10-02T01:47:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:27 crc kubenswrapper[4885]: I1002 01:47:27.904604 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:47:27 crc kubenswrapper[4885]: I1002 01:47:27.904706 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:47:27 crc kubenswrapper[4885]: I1002 01:47:27.904763 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:47:27 crc kubenswrapper[4885]: E1002 01:47:27.904867 4885 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 01:47:27 crc kubenswrapper[4885]: E1002 01:47:27.904870 4885 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 01:47:27 crc kubenswrapper[4885]: E1002 01:47:27.904909 4885 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 01:47:27 crc kubenswrapper[4885]: E1002 01:47:27.904930 4885 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 01:47:27 crc kubenswrapper[4885]: E1002 01:47:27.904954 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 01:47:43.904932454 +0000 UTC m=+52.716679893 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 01:47:27 crc kubenswrapper[4885]: E1002 01:47:27.904966 4885 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 01:47:27 crc kubenswrapper[4885]: E1002 01:47:27.905022 4885 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 01:47:27 crc kubenswrapper[4885]: E1002 01:47:27.905043 4885 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 01:47:27 crc kubenswrapper[4885]: E1002 01:47:27.904987 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-02 01:47:43.904968245 +0000 UTC m=+52.716715674 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 01:47:27 crc kubenswrapper[4885]: E1002 01:47:27.905126 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-02 01:47:43.905111789 +0000 UTC m=+52.716859218 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 01:47:27 crc kubenswrapper[4885]: I1002 01:47:27.946007 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:27 crc kubenswrapper[4885]: I1002 01:47:27.946061 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:27 crc kubenswrapper[4885]: I1002 01:47:27.946078 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:27 crc kubenswrapper[4885]: I1002 01:47:27.946102 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:27 crc kubenswrapper[4885]: I1002 01:47:27.946120 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:27Z","lastTransitionTime":"2025-10-02T01:47:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:28 crc kubenswrapper[4885]: I1002 01:47:28.046050 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:47:28 crc kubenswrapper[4885]: I1002 01:47:28.046102 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:47:28 crc kubenswrapper[4885]: I1002 01:47:28.046142 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:47:28 crc kubenswrapper[4885]: E1002 01:47:28.046251 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:47:28 crc kubenswrapper[4885]: E1002 01:47:28.046382 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:47:28 crc kubenswrapper[4885]: E1002 01:47:28.046564 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:47:28 crc kubenswrapper[4885]: I1002 01:47:28.048836 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:28 crc kubenswrapper[4885]: I1002 01:47:28.048884 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:28 crc kubenswrapper[4885]: I1002 01:47:28.048902 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:28 crc kubenswrapper[4885]: I1002 01:47:28.048926 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:28 crc kubenswrapper[4885]: I1002 01:47:28.048943 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:28Z","lastTransitionTime":"2025-10-02T01:47:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:28 crc kubenswrapper[4885]: I1002 01:47:28.151582 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:28 crc kubenswrapper[4885]: I1002 01:47:28.151617 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:28 crc kubenswrapper[4885]: I1002 01:47:28.151626 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:28 crc kubenswrapper[4885]: I1002 01:47:28.151640 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:28 crc kubenswrapper[4885]: I1002 01:47:28.151650 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:28Z","lastTransitionTime":"2025-10-02T01:47:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:28 crc kubenswrapper[4885]: I1002 01:47:28.184054 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-sv4md"] Oct 02 01:47:28 crc kubenswrapper[4885]: I1002 01:47:28.184797 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sv4md" Oct 02 01:47:28 crc kubenswrapper[4885]: E1002 01:47:28.184900 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sv4md" podUID="724a3e49-001c-45a4-a896-82c13e0ee0e7" Oct 02 01:47:28 crc kubenswrapper[4885]: I1002 01:47:28.197173 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-58lhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac6a19c4-f259-4dbd-a016-d6c2c849c6cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed76c0f968c063f7795d363af194a222487b647ceda234de0dabb39b25eafe0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hzc4c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-58lhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:28Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:28 crc kubenswrapper[4885]: I1002 01:47:28.212909 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d169c8ba-000a-4218-8e3b-4ae53035b110\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61e436b7876107ce5ea9e668c0bbc367be6e663fc804fa3a65b51adf8be8d51f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-77rzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09536e182a46e77318c10812d1777e551f834907c006cab5c726dc66f5bce88a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-77rzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rttx8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:28Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:28 crc kubenswrapper[4885]: I1002 01:47:28.226255 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jgdvt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81733e8d-67b0-46fa-8a41-08b40645bd3e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcef60c816ca1ba9a298da658e84c4d717bea3d352c4c2f807ebd8e56327b384\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cwbl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jgdvt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:28Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:28 crc kubenswrapper[4885]: I1002 01:47:28.235811 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:28 crc kubenswrapper[4885]: I1002 01:47:28.235866 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:28 crc kubenswrapper[4885]: I1002 01:47:28.235889 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:28 crc kubenswrapper[4885]: I1002 01:47:28.235919 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:28 crc kubenswrapper[4885]: I1002 01:47:28.235942 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:28Z","lastTransitionTime":"2025-10-02T01:47:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:28 crc kubenswrapper[4885]: I1002 01:47:28.245824 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a0bb885-2eb0-419c-8a7a-ef634cb3c144\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a89b49cbe08d927629564e607de13655ddf3d01ac4eb5d688b325210306c120c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06c5942805a498d9621a906088aa54538b4eed583d1b8d0e3ba2a6718378a36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b9f6f7fe44d5771a838d0b89c5bdea3ebb79023ae2675ee907ec2766b3d35a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://178b4340ee6904c60c814b99d83ee225161aeb23e438d8d02f856579a13ec4e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bb63bf30d6ecce9e71f0db0c16632bdfa6b33924b673378eb0c38ab229fa837\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T01:47:06Z\\\",\\\"message\\\":\\\"W1002 01:46:55.406027 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1002 01:46:55.406714 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759369615 cert, and key in /tmp/serving-cert-3795040473/serving-signer.crt, /tmp/serving-cert-3795040473/serving-signer.key\\\\nI1002 01:46:55.614670 1 observer_polling.go:159] Starting file observer\\\\nW1002 01:46:55.617662 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1002 01:46:55.618112 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 01:46:55.620347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3795040473/tls.crt::/tmp/serving-cert-3795040473/tls.key\\\\\\\"\\\\nF1002 01:47:06.224459 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90535f97ab2ed4626d47540ebd7baada837fd332140323bec0f49aa38d3c0009\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://592f89a326c9c1559193a71904b47ca27b1dbe1a7b30e03bc48924bb2f35f921\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://592f89a326c9c1559193a71904b47ca27b1dbe1a7b30e03bc48924bb2f35f921\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:28Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:28 crc kubenswrapper[4885]: E1002 01:47:28.259648 4885 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"92603ae5-ba38-4201-98e9-979fd56cb50d\\\",\\\"systemUUID\\\":\\\"50ddf2c2-ea95-4960-ae48-d17e9484ee2f\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:28Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:28 crc kubenswrapper[4885]: I1002 01:47:28.265097 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:28 crc kubenswrapper[4885]: I1002 01:47:28.265221 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:28 crc kubenswrapper[4885]: I1002 01:47:28.265316 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:28 crc kubenswrapper[4885]: I1002 01:47:28.265413 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:28 crc kubenswrapper[4885]: I1002 01:47:28.265455 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:28Z","lastTransitionTime":"2025-10-02T01:47:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:28 crc kubenswrapper[4885]: I1002 01:47:28.273426 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b087880e05039816dcab154643ac72c7e4a0a1124817c51dc2c639b62bbee0fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:28Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:28 crc kubenswrapper[4885]: E1002 01:47:28.284976 4885 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"92603ae5-ba38-4201-98e9-979fd56cb50d\\\",\\\"systemUUID\\\":\\\"50ddf2c2-ea95-4960-ae48-d17e9484ee2f\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:28Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:28 crc kubenswrapper[4885]: I1002 01:47:28.290201 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:28 crc kubenswrapper[4885]: I1002 01:47:28.290309 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:28 crc kubenswrapper[4885]: I1002 01:47:28.290337 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:28 crc kubenswrapper[4885]: I1002 01:47:28.290373 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:28 crc kubenswrapper[4885]: I1002 01:47:28.290400 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:28Z","lastTransitionTime":"2025-10-02T01:47:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:28 crc kubenswrapper[4885]: I1002 01:47:28.292512 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:28Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:28 crc kubenswrapper[4885]: E1002 01:47:28.310008 4885 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"92603ae5-ba38-4201-98e9-979fd56cb50d\\\",\\\"systemUUID\\\":\\\"50ddf2c2-ea95-4960-ae48-d17e9484ee2f\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:28Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:28 crc kubenswrapper[4885]: I1002 01:47:28.310366 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hpqgx\" (UniqueName: \"kubernetes.io/projected/724a3e49-001c-45a4-a896-82c13e0ee0e7-kube-api-access-hpqgx\") pod \"network-metrics-daemon-sv4md\" (UID: \"724a3e49-001c-45a4-a896-82c13e0ee0e7\") " pod="openshift-multus/network-metrics-daemon-sv4md" Oct 02 01:47:28 crc kubenswrapper[4885]: I1002 01:47:28.310538 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/724a3e49-001c-45a4-a896-82c13e0ee0e7-metrics-certs\") pod \"network-metrics-daemon-sv4md\" (UID: \"724a3e49-001c-45a4-a896-82c13e0ee0e7\") " pod="openshift-multus/network-metrics-daemon-sv4md" Oct 02 01:47:28 crc kubenswrapper[4885]: I1002 01:47:28.310904 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:28Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:28 crc kubenswrapper[4885]: I1002 01:47:28.314808 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:28 crc kubenswrapper[4885]: I1002 01:47:28.314863 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:28 crc kubenswrapper[4885]: I1002 01:47:28.314879 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:28 crc kubenswrapper[4885]: I1002 01:47:28.314946 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:28 crc kubenswrapper[4885]: I1002 01:47:28.314968 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:28Z","lastTransitionTime":"2025-10-02T01:47:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:28 crc kubenswrapper[4885]: I1002 01:47:28.327977 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"506a599f-fb1c-4d14-be88-0e4833f3d624\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31c2f0193ffef54965b79a3569b0f6f4547e540bac0774155159a1c996c196c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://533e73e5f90b7258acbcdf018ee01323c118f9dc45035b25b065ba01cabecccf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c00db833a593ec40eb63a6b383455e19e960c2e98b9b66f76877ee993057e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21e2f462a0deb0d740185663b2014f4a1e57b4fdbeb82571264b002f928035bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:28Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:28 crc kubenswrapper[4885]: E1002 01:47:28.332736 4885 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"92603ae5-ba38-4201-98e9-979fd56cb50d\\\",\\\"systemUUID\\\":\\\"50ddf2c2-ea95-4960-ae48-d17e9484ee2f\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:28Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:28 crc kubenswrapper[4885]: I1002 01:47:28.337774 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:28 crc kubenswrapper[4885]: I1002 01:47:28.337836 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:28 crc kubenswrapper[4885]: I1002 01:47:28.337860 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:28 crc kubenswrapper[4885]: I1002 01:47:28.337891 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:28 crc kubenswrapper[4885]: I1002 01:47:28.337913 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:28Z","lastTransitionTime":"2025-10-02T01:47:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:28 crc kubenswrapper[4885]: E1002 01:47:28.358140 4885 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"92603ae5-ba38-4201-98e9-979fd56cb50d\\\",\\\"systemUUID\\\":\\\"50ddf2c2-ea95-4960-ae48-d17e9484ee2f\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:28Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:28 crc kubenswrapper[4885]: E1002 01:47:28.358424 4885 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 02 01:47:28 crc kubenswrapper[4885]: I1002 01:47:28.360697 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:28 crc kubenswrapper[4885]: I1002 01:47:28.360788 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:28 crc kubenswrapper[4885]: I1002 01:47:28.360809 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:28 crc kubenswrapper[4885]: I1002 01:47:28.360832 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:28 crc kubenswrapper[4885]: I1002 01:47:28.360849 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:28Z","lastTransitionTime":"2025-10-02T01:47:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:28 crc kubenswrapper[4885]: I1002 01:47:28.360937 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e4679dd-f870-41e9-a32b-360f5d2ee81b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4310909ad9c7b609ed6814e04a525a6afae747775598e7275080413dec8b23a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7642033b0b5995d332ccf7992d6d7e6a5e9c9369ae2d1071ccb2f536938c8f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9f99f9e26a04fb42b01a0783a5c4e360514818ce9102edc7fa08626d50b9a34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6fa72b030d309656b73121e4ac2a3924a4e449aad66683dd9de5ee5f22d5a4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef403e519d197fb37b3e6b76a14a5128ba3f30a7a5e8bf0a420cf6a39b89575f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8482ba457683ffa03100ec1f07ea9d3f918be5e490b5ba05c2ba965d03033191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f85438a1b57b0a590e4d005978fe70ff8c891980badb20d65e1f0e14080fae6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5082a806f0c20edb9171182c100dd73895a0e51869f8252c191811e678795463\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T01:47:23Z\\\",\\\"message\\\":\\\" from k8s.io/client-go/informers/factory.go:160\\\\nI1002 01:47:23.739948 6233 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 01:47:23.740066 6233 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 01:47:23.740783 6233 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1002 01:47:23.740803 6233 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1002 01:47:23.740832 6233 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1002 01:47:23.740854 6233 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1002 01:47:23.740862 6233 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1002 01:47:23.740890 6233 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1002 01:47:23.740899 6233 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1002 01:47:23.740929 6233 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1002 01:47:23.741528 6233 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1002 01:47:23.741591 6233 factory.go:656] Stopping watch factory\\\\nI1002 01:47:23.741615 6233 ovnkube.go:599] Stopped ovnkube\\\\nI1002 0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:20Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f85438a1b57b0a590e4d005978fe70ff8c891980badb20d65e1f0e14080fae6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T01:47:25Z\\\",\\\"message\\\":\\\"dressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-api/machine-api-operator-machine-webhook_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-operator-machine-webhook\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.250\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1002 01:47:25.452713 6353 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1002 01:47:25.452745 6353 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-source-55646444c4-trplf] creating logical port openshift-network-diagnostics_network-check-source-55646444c4-trplf for pod on switch crc\\\\nF1002 01:47:25.452781 6353 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fafdd1094bd02786d745c984989852b0c690db38655f92b419733a2004f417c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0125e907d061a97286026ee823c4eed303734ed03ab4864f742592b23a1aa2f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0125e907d061a97286026ee823c4eed303734ed03ab4864f742592b23a1aa2f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dngcm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:28Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:28 crc kubenswrapper[4885]: I1002 01:47:28.376937 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-sv4md" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"724a3e49-001c-45a4-a896-82c13e0ee0e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hpqgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hpqgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-sv4md\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:28Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:28 crc kubenswrapper[4885]: I1002 01:47:28.410663 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8070bc1c-e3de-4e9a-985e-40d1a9858056\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3e634b3727c517cc48b17c2cfb38f26460f53d1bf25d631d08ec9acb8d780e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4adcc537c23697a324d82af94c31059dd444245efdf7c978db58ede256e112ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edcd1d99644a778dfb6d0a17d2a4abd906400fe8209e611ebf8d7a9abe7ccf28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://97cb59adc2ddfa365ebc9dec649600dc90466ffe59134a6d15a9d3134a7c9b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d2808bc658ea80d68469ed16422d56ef19447f4c90b4af7b2597c64406f8a44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://83acbc20919f7c1958b6d492fadb203855df02761662425df9f91a0757fe2bf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83acbc20919f7c1958b6d492fadb203855df02761662425df9f91a0757fe2bf7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e86ecd03835efc9a253b6804bc61e6ac690e75b8d1f3cc77f5647730e586574\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e86ecd03835efc9a253b6804bc61e6ac690e75b8d1f3cc77f5647730e586574\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://48f95af51feaf014a18496bdfadd59d7f1e11167c5c8b677fa7037e48368ff2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48f95af51feaf014a18496bdfadd59d7f1e11167c5c8b677fa7037e48368ff2a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:28Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:28 crc kubenswrapper[4885]: I1002 01:47:28.411792 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/724a3e49-001c-45a4-a896-82c13e0ee0e7-metrics-certs\") pod \"network-metrics-daemon-sv4md\" (UID: \"724a3e49-001c-45a4-a896-82c13e0ee0e7\") " pod="openshift-multus/network-metrics-daemon-sv4md" Oct 02 01:47:28 crc kubenswrapper[4885]: I1002 01:47:28.411937 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hpqgx\" (UniqueName: \"kubernetes.io/projected/724a3e49-001c-45a4-a896-82c13e0ee0e7-kube-api-access-hpqgx\") pod \"network-metrics-daemon-sv4md\" (UID: \"724a3e49-001c-45a4-a896-82c13e0ee0e7\") " pod="openshift-multus/network-metrics-daemon-sv4md" Oct 02 01:47:28 crc kubenswrapper[4885]: E1002 01:47:28.412061 4885 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 01:47:28 crc kubenswrapper[4885]: E1002 01:47:28.412200 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/724a3e49-001c-45a4-a896-82c13e0ee0e7-metrics-certs podName:724a3e49-001c-45a4-a896-82c13e0ee0e7 nodeName:}" failed. No retries permitted until 2025-10-02 01:47:28.912166352 +0000 UTC m=+37.723913821 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/724a3e49-001c-45a4-a896-82c13e0ee0e7-metrics-certs") pod "network-metrics-daemon-sv4md" (UID: "724a3e49-001c-45a4-a896-82c13e0ee0e7") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 01:47:28 crc kubenswrapper[4885]: I1002 01:47:28.432446 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5c1f95f188569c9aa47e51f2916939fa0b65a5cdf4c16b798612d41f3c8411a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:28Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:28 crc kubenswrapper[4885]: I1002 01:47:28.445187 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hpqgx\" (UniqueName: \"kubernetes.io/projected/724a3e49-001c-45a4-a896-82c13e0ee0e7-kube-api-access-hpqgx\") pod \"network-metrics-daemon-sv4md\" (UID: \"724a3e49-001c-45a4-a896-82c13e0ee0e7\") " pod="openshift-multus/network-metrics-daemon-sv4md" Oct 02 01:47:28 crc kubenswrapper[4885]: I1002 01:47:28.456384 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vhgbd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c63a7d20d39b65acfb5a9be4c28b21c72111b398bea76c238e15ddb4dd63e139\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7zwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vhgbd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:28Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:28 crc kubenswrapper[4885]: I1002 01:47:28.464900 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:28 crc kubenswrapper[4885]: I1002 01:47:28.464947 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:28 crc kubenswrapper[4885]: I1002 01:47:28.464965 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:28 crc kubenswrapper[4885]: I1002 01:47:28.464990 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:28 crc kubenswrapper[4885]: I1002 01:47:28.465007 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:28Z","lastTransitionTime":"2025-10-02T01:47:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:28 crc kubenswrapper[4885]: I1002 01:47:28.477559 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rrn8b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52967a77-525c-4537-9b1f-d7bb754a8494\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3dee451ebe9006bf1c519fed4c65267570b6139d76841858352e087ae78d26a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtr74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e9412788d6df7332592beac48281822b26e1aa19b39addce8e1be556af9d982\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtr74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:26Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rrn8b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:28Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:28 crc kubenswrapper[4885]: I1002 01:47:28.498653 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:28Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:28 crc kubenswrapper[4885]: I1002 01:47:28.521358 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74adcf912ccedb6b5d4f2803fd215e3e2ef62367597aa6da528354c4511ee536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7403ef968d040ac09d722b7557438bef23892d578992826f42b700b370d495d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:28Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:28 crc kubenswrapper[4885]: I1002 01:47:28.545708 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2ll9b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52b39712-5fa1-4fe0-814b-b170a6e3938b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9a96383d0fbfd116c419054aaf1a7adba1937e8b0bad294cefd68b4ab0d2298\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c979b91ba748aa4647826a8926337662ca929c54090fb62285572819ec2fbf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c979b91ba748aa4647826a8926337662ca929c54090fb62285572819ec2fbf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97d9d693e8e60ee9c183bb44461e698e8b3cfd20a516605aa8754c9205d2bdc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://97d9d693e8e60ee9c183bb44461e698e8b3cfd20a516605aa8754c9205d2bdc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c44252676359c671d65c9588b05572d76c99cf4c90666f0ac089b2f778450a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c44252676359c671d65c9588b05572d76c99cf4c90666f0ac089b2f778450a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80c982f07874df686fe7e29cc91043eb6d58f1be5fc93cacd46d33a21dd9b622\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80c982f07874df686fe7e29cc91043eb6d58f1be5fc93cacd46d33a21dd9b622\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a354c985652c7ad7fa8944fbfdfe223cfb0f97fb11eecd9bd429020d4892b78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a354c985652c7ad7fa8944fbfdfe223cfb0f97fb11eecd9bd429020d4892b78\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2efd7283df313d086c1b8b88f289ace28764de256205bee7610d6711002d8515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2efd7283df313d086c1b8b88f289ace28764de256205bee7610d6711002d8515\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2ll9b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:28Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:28 crc kubenswrapper[4885]: I1002 01:47:28.568420 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:28 crc kubenswrapper[4885]: I1002 01:47:28.568483 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:28 crc kubenswrapper[4885]: I1002 01:47:28.568501 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:28 crc kubenswrapper[4885]: I1002 01:47:28.568525 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:28 crc kubenswrapper[4885]: I1002 01:47:28.568543 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:28Z","lastTransitionTime":"2025-10-02T01:47:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:28 crc kubenswrapper[4885]: I1002 01:47:28.671645 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:28 crc kubenswrapper[4885]: I1002 01:47:28.671714 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:28 crc kubenswrapper[4885]: I1002 01:47:28.671731 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:28 crc kubenswrapper[4885]: I1002 01:47:28.671754 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:28 crc kubenswrapper[4885]: I1002 01:47:28.671771 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:28Z","lastTransitionTime":"2025-10-02T01:47:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:28 crc kubenswrapper[4885]: I1002 01:47:28.775169 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:28 crc kubenswrapper[4885]: I1002 01:47:28.775234 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:28 crc kubenswrapper[4885]: I1002 01:47:28.775251 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:28 crc kubenswrapper[4885]: I1002 01:47:28.775304 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:28 crc kubenswrapper[4885]: I1002 01:47:28.775322 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:28Z","lastTransitionTime":"2025-10-02T01:47:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:28 crc kubenswrapper[4885]: I1002 01:47:28.878125 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:28 crc kubenswrapper[4885]: I1002 01:47:28.878176 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:28 crc kubenswrapper[4885]: I1002 01:47:28.878193 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:28 crc kubenswrapper[4885]: I1002 01:47:28.878218 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:28 crc kubenswrapper[4885]: I1002 01:47:28.878236 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:28Z","lastTransitionTime":"2025-10-02T01:47:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:28 crc kubenswrapper[4885]: I1002 01:47:28.918470 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/724a3e49-001c-45a4-a896-82c13e0ee0e7-metrics-certs\") pod \"network-metrics-daemon-sv4md\" (UID: \"724a3e49-001c-45a4-a896-82c13e0ee0e7\") " pod="openshift-multus/network-metrics-daemon-sv4md" Oct 02 01:47:28 crc kubenswrapper[4885]: E1002 01:47:28.918657 4885 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 01:47:28 crc kubenswrapper[4885]: E1002 01:47:28.918770 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/724a3e49-001c-45a4-a896-82c13e0ee0e7-metrics-certs podName:724a3e49-001c-45a4-a896-82c13e0ee0e7 nodeName:}" failed. No retries permitted until 2025-10-02 01:47:29.918737852 +0000 UTC m=+38.730485301 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/724a3e49-001c-45a4-a896-82c13e0ee0e7-metrics-certs") pod "network-metrics-daemon-sv4md" (UID: "724a3e49-001c-45a4-a896-82c13e0ee0e7") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 01:47:28 crc kubenswrapper[4885]: I1002 01:47:28.981605 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:28 crc kubenswrapper[4885]: I1002 01:47:28.981664 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:28 crc kubenswrapper[4885]: I1002 01:47:28.981681 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:28 crc kubenswrapper[4885]: I1002 01:47:28.981706 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:28 crc kubenswrapper[4885]: I1002 01:47:28.981725 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:28Z","lastTransitionTime":"2025-10-02T01:47:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:29 crc kubenswrapper[4885]: I1002 01:47:29.083899 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:29 crc kubenswrapper[4885]: I1002 01:47:29.083967 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:29 crc kubenswrapper[4885]: I1002 01:47:29.083987 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:29 crc kubenswrapper[4885]: I1002 01:47:29.084016 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:29 crc kubenswrapper[4885]: I1002 01:47:29.084034 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:29Z","lastTransitionTime":"2025-10-02T01:47:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:29 crc kubenswrapper[4885]: I1002 01:47:29.186811 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:29 crc kubenswrapper[4885]: I1002 01:47:29.186913 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:29 crc kubenswrapper[4885]: I1002 01:47:29.186942 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:29 crc kubenswrapper[4885]: I1002 01:47:29.186971 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:29 crc kubenswrapper[4885]: I1002 01:47:29.186996 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:29Z","lastTransitionTime":"2025-10-02T01:47:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:29 crc kubenswrapper[4885]: I1002 01:47:29.290450 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:29 crc kubenswrapper[4885]: I1002 01:47:29.290530 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:29 crc kubenswrapper[4885]: I1002 01:47:29.290550 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:29 crc kubenswrapper[4885]: I1002 01:47:29.290575 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:29 crc kubenswrapper[4885]: I1002 01:47:29.290596 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:29Z","lastTransitionTime":"2025-10-02T01:47:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:29 crc kubenswrapper[4885]: I1002 01:47:29.393441 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:29 crc kubenswrapper[4885]: I1002 01:47:29.393528 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:29 crc kubenswrapper[4885]: I1002 01:47:29.393546 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:29 crc kubenswrapper[4885]: I1002 01:47:29.393570 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:29 crc kubenswrapper[4885]: I1002 01:47:29.393588 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:29Z","lastTransitionTime":"2025-10-02T01:47:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:29 crc kubenswrapper[4885]: I1002 01:47:29.496751 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:29 crc kubenswrapper[4885]: I1002 01:47:29.496838 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:29 crc kubenswrapper[4885]: I1002 01:47:29.496859 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:29 crc kubenswrapper[4885]: I1002 01:47:29.496882 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:29 crc kubenswrapper[4885]: I1002 01:47:29.496900 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:29Z","lastTransitionTime":"2025-10-02T01:47:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:29 crc kubenswrapper[4885]: I1002 01:47:29.600453 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:29 crc kubenswrapper[4885]: I1002 01:47:29.600850 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:29 crc kubenswrapper[4885]: I1002 01:47:29.600868 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:29 crc kubenswrapper[4885]: I1002 01:47:29.600896 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:29 crc kubenswrapper[4885]: I1002 01:47:29.600915 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:29Z","lastTransitionTime":"2025-10-02T01:47:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:29 crc kubenswrapper[4885]: I1002 01:47:29.704195 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:29 crc kubenswrapper[4885]: I1002 01:47:29.704254 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:29 crc kubenswrapper[4885]: I1002 01:47:29.704311 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:29 crc kubenswrapper[4885]: I1002 01:47:29.704340 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:29 crc kubenswrapper[4885]: I1002 01:47:29.704357 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:29Z","lastTransitionTime":"2025-10-02T01:47:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:29 crc kubenswrapper[4885]: I1002 01:47:29.806311 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:29 crc kubenswrapper[4885]: I1002 01:47:29.806368 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:29 crc kubenswrapper[4885]: I1002 01:47:29.806385 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:29 crc kubenswrapper[4885]: I1002 01:47:29.806406 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:29 crc kubenswrapper[4885]: I1002 01:47:29.806417 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:29Z","lastTransitionTime":"2025-10-02T01:47:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:29 crc kubenswrapper[4885]: I1002 01:47:29.909707 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:29 crc kubenswrapper[4885]: I1002 01:47:29.909761 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:29 crc kubenswrapper[4885]: I1002 01:47:29.909778 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:29 crc kubenswrapper[4885]: I1002 01:47:29.909801 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:29 crc kubenswrapper[4885]: I1002 01:47:29.909817 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:29Z","lastTransitionTime":"2025-10-02T01:47:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:29 crc kubenswrapper[4885]: I1002 01:47:29.931227 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/724a3e49-001c-45a4-a896-82c13e0ee0e7-metrics-certs\") pod \"network-metrics-daemon-sv4md\" (UID: \"724a3e49-001c-45a4-a896-82c13e0ee0e7\") " pod="openshift-multus/network-metrics-daemon-sv4md" Oct 02 01:47:29 crc kubenswrapper[4885]: E1002 01:47:29.931404 4885 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 01:47:29 crc kubenswrapper[4885]: E1002 01:47:29.931495 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/724a3e49-001c-45a4-a896-82c13e0ee0e7-metrics-certs podName:724a3e49-001c-45a4-a896-82c13e0ee0e7 nodeName:}" failed. No retries permitted until 2025-10-02 01:47:31.931467278 +0000 UTC m=+40.743214717 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/724a3e49-001c-45a4-a896-82c13e0ee0e7-metrics-certs") pod "network-metrics-daemon-sv4md" (UID: "724a3e49-001c-45a4-a896-82c13e0ee0e7") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 01:47:30 crc kubenswrapper[4885]: I1002 01:47:30.013065 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:30 crc kubenswrapper[4885]: I1002 01:47:30.013138 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:30 crc kubenswrapper[4885]: I1002 01:47:30.013157 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:30 crc kubenswrapper[4885]: I1002 01:47:30.013188 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:30 crc kubenswrapper[4885]: I1002 01:47:30.013207 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:30Z","lastTransitionTime":"2025-10-02T01:47:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:30 crc kubenswrapper[4885]: I1002 01:47:30.045537 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sv4md" Oct 02 01:47:30 crc kubenswrapper[4885]: I1002 01:47:30.045564 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:47:30 crc kubenswrapper[4885]: E1002 01:47:30.045715 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sv4md" podUID="724a3e49-001c-45a4-a896-82c13e0ee0e7" Oct 02 01:47:30 crc kubenswrapper[4885]: E1002 01:47:30.045796 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:47:30 crc kubenswrapper[4885]: I1002 01:47:30.045890 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:47:30 crc kubenswrapper[4885]: E1002 01:47:30.046050 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:47:30 crc kubenswrapper[4885]: I1002 01:47:30.046177 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:47:30 crc kubenswrapper[4885]: E1002 01:47:30.046291 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:47:30 crc kubenswrapper[4885]: I1002 01:47:30.116288 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:30 crc kubenswrapper[4885]: I1002 01:47:30.116351 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:30 crc kubenswrapper[4885]: I1002 01:47:30.116374 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:30 crc kubenswrapper[4885]: I1002 01:47:30.116400 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:30 crc kubenswrapper[4885]: I1002 01:47:30.116419 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:30Z","lastTransitionTime":"2025-10-02T01:47:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:30 crc kubenswrapper[4885]: I1002 01:47:30.225065 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:30 crc kubenswrapper[4885]: I1002 01:47:30.225114 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:30 crc kubenswrapper[4885]: I1002 01:47:30.225131 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:30 crc kubenswrapper[4885]: I1002 01:47:30.225156 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:30 crc kubenswrapper[4885]: I1002 01:47:30.225175 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:30Z","lastTransitionTime":"2025-10-02T01:47:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:30 crc kubenswrapper[4885]: I1002 01:47:30.327450 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:30 crc kubenswrapper[4885]: I1002 01:47:30.327498 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:30 crc kubenswrapper[4885]: I1002 01:47:30.327509 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:30 crc kubenswrapper[4885]: I1002 01:47:30.327526 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:30 crc kubenswrapper[4885]: I1002 01:47:30.327537 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:30Z","lastTransitionTime":"2025-10-02T01:47:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:30 crc kubenswrapper[4885]: I1002 01:47:30.431221 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:30 crc kubenswrapper[4885]: I1002 01:47:30.431334 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:30 crc kubenswrapper[4885]: I1002 01:47:30.431363 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:30 crc kubenswrapper[4885]: I1002 01:47:30.431393 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:30 crc kubenswrapper[4885]: I1002 01:47:30.431416 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:30Z","lastTransitionTime":"2025-10-02T01:47:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:30 crc kubenswrapper[4885]: I1002 01:47:30.534115 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:30 crc kubenswrapper[4885]: I1002 01:47:30.534167 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:30 crc kubenswrapper[4885]: I1002 01:47:30.534182 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:30 crc kubenswrapper[4885]: I1002 01:47:30.534203 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:30 crc kubenswrapper[4885]: I1002 01:47:30.534217 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:30Z","lastTransitionTime":"2025-10-02T01:47:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:30 crc kubenswrapper[4885]: I1002 01:47:30.637600 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:30 crc kubenswrapper[4885]: I1002 01:47:30.637643 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:30 crc kubenswrapper[4885]: I1002 01:47:30.637678 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:30 crc kubenswrapper[4885]: I1002 01:47:30.637705 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:30 crc kubenswrapper[4885]: I1002 01:47:30.637719 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:30Z","lastTransitionTime":"2025-10-02T01:47:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:30 crc kubenswrapper[4885]: I1002 01:47:30.740735 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:30 crc kubenswrapper[4885]: I1002 01:47:30.740804 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:30 crc kubenswrapper[4885]: I1002 01:47:30.740824 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:30 crc kubenswrapper[4885]: I1002 01:47:30.740853 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:30 crc kubenswrapper[4885]: I1002 01:47:30.740873 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:30Z","lastTransitionTime":"2025-10-02T01:47:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:30 crc kubenswrapper[4885]: I1002 01:47:30.843363 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:30 crc kubenswrapper[4885]: I1002 01:47:30.843429 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:30 crc kubenswrapper[4885]: I1002 01:47:30.843448 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:30 crc kubenswrapper[4885]: I1002 01:47:30.843476 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:30 crc kubenswrapper[4885]: I1002 01:47:30.843497 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:30Z","lastTransitionTime":"2025-10-02T01:47:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:30 crc kubenswrapper[4885]: I1002 01:47:30.947502 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:30 crc kubenswrapper[4885]: I1002 01:47:30.947563 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:30 crc kubenswrapper[4885]: I1002 01:47:30.947580 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:30 crc kubenswrapper[4885]: I1002 01:47:30.947605 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:30 crc kubenswrapper[4885]: I1002 01:47:30.947625 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:30Z","lastTransitionTime":"2025-10-02T01:47:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:31 crc kubenswrapper[4885]: I1002 01:47:31.050627 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:31 crc kubenswrapper[4885]: I1002 01:47:31.050688 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:31 crc kubenswrapper[4885]: I1002 01:47:31.050705 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:31 crc kubenswrapper[4885]: I1002 01:47:31.050727 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:31 crc kubenswrapper[4885]: I1002 01:47:31.050745 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:31Z","lastTransitionTime":"2025-10-02T01:47:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:31 crc kubenswrapper[4885]: I1002 01:47:31.154002 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:31 crc kubenswrapper[4885]: I1002 01:47:31.154184 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:31 crc kubenswrapper[4885]: I1002 01:47:31.154214 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:31 crc kubenswrapper[4885]: I1002 01:47:31.154247 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:31 crc kubenswrapper[4885]: I1002 01:47:31.154337 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:31Z","lastTransitionTime":"2025-10-02T01:47:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:31 crc kubenswrapper[4885]: I1002 01:47:31.257530 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:31 crc kubenswrapper[4885]: I1002 01:47:31.257589 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:31 crc kubenswrapper[4885]: I1002 01:47:31.257605 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:31 crc kubenswrapper[4885]: I1002 01:47:31.257632 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:31 crc kubenswrapper[4885]: I1002 01:47:31.257651 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:31Z","lastTransitionTime":"2025-10-02T01:47:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:31 crc kubenswrapper[4885]: I1002 01:47:31.360251 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:31 crc kubenswrapper[4885]: I1002 01:47:31.360342 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:31 crc kubenswrapper[4885]: I1002 01:47:31.360360 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:31 crc kubenswrapper[4885]: I1002 01:47:31.360382 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:31 crc kubenswrapper[4885]: I1002 01:47:31.360401 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:31Z","lastTransitionTime":"2025-10-02T01:47:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:31 crc kubenswrapper[4885]: I1002 01:47:31.463343 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:31 crc kubenswrapper[4885]: I1002 01:47:31.463403 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:31 crc kubenswrapper[4885]: I1002 01:47:31.463425 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:31 crc kubenswrapper[4885]: I1002 01:47:31.463451 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:31 crc kubenswrapper[4885]: I1002 01:47:31.463471 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:31Z","lastTransitionTime":"2025-10-02T01:47:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:31 crc kubenswrapper[4885]: I1002 01:47:31.566990 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:31 crc kubenswrapper[4885]: I1002 01:47:31.567043 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:31 crc kubenswrapper[4885]: I1002 01:47:31.567062 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:31 crc kubenswrapper[4885]: I1002 01:47:31.567085 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:31 crc kubenswrapper[4885]: I1002 01:47:31.567102 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:31Z","lastTransitionTime":"2025-10-02T01:47:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:31 crc kubenswrapper[4885]: I1002 01:47:31.670517 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:31 crc kubenswrapper[4885]: I1002 01:47:31.670575 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:31 crc kubenswrapper[4885]: I1002 01:47:31.670595 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:31 crc kubenswrapper[4885]: I1002 01:47:31.670620 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:31 crc kubenswrapper[4885]: I1002 01:47:31.670638 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:31Z","lastTransitionTime":"2025-10-02T01:47:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:31 crc kubenswrapper[4885]: I1002 01:47:31.772974 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:31 crc kubenswrapper[4885]: I1002 01:47:31.773051 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:31 crc kubenswrapper[4885]: I1002 01:47:31.773068 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:31 crc kubenswrapper[4885]: I1002 01:47:31.773088 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:31 crc kubenswrapper[4885]: I1002 01:47:31.773105 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:31Z","lastTransitionTime":"2025-10-02T01:47:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:31 crc kubenswrapper[4885]: I1002 01:47:31.876054 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:31 crc kubenswrapper[4885]: I1002 01:47:31.876121 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:31 crc kubenswrapper[4885]: I1002 01:47:31.876140 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:31 crc kubenswrapper[4885]: I1002 01:47:31.876166 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:31 crc kubenswrapper[4885]: I1002 01:47:31.876184 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:31Z","lastTransitionTime":"2025-10-02T01:47:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:31 crc kubenswrapper[4885]: I1002 01:47:31.954311 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/724a3e49-001c-45a4-a896-82c13e0ee0e7-metrics-certs\") pod \"network-metrics-daemon-sv4md\" (UID: \"724a3e49-001c-45a4-a896-82c13e0ee0e7\") " pod="openshift-multus/network-metrics-daemon-sv4md" Oct 02 01:47:31 crc kubenswrapper[4885]: E1002 01:47:31.954538 4885 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 01:47:31 crc kubenswrapper[4885]: E1002 01:47:31.954647 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/724a3e49-001c-45a4-a896-82c13e0ee0e7-metrics-certs podName:724a3e49-001c-45a4-a896-82c13e0ee0e7 nodeName:}" failed. No retries permitted until 2025-10-02 01:47:35.954614865 +0000 UTC m=+44.766362304 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/724a3e49-001c-45a4-a896-82c13e0ee0e7-metrics-certs") pod "network-metrics-daemon-sv4md" (UID: "724a3e49-001c-45a4-a896-82c13e0ee0e7") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 01:47:31 crc kubenswrapper[4885]: I1002 01:47:31.978776 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:31 crc kubenswrapper[4885]: I1002 01:47:31.978838 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:31 crc kubenswrapper[4885]: I1002 01:47:31.978856 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:31 crc kubenswrapper[4885]: I1002 01:47:31.978884 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:31 crc kubenswrapper[4885]: I1002 01:47:31.978903 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:31Z","lastTransitionTime":"2025-10-02T01:47:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:32 crc kubenswrapper[4885]: I1002 01:47:32.046234 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:47:32 crc kubenswrapper[4885]: I1002 01:47:32.046347 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sv4md" Oct 02 01:47:32 crc kubenswrapper[4885]: E1002 01:47:32.046432 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:47:32 crc kubenswrapper[4885]: E1002 01:47:32.046741 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sv4md" podUID="724a3e49-001c-45a4-a896-82c13e0ee0e7" Oct 02 01:47:32 crc kubenswrapper[4885]: I1002 01:47:32.046813 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:47:32 crc kubenswrapper[4885]: I1002 01:47:32.046912 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:47:32 crc kubenswrapper[4885]: E1002 01:47:32.047020 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:47:32 crc kubenswrapper[4885]: E1002 01:47:32.047122 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:47:32 crc kubenswrapper[4885]: I1002 01:47:32.064995 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"506a599f-fb1c-4d14-be88-0e4833f3d624\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31c2f0193ffef54965b79a3569b0f6f4547e540bac0774155159a1c996c196c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://533e73e5f90b7258acbcdf018ee01323c118f9dc45035b25b065ba01cabecccf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c00db833a593ec40eb63a6b383455e19e960c2e98b9b66f76877ee993057e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21e2f462a0deb0d740185663b2014f4a1e57b4fdbeb82571264b002f928035bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:32Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:32 crc kubenswrapper[4885]: I1002 01:47:32.082642 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:32 crc kubenswrapper[4885]: I1002 01:47:32.082685 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:32 crc kubenswrapper[4885]: I1002 01:47:32.082704 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:32 crc kubenswrapper[4885]: I1002 01:47:32.082729 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:32 crc kubenswrapper[4885]: I1002 01:47:32.082746 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:32Z","lastTransitionTime":"2025-10-02T01:47:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:32 crc kubenswrapper[4885]: I1002 01:47:32.096460 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e4679dd-f870-41e9-a32b-360f5d2ee81b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4310909ad9c7b609ed6814e04a525a6afae747775598e7275080413dec8b23a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7642033b0b5995d332ccf7992d6d7e6a5e9c9369ae2d1071ccb2f536938c8f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9f99f9e26a04fb42b01a0783a5c4e360514818ce9102edc7fa08626d50b9a34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6fa72b030d309656b73121e4ac2a3924a4e449aad66683dd9de5ee5f22d5a4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef403e519d197fb37b3e6b76a14a5128ba3f30a7a5e8bf0a420cf6a39b89575f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8482ba457683ffa03100ec1f07ea9d3f918be5e490b5ba05c2ba965d03033191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f85438a1b57b0a590e4d005978fe70ff8c891980badb20d65e1f0e14080fae6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5082a806f0c20edb9171182c100dd73895a0e51869f8252c191811e678795463\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T01:47:23Z\\\",\\\"message\\\":\\\" from k8s.io/client-go/informers/factory.go:160\\\\nI1002 01:47:23.739948 6233 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 01:47:23.740066 6233 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 01:47:23.740783 6233 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1002 01:47:23.740803 6233 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1002 01:47:23.740832 6233 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1002 01:47:23.740854 6233 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1002 01:47:23.740862 6233 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1002 01:47:23.740890 6233 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1002 01:47:23.740899 6233 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1002 01:47:23.740929 6233 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1002 01:47:23.741528 6233 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1002 01:47:23.741591 6233 factory.go:656] Stopping watch factory\\\\nI1002 01:47:23.741615 6233 ovnkube.go:599] Stopped ovnkube\\\\nI1002 0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:20Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f85438a1b57b0a590e4d005978fe70ff8c891980badb20d65e1f0e14080fae6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T01:47:25Z\\\",\\\"message\\\":\\\"dressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-api/machine-api-operator-machine-webhook_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-operator-machine-webhook\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.250\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1002 01:47:25.452713 6353 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1002 01:47:25.452745 6353 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-source-55646444c4-trplf] creating logical port openshift-network-diagnostics_network-check-source-55646444c4-trplf for pod on switch crc\\\\nF1002 01:47:25.452781 6353 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fafdd1094bd02786d745c984989852b0c690db38655f92b419733a2004f417c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0125e907d061a97286026ee823c4eed303734ed03ab4864f742592b23a1aa2f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0125e907d061a97286026ee823c4eed303734ed03ab4864f742592b23a1aa2f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dngcm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:32Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:32 crc kubenswrapper[4885]: I1002 01:47:32.111060 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-sv4md" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"724a3e49-001c-45a4-a896-82c13e0ee0e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hpqgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hpqgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-sv4md\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:32Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:32 crc kubenswrapper[4885]: I1002 01:47:32.141723 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8070bc1c-e3de-4e9a-985e-40d1a9858056\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3e634b3727c517cc48b17c2cfb38f26460f53d1bf25d631d08ec9acb8d780e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4adcc537c23697a324d82af94c31059dd444245efdf7c978db58ede256e112ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edcd1d99644a778dfb6d0a17d2a4abd906400fe8209e611ebf8d7a9abe7ccf28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://97cb59adc2ddfa365ebc9dec649600dc90466ffe59134a6d15a9d3134a7c9b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d2808bc658ea80d68469ed16422d56ef19447f4c90b4af7b2597c64406f8a44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://83acbc20919f7c1958b6d492fadb203855df02761662425df9f91a0757fe2bf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83acbc20919f7c1958b6d492fadb203855df02761662425df9f91a0757fe2bf7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e86ecd03835efc9a253b6804bc61e6ac690e75b8d1f3cc77f5647730e586574\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e86ecd03835efc9a253b6804bc61e6ac690e75b8d1f3cc77f5647730e586574\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://48f95af51feaf014a18496bdfadd59d7f1e11167c5c8b677fa7037e48368ff2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48f95af51feaf014a18496bdfadd59d7f1e11167c5c8b677fa7037e48368ff2a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:32Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:32 crc kubenswrapper[4885]: I1002 01:47:32.160790 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5c1f95f188569c9aa47e51f2916939fa0b65a5cdf4c16b798612d41f3c8411a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:32Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:32 crc kubenswrapper[4885]: I1002 01:47:32.180879 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vhgbd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c63a7d20d39b65acfb5a9be4c28b21c72111b398bea76c238e15ddb4dd63e139\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7zwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vhgbd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:32Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:32 crc kubenswrapper[4885]: I1002 01:47:32.185474 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:32 crc kubenswrapper[4885]: I1002 01:47:32.185513 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:32 crc kubenswrapper[4885]: I1002 01:47:32.185532 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:32 crc kubenswrapper[4885]: I1002 01:47:32.185556 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:32 crc kubenswrapper[4885]: I1002 01:47:32.185572 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:32Z","lastTransitionTime":"2025-10-02T01:47:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:32 crc kubenswrapper[4885]: I1002 01:47:32.195992 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rrn8b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52967a77-525c-4537-9b1f-d7bb754a8494\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3dee451ebe9006bf1c519fed4c65267570b6139d76841858352e087ae78d26a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtr74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e9412788d6df7332592beac48281822b26e1aa19b39addce8e1be556af9d982\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtr74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:26Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rrn8b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:32Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:32 crc kubenswrapper[4885]: I1002 01:47:32.213052 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:32Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:32 crc kubenswrapper[4885]: I1002 01:47:32.230321 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74adcf912ccedb6b5d4f2803fd215e3e2ef62367597aa6da528354c4511ee536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7403ef968d040ac09d722b7557438bef23892d578992826f42b700b370d495d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:32Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:32 crc kubenswrapper[4885]: I1002 01:47:32.250211 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2ll9b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52b39712-5fa1-4fe0-814b-b170a6e3938b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9a96383d0fbfd116c419054aaf1a7adba1937e8b0bad294cefd68b4ab0d2298\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c979b91ba748aa4647826a8926337662ca929c54090fb62285572819ec2fbf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c979b91ba748aa4647826a8926337662ca929c54090fb62285572819ec2fbf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97d9d693e8e60ee9c183bb44461e698e8b3cfd20a516605aa8754c9205d2bdc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://97d9d693e8e60ee9c183bb44461e698e8b3cfd20a516605aa8754c9205d2bdc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c44252676359c671d65c9588b05572d76c99cf4c90666f0ac089b2f778450a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c44252676359c671d65c9588b05572d76c99cf4c90666f0ac089b2f778450a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80c982f07874df686fe7e29cc91043eb6d58f1be5fc93cacd46d33a21dd9b622\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80c982f07874df686fe7e29cc91043eb6d58f1be5fc93cacd46d33a21dd9b622\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a354c985652c7ad7fa8944fbfdfe223cfb0f97fb11eecd9bd429020d4892b78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a354c985652c7ad7fa8944fbfdfe223cfb0f97fb11eecd9bd429020d4892b78\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2efd7283df313d086c1b8b88f289ace28764de256205bee7610d6711002d8515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2efd7283df313d086c1b8b88f289ace28764de256205bee7610d6711002d8515\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2ll9b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:32Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:32 crc kubenswrapper[4885]: I1002 01:47:32.267615 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a0bb885-2eb0-419c-8a7a-ef634cb3c144\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a89b49cbe08d927629564e607de13655ddf3d01ac4eb5d688b325210306c120c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06c5942805a498d9621a906088aa54538b4eed583d1b8d0e3ba2a6718378a36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b9f6f7fe44d5771a838d0b89c5bdea3ebb79023ae2675ee907ec2766b3d35a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://178b4340ee6904c60c814b99d83ee225161aeb23e438d8d02f856579a13ec4e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bb63bf30d6ecce9e71f0db0c16632bdfa6b33924b673378eb0c38ab229fa837\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T01:47:06Z\\\",\\\"message\\\":\\\"W1002 01:46:55.406027 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1002 01:46:55.406714 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759369615 cert, and key in /tmp/serving-cert-3795040473/serving-signer.crt, /tmp/serving-cert-3795040473/serving-signer.key\\\\nI1002 01:46:55.614670 1 observer_polling.go:159] Starting file observer\\\\nW1002 01:46:55.617662 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1002 01:46:55.618112 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 01:46:55.620347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3795040473/tls.crt::/tmp/serving-cert-3795040473/tls.key\\\\\\\"\\\\nF1002 01:47:06.224459 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90535f97ab2ed4626d47540ebd7baada837fd332140323bec0f49aa38d3c0009\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://592f89a326c9c1559193a71904b47ca27b1dbe1a7b30e03bc48924bb2f35f921\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://592f89a326c9c1559193a71904b47ca27b1dbe1a7b30e03bc48924bb2f35f921\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:32Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:32 crc kubenswrapper[4885]: I1002 01:47:32.286358 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b087880e05039816dcab154643ac72c7e4a0a1124817c51dc2c639b62bbee0fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:32Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:32 crc kubenswrapper[4885]: I1002 01:47:32.288144 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:32 crc kubenswrapper[4885]: I1002 01:47:32.288242 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:32 crc kubenswrapper[4885]: I1002 01:47:32.288571 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:32 crc kubenswrapper[4885]: I1002 01:47:32.288868 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:32 crc kubenswrapper[4885]: I1002 01:47:32.288922 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:32Z","lastTransitionTime":"2025-10-02T01:47:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:32 crc kubenswrapper[4885]: I1002 01:47:32.323021 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:32Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:32 crc kubenswrapper[4885]: I1002 01:47:32.356395 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:32Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:32 crc kubenswrapper[4885]: I1002 01:47:32.377452 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-58lhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac6a19c4-f259-4dbd-a016-d6c2c849c6cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed76c0f968c063f7795d363af194a222487b647ceda234de0dabb39b25eafe0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hzc4c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-58lhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:32Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:32 crc kubenswrapper[4885]: I1002 01:47:32.391041 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:32 crc kubenswrapper[4885]: I1002 01:47:32.391088 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:32 crc kubenswrapper[4885]: I1002 01:47:32.391104 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:32 crc kubenswrapper[4885]: I1002 01:47:32.391125 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:32 crc kubenswrapper[4885]: I1002 01:47:32.391140 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:32Z","lastTransitionTime":"2025-10-02T01:47:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:32 crc kubenswrapper[4885]: I1002 01:47:32.393221 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d169c8ba-000a-4218-8e3b-4ae53035b110\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61e436b7876107ce5ea9e668c0bbc367be6e663fc804fa3a65b51adf8be8d51f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-77rzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09536e182a46e77318c10812d1777e551f834907c006cab5c726dc66f5bce88a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-77rzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rttx8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:32Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:32 crc kubenswrapper[4885]: I1002 01:47:32.406003 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jgdvt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81733e8d-67b0-46fa-8a41-08b40645bd3e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcef60c816ca1ba9a298da658e84c4d717bea3d352c4c2f807ebd8e56327b384\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cwbl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jgdvt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:32Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:32 crc kubenswrapper[4885]: I1002 01:47:32.493697 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:32 crc kubenswrapper[4885]: I1002 01:47:32.493730 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:32 crc kubenswrapper[4885]: I1002 01:47:32.493741 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:32 crc kubenswrapper[4885]: I1002 01:47:32.493757 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:32 crc kubenswrapper[4885]: I1002 01:47:32.493769 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:32Z","lastTransitionTime":"2025-10-02T01:47:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:32 crc kubenswrapper[4885]: I1002 01:47:32.596637 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:32 crc kubenswrapper[4885]: I1002 01:47:32.596703 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:32 crc kubenswrapper[4885]: I1002 01:47:32.596724 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:32 crc kubenswrapper[4885]: I1002 01:47:32.596750 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:32 crc kubenswrapper[4885]: I1002 01:47:32.596767 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:32Z","lastTransitionTime":"2025-10-02T01:47:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:32 crc kubenswrapper[4885]: I1002 01:47:32.699882 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:32 crc kubenswrapper[4885]: I1002 01:47:32.699937 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:32 crc kubenswrapper[4885]: I1002 01:47:32.699955 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:32 crc kubenswrapper[4885]: I1002 01:47:32.699979 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:32 crc kubenswrapper[4885]: I1002 01:47:32.699998 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:32Z","lastTransitionTime":"2025-10-02T01:47:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:32 crc kubenswrapper[4885]: I1002 01:47:32.803185 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:32 crc kubenswrapper[4885]: I1002 01:47:32.803241 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:32 crc kubenswrapper[4885]: I1002 01:47:32.803287 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:32 crc kubenswrapper[4885]: I1002 01:47:32.803310 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:32 crc kubenswrapper[4885]: I1002 01:47:32.803329 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:32Z","lastTransitionTime":"2025-10-02T01:47:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:32 crc kubenswrapper[4885]: I1002 01:47:32.906853 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:32 crc kubenswrapper[4885]: I1002 01:47:32.906923 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:32 crc kubenswrapper[4885]: I1002 01:47:32.906941 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:32 crc kubenswrapper[4885]: I1002 01:47:32.906964 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:32 crc kubenswrapper[4885]: I1002 01:47:32.907012 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:32Z","lastTransitionTime":"2025-10-02T01:47:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:33 crc kubenswrapper[4885]: I1002 01:47:33.011018 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:33 crc kubenswrapper[4885]: I1002 01:47:33.011213 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:33 crc kubenswrapper[4885]: I1002 01:47:33.011239 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:33 crc kubenswrapper[4885]: I1002 01:47:33.011324 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:33 crc kubenswrapper[4885]: I1002 01:47:33.011353 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:33Z","lastTransitionTime":"2025-10-02T01:47:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:33 crc kubenswrapper[4885]: I1002 01:47:33.115008 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:33 crc kubenswrapper[4885]: I1002 01:47:33.115068 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:33 crc kubenswrapper[4885]: I1002 01:47:33.115114 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:33 crc kubenswrapper[4885]: I1002 01:47:33.115141 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:33 crc kubenswrapper[4885]: I1002 01:47:33.115158 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:33Z","lastTransitionTime":"2025-10-02T01:47:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:33 crc kubenswrapper[4885]: I1002 01:47:33.218096 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:33 crc kubenswrapper[4885]: I1002 01:47:33.218157 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:33 crc kubenswrapper[4885]: I1002 01:47:33.218177 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:33 crc kubenswrapper[4885]: I1002 01:47:33.218204 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:33 crc kubenswrapper[4885]: I1002 01:47:33.218221 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:33Z","lastTransitionTime":"2025-10-02T01:47:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:33 crc kubenswrapper[4885]: I1002 01:47:33.321506 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:33 crc kubenswrapper[4885]: I1002 01:47:33.321573 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:33 crc kubenswrapper[4885]: I1002 01:47:33.321597 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:33 crc kubenswrapper[4885]: I1002 01:47:33.321631 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:33 crc kubenswrapper[4885]: I1002 01:47:33.321649 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:33Z","lastTransitionTime":"2025-10-02T01:47:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:33 crc kubenswrapper[4885]: I1002 01:47:33.425056 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:33 crc kubenswrapper[4885]: I1002 01:47:33.425173 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:33 crc kubenswrapper[4885]: I1002 01:47:33.425193 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:33 crc kubenswrapper[4885]: I1002 01:47:33.425220 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:33 crc kubenswrapper[4885]: I1002 01:47:33.425238 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:33Z","lastTransitionTime":"2025-10-02T01:47:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:33 crc kubenswrapper[4885]: I1002 01:47:33.528600 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:33 crc kubenswrapper[4885]: I1002 01:47:33.528703 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:33 crc kubenswrapper[4885]: I1002 01:47:33.528823 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:33 crc kubenswrapper[4885]: I1002 01:47:33.528852 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:33 crc kubenswrapper[4885]: I1002 01:47:33.528872 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:33Z","lastTransitionTime":"2025-10-02T01:47:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:33 crc kubenswrapper[4885]: I1002 01:47:33.631642 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:33 crc kubenswrapper[4885]: I1002 01:47:33.631704 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:33 crc kubenswrapper[4885]: I1002 01:47:33.631723 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:33 crc kubenswrapper[4885]: I1002 01:47:33.631755 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:33 crc kubenswrapper[4885]: I1002 01:47:33.631776 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:33Z","lastTransitionTime":"2025-10-02T01:47:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:33 crc kubenswrapper[4885]: I1002 01:47:33.734398 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:33 crc kubenswrapper[4885]: I1002 01:47:33.734459 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:33 crc kubenswrapper[4885]: I1002 01:47:33.734478 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:33 crc kubenswrapper[4885]: I1002 01:47:33.734503 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:33 crc kubenswrapper[4885]: I1002 01:47:33.734521 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:33Z","lastTransitionTime":"2025-10-02T01:47:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:33 crc kubenswrapper[4885]: I1002 01:47:33.837905 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:33 crc kubenswrapper[4885]: I1002 01:47:33.837973 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:33 crc kubenswrapper[4885]: I1002 01:47:33.837996 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:33 crc kubenswrapper[4885]: I1002 01:47:33.838022 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:33 crc kubenswrapper[4885]: I1002 01:47:33.838040 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:33Z","lastTransitionTime":"2025-10-02T01:47:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:33 crc kubenswrapper[4885]: I1002 01:47:33.941610 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:33 crc kubenswrapper[4885]: I1002 01:47:33.941658 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:33 crc kubenswrapper[4885]: I1002 01:47:33.941671 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:33 crc kubenswrapper[4885]: I1002 01:47:33.941691 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:33 crc kubenswrapper[4885]: I1002 01:47:33.941703 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:33Z","lastTransitionTime":"2025-10-02T01:47:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:34 crc kubenswrapper[4885]: I1002 01:47:34.044816 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:34 crc kubenswrapper[4885]: I1002 01:47:34.044937 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:34 crc kubenswrapper[4885]: I1002 01:47:34.044963 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:34 crc kubenswrapper[4885]: I1002 01:47:34.044992 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:34 crc kubenswrapper[4885]: I1002 01:47:34.045011 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:34Z","lastTransitionTime":"2025-10-02T01:47:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:34 crc kubenswrapper[4885]: I1002 01:47:34.045877 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:47:34 crc kubenswrapper[4885]: I1002 01:47:34.045907 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:47:34 crc kubenswrapper[4885]: I1002 01:47:34.045897 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sv4md" Oct 02 01:47:34 crc kubenswrapper[4885]: E1002 01:47:34.046088 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:47:34 crc kubenswrapper[4885]: E1002 01:47:34.046224 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:47:34 crc kubenswrapper[4885]: E1002 01:47:34.046380 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sv4md" podUID="724a3e49-001c-45a4-a896-82c13e0ee0e7" Oct 02 01:47:34 crc kubenswrapper[4885]: I1002 01:47:34.046431 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:47:34 crc kubenswrapper[4885]: E1002 01:47:34.046521 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:47:34 crc kubenswrapper[4885]: I1002 01:47:34.147969 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:34 crc kubenswrapper[4885]: I1002 01:47:34.148050 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:34 crc kubenswrapper[4885]: I1002 01:47:34.148071 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:34 crc kubenswrapper[4885]: I1002 01:47:34.148105 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:34 crc kubenswrapper[4885]: I1002 01:47:34.148131 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:34Z","lastTransitionTime":"2025-10-02T01:47:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:34 crc kubenswrapper[4885]: I1002 01:47:34.251071 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:34 crc kubenswrapper[4885]: I1002 01:47:34.251143 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:34 crc kubenswrapper[4885]: I1002 01:47:34.251164 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:34 crc kubenswrapper[4885]: I1002 01:47:34.251195 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:34 crc kubenswrapper[4885]: I1002 01:47:34.251215 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:34Z","lastTransitionTime":"2025-10-02T01:47:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:34 crc kubenswrapper[4885]: I1002 01:47:34.353632 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:34 crc kubenswrapper[4885]: I1002 01:47:34.353691 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:34 crc kubenswrapper[4885]: I1002 01:47:34.353709 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:34 crc kubenswrapper[4885]: I1002 01:47:34.353730 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:34 crc kubenswrapper[4885]: I1002 01:47:34.353746 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:34Z","lastTransitionTime":"2025-10-02T01:47:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:34 crc kubenswrapper[4885]: I1002 01:47:34.457069 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:34 crc kubenswrapper[4885]: I1002 01:47:34.457151 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:34 crc kubenswrapper[4885]: I1002 01:47:34.457176 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:34 crc kubenswrapper[4885]: I1002 01:47:34.457210 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:34 crc kubenswrapper[4885]: I1002 01:47:34.457229 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:34Z","lastTransitionTime":"2025-10-02T01:47:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:34 crc kubenswrapper[4885]: I1002 01:47:34.560060 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:34 crc kubenswrapper[4885]: I1002 01:47:34.560435 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:34 crc kubenswrapper[4885]: I1002 01:47:34.560581 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:34 crc kubenswrapper[4885]: I1002 01:47:34.560721 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:34 crc kubenswrapper[4885]: I1002 01:47:34.560841 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:34Z","lastTransitionTime":"2025-10-02T01:47:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:34 crc kubenswrapper[4885]: I1002 01:47:34.663909 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:34 crc kubenswrapper[4885]: I1002 01:47:34.664004 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:34 crc kubenswrapper[4885]: I1002 01:47:34.664031 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:34 crc kubenswrapper[4885]: I1002 01:47:34.664064 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:34 crc kubenswrapper[4885]: I1002 01:47:34.664087 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:34Z","lastTransitionTime":"2025-10-02T01:47:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:34 crc kubenswrapper[4885]: I1002 01:47:34.767740 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:34 crc kubenswrapper[4885]: I1002 01:47:34.767805 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:34 crc kubenswrapper[4885]: I1002 01:47:34.767822 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:34 crc kubenswrapper[4885]: I1002 01:47:34.767847 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:34 crc kubenswrapper[4885]: I1002 01:47:34.767868 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:34Z","lastTransitionTime":"2025-10-02T01:47:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:34 crc kubenswrapper[4885]: I1002 01:47:34.871538 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:34 crc kubenswrapper[4885]: I1002 01:47:34.871612 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:34 crc kubenswrapper[4885]: I1002 01:47:34.871630 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:34 crc kubenswrapper[4885]: I1002 01:47:34.871657 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:34 crc kubenswrapper[4885]: I1002 01:47:34.871678 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:34Z","lastTransitionTime":"2025-10-02T01:47:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:34 crc kubenswrapper[4885]: I1002 01:47:34.974857 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:34 crc kubenswrapper[4885]: I1002 01:47:34.974923 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:34 crc kubenswrapper[4885]: I1002 01:47:34.974940 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:34 crc kubenswrapper[4885]: I1002 01:47:34.974968 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:34 crc kubenswrapper[4885]: I1002 01:47:34.974988 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:34Z","lastTransitionTime":"2025-10-02T01:47:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:35 crc kubenswrapper[4885]: I1002 01:47:35.078567 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:35 crc kubenswrapper[4885]: I1002 01:47:35.078625 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:35 crc kubenswrapper[4885]: I1002 01:47:35.078645 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:35 crc kubenswrapper[4885]: I1002 01:47:35.078669 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:35 crc kubenswrapper[4885]: I1002 01:47:35.078687 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:35Z","lastTransitionTime":"2025-10-02T01:47:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:35 crc kubenswrapper[4885]: I1002 01:47:35.181310 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:35 crc kubenswrapper[4885]: I1002 01:47:35.181377 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:35 crc kubenswrapper[4885]: I1002 01:47:35.181394 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:35 crc kubenswrapper[4885]: I1002 01:47:35.181418 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:35 crc kubenswrapper[4885]: I1002 01:47:35.181436 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:35Z","lastTransitionTime":"2025-10-02T01:47:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:35 crc kubenswrapper[4885]: I1002 01:47:35.283961 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:35 crc kubenswrapper[4885]: I1002 01:47:35.284023 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:35 crc kubenswrapper[4885]: I1002 01:47:35.284043 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:35 crc kubenswrapper[4885]: I1002 01:47:35.284067 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:35 crc kubenswrapper[4885]: I1002 01:47:35.284084 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:35Z","lastTransitionTime":"2025-10-02T01:47:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:35 crc kubenswrapper[4885]: I1002 01:47:35.387305 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:35 crc kubenswrapper[4885]: I1002 01:47:35.387374 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:35 crc kubenswrapper[4885]: I1002 01:47:35.387393 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:35 crc kubenswrapper[4885]: I1002 01:47:35.387423 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:35 crc kubenswrapper[4885]: I1002 01:47:35.387445 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:35Z","lastTransitionTime":"2025-10-02T01:47:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:35 crc kubenswrapper[4885]: I1002 01:47:35.490783 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:35 crc kubenswrapper[4885]: I1002 01:47:35.490890 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:35 crc kubenswrapper[4885]: I1002 01:47:35.490907 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:35 crc kubenswrapper[4885]: I1002 01:47:35.490930 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:35 crc kubenswrapper[4885]: I1002 01:47:35.490953 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:35Z","lastTransitionTime":"2025-10-02T01:47:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:35 crc kubenswrapper[4885]: I1002 01:47:35.595211 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:35 crc kubenswrapper[4885]: I1002 01:47:35.595318 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:35 crc kubenswrapper[4885]: I1002 01:47:35.595337 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:35 crc kubenswrapper[4885]: I1002 01:47:35.595367 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:35 crc kubenswrapper[4885]: I1002 01:47:35.595391 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:35Z","lastTransitionTime":"2025-10-02T01:47:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:35 crc kubenswrapper[4885]: I1002 01:47:35.698919 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:35 crc kubenswrapper[4885]: I1002 01:47:35.698988 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:35 crc kubenswrapper[4885]: I1002 01:47:35.699006 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:35 crc kubenswrapper[4885]: I1002 01:47:35.699030 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:35 crc kubenswrapper[4885]: I1002 01:47:35.699050 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:35Z","lastTransitionTime":"2025-10-02T01:47:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:35 crc kubenswrapper[4885]: I1002 01:47:35.801798 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:35 crc kubenswrapper[4885]: I1002 01:47:35.801848 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:35 crc kubenswrapper[4885]: I1002 01:47:35.801863 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:35 crc kubenswrapper[4885]: I1002 01:47:35.801887 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:35 crc kubenswrapper[4885]: I1002 01:47:35.801911 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:35Z","lastTransitionTime":"2025-10-02T01:47:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:35 crc kubenswrapper[4885]: I1002 01:47:35.905732 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:35 crc kubenswrapper[4885]: I1002 01:47:35.905803 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:35 crc kubenswrapper[4885]: I1002 01:47:35.905824 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:35 crc kubenswrapper[4885]: I1002 01:47:35.905850 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:35 crc kubenswrapper[4885]: I1002 01:47:35.905868 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:35Z","lastTransitionTime":"2025-10-02T01:47:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:36 crc kubenswrapper[4885]: I1002 01:47:36.001097 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/724a3e49-001c-45a4-a896-82c13e0ee0e7-metrics-certs\") pod \"network-metrics-daemon-sv4md\" (UID: \"724a3e49-001c-45a4-a896-82c13e0ee0e7\") " pod="openshift-multus/network-metrics-daemon-sv4md" Oct 02 01:47:36 crc kubenswrapper[4885]: E1002 01:47:36.001413 4885 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 01:47:36 crc kubenswrapper[4885]: E1002 01:47:36.001602 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/724a3e49-001c-45a4-a896-82c13e0ee0e7-metrics-certs podName:724a3e49-001c-45a4-a896-82c13e0ee0e7 nodeName:}" failed. No retries permitted until 2025-10-02 01:47:44.001563015 +0000 UTC m=+52.813310494 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/724a3e49-001c-45a4-a896-82c13e0ee0e7-metrics-certs") pod "network-metrics-daemon-sv4md" (UID: "724a3e49-001c-45a4-a896-82c13e0ee0e7") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 01:47:36 crc kubenswrapper[4885]: I1002 01:47:36.009574 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:36 crc kubenswrapper[4885]: I1002 01:47:36.009622 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:36 crc kubenswrapper[4885]: I1002 01:47:36.009639 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:36 crc kubenswrapper[4885]: I1002 01:47:36.009667 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:36 crc kubenswrapper[4885]: I1002 01:47:36.009690 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:36Z","lastTransitionTime":"2025-10-02T01:47:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:36 crc kubenswrapper[4885]: I1002 01:47:36.046321 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:47:36 crc kubenswrapper[4885]: I1002 01:47:36.046368 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sv4md" Oct 02 01:47:36 crc kubenswrapper[4885]: I1002 01:47:36.046471 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:47:36 crc kubenswrapper[4885]: I1002 01:47:36.046701 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:47:36 crc kubenswrapper[4885]: E1002 01:47:36.046687 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:47:36 crc kubenswrapper[4885]: E1002 01:47:36.046873 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sv4md" podUID="724a3e49-001c-45a4-a896-82c13e0ee0e7" Oct 02 01:47:36 crc kubenswrapper[4885]: E1002 01:47:36.047015 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:47:36 crc kubenswrapper[4885]: E1002 01:47:36.047221 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:47:36 crc kubenswrapper[4885]: I1002 01:47:36.112620 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:36 crc kubenswrapper[4885]: I1002 01:47:36.112692 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:36 crc kubenswrapper[4885]: I1002 01:47:36.112711 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:36 crc kubenswrapper[4885]: I1002 01:47:36.112735 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:36 crc kubenswrapper[4885]: I1002 01:47:36.112754 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:36Z","lastTransitionTime":"2025-10-02T01:47:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:36 crc kubenswrapper[4885]: I1002 01:47:36.215918 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:36 crc kubenswrapper[4885]: I1002 01:47:36.215988 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:36 crc kubenswrapper[4885]: I1002 01:47:36.216007 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:36 crc kubenswrapper[4885]: I1002 01:47:36.216033 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:36 crc kubenswrapper[4885]: I1002 01:47:36.216052 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:36Z","lastTransitionTime":"2025-10-02T01:47:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:36 crc kubenswrapper[4885]: I1002 01:47:36.319255 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:36 crc kubenswrapper[4885]: I1002 01:47:36.319359 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:36 crc kubenswrapper[4885]: I1002 01:47:36.319380 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:36 crc kubenswrapper[4885]: I1002 01:47:36.319404 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:36 crc kubenswrapper[4885]: I1002 01:47:36.319422 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:36Z","lastTransitionTime":"2025-10-02T01:47:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:36 crc kubenswrapper[4885]: I1002 01:47:36.423098 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:36 crc kubenswrapper[4885]: I1002 01:47:36.423154 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:36 crc kubenswrapper[4885]: I1002 01:47:36.423173 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:36 crc kubenswrapper[4885]: I1002 01:47:36.423196 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:36 crc kubenswrapper[4885]: I1002 01:47:36.423212 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:36Z","lastTransitionTime":"2025-10-02T01:47:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:36 crc kubenswrapper[4885]: I1002 01:47:36.526921 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:36 crc kubenswrapper[4885]: I1002 01:47:36.526979 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:36 crc kubenswrapper[4885]: I1002 01:47:36.526998 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:36 crc kubenswrapper[4885]: I1002 01:47:36.527026 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:36 crc kubenswrapper[4885]: I1002 01:47:36.527087 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:36Z","lastTransitionTime":"2025-10-02T01:47:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:36 crc kubenswrapper[4885]: I1002 01:47:36.630028 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:36 crc kubenswrapper[4885]: I1002 01:47:36.630085 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:36 crc kubenswrapper[4885]: I1002 01:47:36.630102 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:36 crc kubenswrapper[4885]: I1002 01:47:36.630125 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:36 crc kubenswrapper[4885]: I1002 01:47:36.630145 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:36Z","lastTransitionTime":"2025-10-02T01:47:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:36 crc kubenswrapper[4885]: I1002 01:47:36.732952 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:36 crc kubenswrapper[4885]: I1002 01:47:36.733030 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:36 crc kubenswrapper[4885]: I1002 01:47:36.733054 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:36 crc kubenswrapper[4885]: I1002 01:47:36.733087 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:36 crc kubenswrapper[4885]: I1002 01:47:36.733112 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:36Z","lastTransitionTime":"2025-10-02T01:47:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:36 crc kubenswrapper[4885]: I1002 01:47:36.836360 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:36 crc kubenswrapper[4885]: I1002 01:47:36.836415 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:36 crc kubenswrapper[4885]: I1002 01:47:36.836434 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:36 crc kubenswrapper[4885]: I1002 01:47:36.836457 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:36 crc kubenswrapper[4885]: I1002 01:47:36.836474 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:36Z","lastTransitionTime":"2025-10-02T01:47:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:36 crc kubenswrapper[4885]: I1002 01:47:36.939671 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:36 crc kubenswrapper[4885]: I1002 01:47:36.939754 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:36 crc kubenswrapper[4885]: I1002 01:47:36.939776 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:36 crc kubenswrapper[4885]: I1002 01:47:36.939812 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:36 crc kubenswrapper[4885]: I1002 01:47:36.939834 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:36Z","lastTransitionTime":"2025-10-02T01:47:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:37 crc kubenswrapper[4885]: I1002 01:47:37.043255 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:37 crc kubenswrapper[4885]: I1002 01:47:37.043354 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:37 crc kubenswrapper[4885]: I1002 01:47:37.043370 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:37 crc kubenswrapper[4885]: I1002 01:47:37.043395 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:37 crc kubenswrapper[4885]: I1002 01:47:37.043413 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:37Z","lastTransitionTime":"2025-10-02T01:47:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:37 crc kubenswrapper[4885]: I1002 01:47:37.148704 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:37 crc kubenswrapper[4885]: I1002 01:47:37.150207 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:37 crc kubenswrapper[4885]: I1002 01:47:37.150309 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:37 crc kubenswrapper[4885]: I1002 01:47:37.150351 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:37 crc kubenswrapper[4885]: I1002 01:47:37.150374 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:37Z","lastTransitionTime":"2025-10-02T01:47:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:37 crc kubenswrapper[4885]: I1002 01:47:37.253501 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:37 crc kubenswrapper[4885]: I1002 01:47:37.253569 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:37 crc kubenswrapper[4885]: I1002 01:47:37.253594 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:37 crc kubenswrapper[4885]: I1002 01:47:37.253625 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:37 crc kubenswrapper[4885]: I1002 01:47:37.253648 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:37Z","lastTransitionTime":"2025-10-02T01:47:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:37 crc kubenswrapper[4885]: I1002 01:47:37.357894 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:37 crc kubenswrapper[4885]: I1002 01:47:37.357949 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:37 crc kubenswrapper[4885]: I1002 01:47:37.357967 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:37 crc kubenswrapper[4885]: I1002 01:47:37.357991 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:37 crc kubenswrapper[4885]: I1002 01:47:37.358009 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:37Z","lastTransitionTime":"2025-10-02T01:47:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:37 crc kubenswrapper[4885]: I1002 01:47:37.462558 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:37 crc kubenswrapper[4885]: I1002 01:47:37.462613 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:37 crc kubenswrapper[4885]: I1002 01:47:37.462630 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:37 crc kubenswrapper[4885]: I1002 01:47:37.462654 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:37 crc kubenswrapper[4885]: I1002 01:47:37.462670 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:37Z","lastTransitionTime":"2025-10-02T01:47:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:37 crc kubenswrapper[4885]: I1002 01:47:37.565792 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:37 crc kubenswrapper[4885]: I1002 01:47:37.565853 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:37 crc kubenswrapper[4885]: I1002 01:47:37.565869 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:37 crc kubenswrapper[4885]: I1002 01:47:37.565892 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:37 crc kubenswrapper[4885]: I1002 01:47:37.565909 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:37Z","lastTransitionTime":"2025-10-02T01:47:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:37 crc kubenswrapper[4885]: I1002 01:47:37.669044 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:37 crc kubenswrapper[4885]: I1002 01:47:37.669118 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:37 crc kubenswrapper[4885]: I1002 01:47:37.669141 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:37 crc kubenswrapper[4885]: I1002 01:47:37.669171 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:37 crc kubenswrapper[4885]: I1002 01:47:37.669192 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:37Z","lastTransitionTime":"2025-10-02T01:47:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:37 crc kubenswrapper[4885]: I1002 01:47:37.772666 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:37 crc kubenswrapper[4885]: I1002 01:47:37.772725 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:37 crc kubenswrapper[4885]: I1002 01:47:37.772743 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:37 crc kubenswrapper[4885]: I1002 01:47:37.772767 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:37 crc kubenswrapper[4885]: I1002 01:47:37.772786 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:37Z","lastTransitionTime":"2025-10-02T01:47:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:37 crc kubenswrapper[4885]: I1002 01:47:37.876131 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:37 crc kubenswrapper[4885]: I1002 01:47:37.876553 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:37 crc kubenswrapper[4885]: I1002 01:47:37.876686 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:37 crc kubenswrapper[4885]: I1002 01:47:37.876827 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:37 crc kubenswrapper[4885]: I1002 01:47:37.876963 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:37Z","lastTransitionTime":"2025-10-02T01:47:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:37 crc kubenswrapper[4885]: I1002 01:47:37.980460 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:37 crc kubenswrapper[4885]: I1002 01:47:37.980522 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:37 crc kubenswrapper[4885]: I1002 01:47:37.980540 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:37 crc kubenswrapper[4885]: I1002 01:47:37.980563 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:37 crc kubenswrapper[4885]: I1002 01:47:37.980581 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:37Z","lastTransitionTime":"2025-10-02T01:47:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:38 crc kubenswrapper[4885]: I1002 01:47:38.045999 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:47:38 crc kubenswrapper[4885]: E1002 01:47:38.046593 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:47:38 crc kubenswrapper[4885]: I1002 01:47:38.046177 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:47:38 crc kubenswrapper[4885]: I1002 01:47:38.046387 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:47:38 crc kubenswrapper[4885]: I1002 01:47:38.046118 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sv4md" Oct 02 01:47:38 crc kubenswrapper[4885]: E1002 01:47:38.047065 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:47:38 crc kubenswrapper[4885]: E1002 01:47:38.047226 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:47:38 crc kubenswrapper[4885]: E1002 01:47:38.047458 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sv4md" podUID="724a3e49-001c-45a4-a896-82c13e0ee0e7" Oct 02 01:47:38 crc kubenswrapper[4885]: I1002 01:47:38.083145 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:38 crc kubenswrapper[4885]: I1002 01:47:38.083199 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:38 crc kubenswrapper[4885]: I1002 01:47:38.083216 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:38 crc kubenswrapper[4885]: I1002 01:47:38.083240 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:38 crc kubenswrapper[4885]: I1002 01:47:38.083295 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:38Z","lastTransitionTime":"2025-10-02T01:47:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:38 crc kubenswrapper[4885]: I1002 01:47:38.186695 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:38 crc kubenswrapper[4885]: I1002 01:47:38.186754 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:38 crc kubenswrapper[4885]: I1002 01:47:38.186780 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:38 crc kubenswrapper[4885]: I1002 01:47:38.186812 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:38 crc kubenswrapper[4885]: I1002 01:47:38.186836 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:38Z","lastTransitionTime":"2025-10-02T01:47:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:38 crc kubenswrapper[4885]: I1002 01:47:38.290024 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:38 crc kubenswrapper[4885]: I1002 01:47:38.290080 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:38 crc kubenswrapper[4885]: I1002 01:47:38.290096 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:38 crc kubenswrapper[4885]: I1002 01:47:38.290120 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:38 crc kubenswrapper[4885]: I1002 01:47:38.290138 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:38Z","lastTransitionTime":"2025-10-02T01:47:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:38 crc kubenswrapper[4885]: I1002 01:47:38.393761 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:38 crc kubenswrapper[4885]: I1002 01:47:38.393822 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:38 crc kubenswrapper[4885]: I1002 01:47:38.393840 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:38 crc kubenswrapper[4885]: I1002 01:47:38.393864 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:38 crc kubenswrapper[4885]: I1002 01:47:38.393882 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:38Z","lastTransitionTime":"2025-10-02T01:47:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:38 crc kubenswrapper[4885]: I1002 01:47:38.496934 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:38 crc kubenswrapper[4885]: I1002 01:47:38.497042 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:38 crc kubenswrapper[4885]: I1002 01:47:38.497060 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:38 crc kubenswrapper[4885]: I1002 01:47:38.497087 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:38 crc kubenswrapper[4885]: I1002 01:47:38.497105 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:38Z","lastTransitionTime":"2025-10-02T01:47:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:38 crc kubenswrapper[4885]: I1002 01:47:38.600440 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:38 crc kubenswrapper[4885]: I1002 01:47:38.600524 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:38 crc kubenswrapper[4885]: I1002 01:47:38.600547 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:38 crc kubenswrapper[4885]: I1002 01:47:38.600578 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:38 crc kubenswrapper[4885]: I1002 01:47:38.600599 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:38Z","lastTransitionTime":"2025-10-02T01:47:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:38 crc kubenswrapper[4885]: I1002 01:47:38.704010 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:38 crc kubenswrapper[4885]: I1002 01:47:38.704065 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:38 crc kubenswrapper[4885]: I1002 01:47:38.704087 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:38 crc kubenswrapper[4885]: I1002 01:47:38.704116 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:38 crc kubenswrapper[4885]: I1002 01:47:38.704137 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:38Z","lastTransitionTime":"2025-10-02T01:47:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:38 crc kubenswrapper[4885]: I1002 01:47:38.724401 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:38 crc kubenswrapper[4885]: I1002 01:47:38.724460 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:38 crc kubenswrapper[4885]: I1002 01:47:38.724479 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:38 crc kubenswrapper[4885]: I1002 01:47:38.724502 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:38 crc kubenswrapper[4885]: I1002 01:47:38.724521 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:38Z","lastTransitionTime":"2025-10-02T01:47:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:38 crc kubenswrapper[4885]: E1002 01:47:38.745358 4885 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:47:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:47:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:47:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:47:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"92603ae5-ba38-4201-98e9-979fd56cb50d\\\",\\\"systemUUID\\\":\\\"50ddf2c2-ea95-4960-ae48-d17e9484ee2f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:38Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:38 crc kubenswrapper[4885]: I1002 01:47:38.750507 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:38 crc kubenswrapper[4885]: I1002 01:47:38.750561 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:38 crc kubenswrapper[4885]: I1002 01:47:38.750578 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:38 crc kubenswrapper[4885]: I1002 01:47:38.750599 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:38 crc kubenswrapper[4885]: I1002 01:47:38.750615 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:38Z","lastTransitionTime":"2025-10-02T01:47:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:38 crc kubenswrapper[4885]: E1002 01:47:38.769783 4885 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:47:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:47:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:47:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:47:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"92603ae5-ba38-4201-98e9-979fd56cb50d\\\",\\\"systemUUID\\\":\\\"50ddf2c2-ea95-4960-ae48-d17e9484ee2f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:38Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:38 crc kubenswrapper[4885]: I1002 01:47:38.774938 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:38 crc kubenswrapper[4885]: I1002 01:47:38.775002 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:38 crc kubenswrapper[4885]: I1002 01:47:38.775021 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:38 crc kubenswrapper[4885]: I1002 01:47:38.775055 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:38 crc kubenswrapper[4885]: I1002 01:47:38.775074 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:38Z","lastTransitionTime":"2025-10-02T01:47:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:38 crc kubenswrapper[4885]: E1002 01:47:38.794782 4885 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:47:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:47:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:47:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:47:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"92603ae5-ba38-4201-98e9-979fd56cb50d\\\",\\\"systemUUID\\\":\\\"50ddf2c2-ea95-4960-ae48-d17e9484ee2f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:38Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:38 crc kubenswrapper[4885]: I1002 01:47:38.805511 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:38 crc kubenswrapper[4885]: I1002 01:47:38.805579 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:38 crc kubenswrapper[4885]: I1002 01:47:38.805842 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:38 crc kubenswrapper[4885]: I1002 01:47:38.805908 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:38 crc kubenswrapper[4885]: I1002 01:47:38.805929 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:38Z","lastTransitionTime":"2025-10-02T01:47:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:38 crc kubenswrapper[4885]: E1002 01:47:38.827112 4885 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:47:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:47:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:47:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:47:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"92603ae5-ba38-4201-98e9-979fd56cb50d\\\",\\\"systemUUID\\\":\\\"50ddf2c2-ea95-4960-ae48-d17e9484ee2f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:38Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:38 crc kubenswrapper[4885]: I1002 01:47:38.831503 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:38 crc kubenswrapper[4885]: I1002 01:47:38.831547 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:38 crc kubenswrapper[4885]: I1002 01:47:38.831573 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:38 crc kubenswrapper[4885]: I1002 01:47:38.831601 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:38 crc kubenswrapper[4885]: I1002 01:47:38.831619 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:38Z","lastTransitionTime":"2025-10-02T01:47:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:38 crc kubenswrapper[4885]: E1002 01:47:38.852937 4885 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:47:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:47:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:47:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:47:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"92603ae5-ba38-4201-98e9-979fd56cb50d\\\",\\\"systemUUID\\\":\\\"50ddf2c2-ea95-4960-ae48-d17e9484ee2f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:38Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:38 crc kubenswrapper[4885]: E1002 01:47:38.853179 4885 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 02 01:47:38 crc kubenswrapper[4885]: I1002 01:47:38.855534 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:38 crc kubenswrapper[4885]: I1002 01:47:38.855607 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:38 crc kubenswrapper[4885]: I1002 01:47:38.855628 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:38 crc kubenswrapper[4885]: I1002 01:47:38.855656 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:38 crc kubenswrapper[4885]: I1002 01:47:38.855675 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:38Z","lastTransitionTime":"2025-10-02T01:47:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:38 crc kubenswrapper[4885]: I1002 01:47:38.959070 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:38 crc kubenswrapper[4885]: I1002 01:47:38.959140 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:38 crc kubenswrapper[4885]: I1002 01:47:38.959158 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:38 crc kubenswrapper[4885]: I1002 01:47:38.959185 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:38 crc kubenswrapper[4885]: I1002 01:47:38.959204 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:38Z","lastTransitionTime":"2025-10-02T01:47:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:39 crc kubenswrapper[4885]: I1002 01:47:39.062066 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:39 crc kubenswrapper[4885]: I1002 01:47:39.062164 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:39 crc kubenswrapper[4885]: I1002 01:47:39.062184 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:39 crc kubenswrapper[4885]: I1002 01:47:39.062212 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:39 crc kubenswrapper[4885]: I1002 01:47:39.062255 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:39Z","lastTransitionTime":"2025-10-02T01:47:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:39 crc kubenswrapper[4885]: I1002 01:47:39.165547 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:39 crc kubenswrapper[4885]: I1002 01:47:39.165606 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:39 crc kubenswrapper[4885]: I1002 01:47:39.165622 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:39 crc kubenswrapper[4885]: I1002 01:47:39.165646 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:39 crc kubenswrapper[4885]: I1002 01:47:39.165666 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:39Z","lastTransitionTime":"2025-10-02T01:47:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:39 crc kubenswrapper[4885]: I1002 01:47:39.269140 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:39 crc kubenswrapper[4885]: I1002 01:47:39.269325 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:39 crc kubenswrapper[4885]: I1002 01:47:39.269358 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:39 crc kubenswrapper[4885]: I1002 01:47:39.269387 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:39 crc kubenswrapper[4885]: I1002 01:47:39.269417 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:39Z","lastTransitionTime":"2025-10-02T01:47:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:39 crc kubenswrapper[4885]: I1002 01:47:39.372805 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:39 crc kubenswrapper[4885]: I1002 01:47:39.372873 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:39 crc kubenswrapper[4885]: I1002 01:47:39.372889 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:39 crc kubenswrapper[4885]: I1002 01:47:39.372918 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:39 crc kubenswrapper[4885]: I1002 01:47:39.372938 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:39Z","lastTransitionTime":"2025-10-02T01:47:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:39 crc kubenswrapper[4885]: I1002 01:47:39.475643 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:39 crc kubenswrapper[4885]: I1002 01:47:39.475702 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:39 crc kubenswrapper[4885]: I1002 01:47:39.475719 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:39 crc kubenswrapper[4885]: I1002 01:47:39.475743 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:39 crc kubenswrapper[4885]: I1002 01:47:39.475761 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:39Z","lastTransitionTime":"2025-10-02T01:47:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:39 crc kubenswrapper[4885]: I1002 01:47:39.581359 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:39 crc kubenswrapper[4885]: I1002 01:47:39.581441 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:39 crc kubenswrapper[4885]: I1002 01:47:39.581467 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:39 crc kubenswrapper[4885]: I1002 01:47:39.581499 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:39 crc kubenswrapper[4885]: I1002 01:47:39.581523 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:39Z","lastTransitionTime":"2025-10-02T01:47:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:39 crc kubenswrapper[4885]: I1002 01:47:39.684582 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:39 crc kubenswrapper[4885]: I1002 01:47:39.684645 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:39 crc kubenswrapper[4885]: I1002 01:47:39.684662 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:39 crc kubenswrapper[4885]: I1002 01:47:39.684687 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:39 crc kubenswrapper[4885]: I1002 01:47:39.684705 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:39Z","lastTransitionTime":"2025-10-02T01:47:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:39 crc kubenswrapper[4885]: I1002 01:47:39.788084 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:39 crc kubenswrapper[4885]: I1002 01:47:39.788167 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:39 crc kubenswrapper[4885]: I1002 01:47:39.788190 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:39 crc kubenswrapper[4885]: I1002 01:47:39.788220 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:39 crc kubenswrapper[4885]: I1002 01:47:39.788245 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:39Z","lastTransitionTime":"2025-10-02T01:47:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:39 crc kubenswrapper[4885]: I1002 01:47:39.891054 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:39 crc kubenswrapper[4885]: I1002 01:47:39.891134 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:39 crc kubenswrapper[4885]: I1002 01:47:39.891157 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:39 crc kubenswrapper[4885]: I1002 01:47:39.891188 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:39 crc kubenswrapper[4885]: I1002 01:47:39.891213 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:39Z","lastTransitionTime":"2025-10-02T01:47:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:39 crc kubenswrapper[4885]: I1002 01:47:39.994160 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:39 crc kubenswrapper[4885]: I1002 01:47:39.994216 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:39 crc kubenswrapper[4885]: I1002 01:47:39.994233 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:39 crc kubenswrapper[4885]: I1002 01:47:39.994256 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:39 crc kubenswrapper[4885]: I1002 01:47:39.994303 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:39Z","lastTransitionTime":"2025-10-02T01:47:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:40 crc kubenswrapper[4885]: I1002 01:47:40.046448 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sv4md" Oct 02 01:47:40 crc kubenswrapper[4885]: I1002 01:47:40.046531 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:47:40 crc kubenswrapper[4885]: I1002 01:47:40.046953 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:47:40 crc kubenswrapper[4885]: E1002 01:47:40.047031 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sv4md" podUID="724a3e49-001c-45a4-a896-82c13e0ee0e7" Oct 02 01:47:40 crc kubenswrapper[4885]: I1002 01:47:40.047193 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:47:40 crc kubenswrapper[4885]: I1002 01:47:40.047445 4885 scope.go:117] "RemoveContainer" containerID="0f85438a1b57b0a590e4d005978fe70ff8c891980badb20d65e1f0e14080fae6" Oct 02 01:47:40 crc kubenswrapper[4885]: E1002 01:47:40.047437 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:47:40 crc kubenswrapper[4885]: E1002 01:47:40.047686 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:47:40 crc kubenswrapper[4885]: E1002 01:47:40.047822 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:47:40 crc kubenswrapper[4885]: I1002 01:47:40.068826 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a0bb885-2eb0-419c-8a7a-ef634cb3c144\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a89b49cbe08d927629564e607de13655ddf3d01ac4eb5d688b325210306c120c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06c5942805a498d9621a906088aa54538b4eed583d1b8d0e3ba2a6718378a36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b9f6f7fe44d5771a838d0b89c5bdea3ebb79023ae2675ee907ec2766b3d35a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://178b4340ee6904c60c814b99d83ee225161aeb23e438d8d02f856579a13ec4e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bb63bf30d6ecce9e71f0db0c16632bdfa6b33924b673378eb0c38ab229fa837\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T01:47:06Z\\\",\\\"message\\\":\\\"W1002 01:46:55.406027 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1002 01:46:55.406714 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759369615 cert, and key in /tmp/serving-cert-3795040473/serving-signer.crt, /tmp/serving-cert-3795040473/serving-signer.key\\\\nI1002 01:46:55.614670 1 observer_polling.go:159] Starting file observer\\\\nW1002 01:46:55.617662 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1002 01:46:55.618112 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 01:46:55.620347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3795040473/tls.crt::/tmp/serving-cert-3795040473/tls.key\\\\\\\"\\\\nF1002 01:47:06.224459 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90535f97ab2ed4626d47540ebd7baada837fd332140323bec0f49aa38d3c0009\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://592f89a326c9c1559193a71904b47ca27b1dbe1a7b30e03bc48924bb2f35f921\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://592f89a326c9c1559193a71904b47ca27b1dbe1a7b30e03bc48924bb2f35f921\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:40Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:40 crc kubenswrapper[4885]: I1002 01:47:40.088650 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b087880e05039816dcab154643ac72c7e4a0a1124817c51dc2c639b62bbee0fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:40Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:40 crc kubenswrapper[4885]: I1002 01:47:40.097116 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:40 crc kubenswrapper[4885]: I1002 01:47:40.097168 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:40 crc kubenswrapper[4885]: I1002 01:47:40.097185 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:40 crc kubenswrapper[4885]: I1002 01:47:40.097208 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:40 crc kubenswrapper[4885]: I1002 01:47:40.097227 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:40Z","lastTransitionTime":"2025-10-02T01:47:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:40 crc kubenswrapper[4885]: I1002 01:47:40.109762 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:40Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:40 crc kubenswrapper[4885]: I1002 01:47:40.132640 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:40Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:40 crc kubenswrapper[4885]: I1002 01:47:40.149906 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-58lhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac6a19c4-f259-4dbd-a016-d6c2c849c6cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed76c0f968c063f7795d363af194a222487b647ceda234de0dabb39b25eafe0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hzc4c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-58lhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:40Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:40 crc kubenswrapper[4885]: I1002 01:47:40.167748 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d169c8ba-000a-4218-8e3b-4ae53035b110\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61e436b7876107ce5ea9e668c0bbc367be6e663fc804fa3a65b51adf8be8d51f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-77rzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09536e182a46e77318c10812d1777e551f834907c006cab5c726dc66f5bce88a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-77rzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rttx8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:40Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:40 crc kubenswrapper[4885]: I1002 01:47:40.184810 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jgdvt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81733e8d-67b0-46fa-8a41-08b40645bd3e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcef60c816ca1ba9a298da658e84c4d717bea3d352c4c2f807ebd8e56327b384\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cwbl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jgdvt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:40Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:40 crc kubenswrapper[4885]: I1002 01:47:40.200815 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:40 crc kubenswrapper[4885]: I1002 01:47:40.200872 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:40 crc kubenswrapper[4885]: I1002 01:47:40.200889 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:40 crc kubenswrapper[4885]: I1002 01:47:40.200915 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:40 crc kubenswrapper[4885]: I1002 01:47:40.200933 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:40Z","lastTransitionTime":"2025-10-02T01:47:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:40 crc kubenswrapper[4885]: I1002 01:47:40.205778 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"506a599f-fb1c-4d14-be88-0e4833f3d624\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31c2f0193ffef54965b79a3569b0f6f4547e540bac0774155159a1c996c196c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://533e73e5f90b7258acbcdf018ee01323c118f9dc45035b25b065ba01cabecccf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c00db833a593ec40eb63a6b383455e19e960c2e98b9b66f76877ee993057e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21e2f462a0deb0d740185663b2014f4a1e57b4fdbeb82571264b002f928035bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:40Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:40 crc kubenswrapper[4885]: I1002 01:47:40.237489 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e4679dd-f870-41e9-a32b-360f5d2ee81b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4310909ad9c7b609ed6814e04a525a6afae747775598e7275080413dec8b23a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7642033b0b5995d332ccf7992d6d7e6a5e9c9369ae2d1071ccb2f536938c8f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9f99f9e26a04fb42b01a0783a5c4e360514818ce9102edc7fa08626d50b9a34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6fa72b030d309656b73121e4ac2a3924a4e449aad66683dd9de5ee5f22d5a4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef403e519d197fb37b3e6b76a14a5128ba3f30a7a5e8bf0a420cf6a39b89575f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8482ba457683ffa03100ec1f07ea9d3f918be5e490b5ba05c2ba965d03033191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f85438a1b57b0a590e4d005978fe70ff8c891980badb20d65e1f0e14080fae6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f85438a1b57b0a590e4d005978fe70ff8c891980badb20d65e1f0e14080fae6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T01:47:25Z\\\",\\\"message\\\":\\\"dressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-api/machine-api-operator-machine-webhook_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-operator-machine-webhook\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.250\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1002 01:47:25.452713 6353 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1002 01:47:25.452745 6353 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-source-55646444c4-trplf] creating logical port openshift-network-diagnostics_network-check-source-55646444c4-trplf for pod on switch crc\\\\nF1002 01:47:25.452781 6353 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:24Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-dngcm_openshift-ovn-kubernetes(9e4679dd-f870-41e9-a32b-360f5d2ee81b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fafdd1094bd02786d745c984989852b0c690db38655f92b419733a2004f417c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0125e907d061a97286026ee823c4eed303734ed03ab4864f742592b23a1aa2f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0125e907d061a97286026ee823c4eed303734ed03ab4864f742592b23a1aa2f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dngcm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:40Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:40 crc kubenswrapper[4885]: I1002 01:47:40.254828 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-sv4md" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"724a3e49-001c-45a4-a896-82c13e0ee0e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hpqgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hpqgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-sv4md\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:40Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:40 crc kubenswrapper[4885]: I1002 01:47:40.291236 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8070bc1c-e3de-4e9a-985e-40d1a9858056\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3e634b3727c517cc48b17c2cfb38f26460f53d1bf25d631d08ec9acb8d780e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4adcc537c23697a324d82af94c31059dd444245efdf7c978db58ede256e112ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edcd1d99644a778dfb6d0a17d2a4abd906400fe8209e611ebf8d7a9abe7ccf28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://97cb59adc2ddfa365ebc9dec649600dc90466ffe59134a6d15a9d3134a7c9b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d2808bc658ea80d68469ed16422d56ef19447f4c90b4af7b2597c64406f8a44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://83acbc20919f7c1958b6d492fadb203855df02761662425df9f91a0757fe2bf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83acbc20919f7c1958b6d492fadb203855df02761662425df9f91a0757fe2bf7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e86ecd03835efc9a253b6804bc61e6ac690e75b8d1f3cc77f5647730e586574\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e86ecd03835efc9a253b6804bc61e6ac690e75b8d1f3cc77f5647730e586574\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://48f95af51feaf014a18496bdfadd59d7f1e11167c5c8b677fa7037e48368ff2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48f95af51feaf014a18496bdfadd59d7f1e11167c5c8b677fa7037e48368ff2a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:40Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:40 crc kubenswrapper[4885]: I1002 01:47:40.306895 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:40 crc kubenswrapper[4885]: I1002 01:47:40.306973 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:40 crc kubenswrapper[4885]: I1002 01:47:40.307000 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:40 crc kubenswrapper[4885]: I1002 01:47:40.307034 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:40 crc kubenswrapper[4885]: I1002 01:47:40.307061 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:40Z","lastTransitionTime":"2025-10-02T01:47:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:40 crc kubenswrapper[4885]: I1002 01:47:40.312362 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5c1f95f188569c9aa47e51f2916939fa0b65a5cdf4c16b798612d41f3c8411a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:40Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:40 crc kubenswrapper[4885]: I1002 01:47:40.333037 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vhgbd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c63a7d20d39b65acfb5a9be4c28b21c72111b398bea76c238e15ddb4dd63e139\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7zwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vhgbd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:40Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:40 crc kubenswrapper[4885]: I1002 01:47:40.351003 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rrn8b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52967a77-525c-4537-9b1f-d7bb754a8494\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3dee451ebe9006bf1c519fed4c65267570b6139d76841858352e087ae78d26a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtr74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e9412788d6df7332592beac48281822b26e1aa19b39addce8e1be556af9d982\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtr74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:26Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rrn8b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:40Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:40 crc kubenswrapper[4885]: I1002 01:47:40.371048 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:40Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:40 crc kubenswrapper[4885]: I1002 01:47:40.396855 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74adcf912ccedb6b5d4f2803fd215e3e2ef62367597aa6da528354c4511ee536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7403ef968d040ac09d722b7557438bef23892d578992826f42b700b370d495d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:40Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:40 crc kubenswrapper[4885]: I1002 01:47:40.409994 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:40 crc kubenswrapper[4885]: I1002 01:47:40.410042 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:40 crc kubenswrapper[4885]: I1002 01:47:40.410053 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:40 crc kubenswrapper[4885]: I1002 01:47:40.410073 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:40 crc kubenswrapper[4885]: I1002 01:47:40.410085 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:40Z","lastTransitionTime":"2025-10-02T01:47:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:40 crc kubenswrapper[4885]: I1002 01:47:40.420544 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2ll9b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52b39712-5fa1-4fe0-814b-b170a6e3938b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9a96383d0fbfd116c419054aaf1a7adba1937e8b0bad294cefd68b4ab0d2298\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c979b91ba748aa4647826a8926337662ca929c54090fb62285572819ec2fbf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c979b91ba748aa4647826a8926337662ca929c54090fb62285572819ec2fbf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97d9d693e8e60ee9c183bb44461e698e8b3cfd20a516605aa8754c9205d2bdc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://97d9d693e8e60ee9c183bb44461e698e8b3cfd20a516605aa8754c9205d2bdc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c44252676359c671d65c9588b05572d76c99cf4c90666f0ac089b2f778450a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c44252676359c671d65c9588b05572d76c99cf4c90666f0ac089b2f778450a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80c982f07874df686fe7e29cc91043eb6d58f1be5fc93cacd46d33a21dd9b622\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80c982f07874df686fe7e29cc91043eb6d58f1be5fc93cacd46d33a21dd9b622\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a354c985652c7ad7fa8944fbfdfe223cfb0f97fb11eecd9bd429020d4892b78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a354c985652c7ad7fa8944fbfdfe223cfb0f97fb11eecd9bd429020d4892b78\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2efd7283df313d086c1b8b88f289ace28764de256205bee7610d6711002d8515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2efd7283df313d086c1b8b88f289ace28764de256205bee7610d6711002d8515\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2ll9b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:40Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:40 crc kubenswrapper[4885]: I1002 01:47:40.478224 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dngcm_9e4679dd-f870-41e9-a32b-360f5d2ee81b/ovnkube-controller/1.log" Oct 02 01:47:40 crc kubenswrapper[4885]: I1002 01:47:40.489446 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" event={"ID":"9e4679dd-f870-41e9-a32b-360f5d2ee81b","Type":"ContainerStarted","Data":"188cb50b58926683ce8862d618258c1d1a6f33310608546c8651ece28f7c338f"} Oct 02 01:47:40 crc kubenswrapper[4885]: I1002 01:47:40.489777 4885 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 02 01:47:40 crc kubenswrapper[4885]: I1002 01:47:40.512378 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:40Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:40 crc kubenswrapper[4885]: I1002 01:47:40.512975 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:40 crc kubenswrapper[4885]: I1002 01:47:40.513045 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:40 crc kubenswrapper[4885]: I1002 01:47:40.513069 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:40 crc kubenswrapper[4885]: I1002 01:47:40.513095 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:40 crc kubenswrapper[4885]: I1002 01:47:40.513116 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:40Z","lastTransitionTime":"2025-10-02T01:47:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:40 crc kubenswrapper[4885]: I1002 01:47:40.533095 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74adcf912ccedb6b5d4f2803fd215e3e2ef62367597aa6da528354c4511ee536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7403ef968d040ac09d722b7557438bef23892d578992826f42b700b370d495d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:40Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:40 crc kubenswrapper[4885]: I1002 01:47:40.560830 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2ll9b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52b39712-5fa1-4fe0-814b-b170a6e3938b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9a96383d0fbfd116c419054aaf1a7adba1937e8b0bad294cefd68b4ab0d2298\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c979b91ba748aa4647826a8926337662ca929c54090fb62285572819ec2fbf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c979b91ba748aa4647826a8926337662ca929c54090fb62285572819ec2fbf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97d9d693e8e60ee9c183bb44461e698e8b3cfd20a516605aa8754c9205d2bdc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://97d9d693e8e60ee9c183bb44461e698e8b3cfd20a516605aa8754c9205d2bdc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c44252676359c671d65c9588b05572d76c99cf4c90666f0ac089b2f778450a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c44252676359c671d65c9588b05572d76c99cf4c90666f0ac089b2f778450a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80c982f07874df686fe7e29cc91043eb6d58f1be5fc93cacd46d33a21dd9b622\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80c982f07874df686fe7e29cc91043eb6d58f1be5fc93cacd46d33a21dd9b622\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a354c985652c7ad7fa8944fbfdfe223cfb0f97fb11eecd9bd429020d4892b78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a354c985652c7ad7fa8944fbfdfe223cfb0f97fb11eecd9bd429020d4892b78\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2efd7283df313d086c1b8b88f289ace28764de256205bee7610d6711002d8515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2efd7283df313d086c1b8b88f289ace28764de256205bee7610d6711002d8515\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2ll9b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:40Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:40 crc kubenswrapper[4885]: I1002 01:47:40.576377 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:40Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:40 crc kubenswrapper[4885]: I1002 01:47:40.598088 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-58lhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac6a19c4-f259-4dbd-a016-d6c2c849c6cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed76c0f968c063f7795d363af194a222487b647ceda234de0dabb39b25eafe0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hzc4c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-58lhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:40Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:40 crc kubenswrapper[4885]: I1002 01:47:40.612391 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d169c8ba-000a-4218-8e3b-4ae53035b110\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61e436b7876107ce5ea9e668c0bbc367be6e663fc804fa3a65b51adf8be8d51f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-77rzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09536e182a46e77318c10812d1777e551f834907c006cab5c726dc66f5bce88a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-77rzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rttx8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:40Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:40 crc kubenswrapper[4885]: I1002 01:47:40.615129 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:40 crc kubenswrapper[4885]: I1002 01:47:40.615191 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:40 crc kubenswrapper[4885]: I1002 01:47:40.615211 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:40 crc kubenswrapper[4885]: I1002 01:47:40.615234 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:40 crc kubenswrapper[4885]: I1002 01:47:40.615250 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:40Z","lastTransitionTime":"2025-10-02T01:47:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:40 crc kubenswrapper[4885]: I1002 01:47:40.625842 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jgdvt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81733e8d-67b0-46fa-8a41-08b40645bd3e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcef60c816ca1ba9a298da658e84c4d717bea3d352c4c2f807ebd8e56327b384\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cwbl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jgdvt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:40Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:40 crc kubenswrapper[4885]: I1002 01:47:40.653606 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a0bb885-2eb0-419c-8a7a-ef634cb3c144\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a89b49cbe08d927629564e607de13655ddf3d01ac4eb5d688b325210306c120c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06c5942805a498d9621a906088aa54538b4eed583d1b8d0e3ba2a6718378a36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b9f6f7fe44d5771a838d0b89c5bdea3ebb79023ae2675ee907ec2766b3d35a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://178b4340ee6904c60c814b99d83ee225161aeb23e438d8d02f856579a13ec4e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bb63bf30d6ecce9e71f0db0c16632bdfa6b33924b673378eb0c38ab229fa837\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T01:47:06Z\\\",\\\"message\\\":\\\"W1002 01:46:55.406027 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1002 01:46:55.406714 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759369615 cert, and key in /tmp/serving-cert-3795040473/serving-signer.crt, /tmp/serving-cert-3795040473/serving-signer.key\\\\nI1002 01:46:55.614670 1 observer_polling.go:159] Starting file observer\\\\nW1002 01:46:55.617662 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1002 01:46:55.618112 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 01:46:55.620347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3795040473/tls.crt::/tmp/serving-cert-3795040473/tls.key\\\\\\\"\\\\nF1002 01:47:06.224459 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90535f97ab2ed4626d47540ebd7baada837fd332140323bec0f49aa38d3c0009\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://592f89a326c9c1559193a71904b47ca27b1dbe1a7b30e03bc48924bb2f35f921\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://592f89a326c9c1559193a71904b47ca27b1dbe1a7b30e03bc48924bb2f35f921\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:40Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:40 crc kubenswrapper[4885]: I1002 01:47:40.670528 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b087880e05039816dcab154643ac72c7e4a0a1124817c51dc2c639b62bbee0fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:40Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:40 crc kubenswrapper[4885]: I1002 01:47:40.683643 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:40Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:40 crc kubenswrapper[4885]: I1002 01:47:40.697458 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"506a599f-fb1c-4d14-be88-0e4833f3d624\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31c2f0193ffef54965b79a3569b0f6f4547e540bac0774155159a1c996c196c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://533e73e5f90b7258acbcdf018ee01323c118f9dc45035b25b065ba01cabecccf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c00db833a593ec40eb63a6b383455e19e960c2e98b9b66f76877ee993057e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21e2f462a0deb0d740185663b2014f4a1e57b4fdbeb82571264b002f928035bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:40Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:40 crc kubenswrapper[4885]: I1002 01:47:40.717957 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:40 crc kubenswrapper[4885]: I1002 01:47:40.718029 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:40 crc kubenswrapper[4885]: I1002 01:47:40.718041 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:40 crc kubenswrapper[4885]: I1002 01:47:40.718060 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:40 crc kubenswrapper[4885]: I1002 01:47:40.718074 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:40Z","lastTransitionTime":"2025-10-02T01:47:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:40 crc kubenswrapper[4885]: I1002 01:47:40.720708 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e4679dd-f870-41e9-a32b-360f5d2ee81b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4310909ad9c7b609ed6814e04a525a6afae747775598e7275080413dec8b23a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7642033b0b5995d332ccf7992d6d7e6a5e9c9369ae2d1071ccb2f536938c8f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9f99f9e26a04fb42b01a0783a5c4e360514818ce9102edc7fa08626d50b9a34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6fa72b030d309656b73121e4ac2a3924a4e449aad66683dd9de5ee5f22d5a4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef403e519d197fb37b3e6b76a14a5128ba3f30a7a5e8bf0a420cf6a39b89575f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8482ba457683ffa03100ec1f07ea9d3f918be5e490b5ba05c2ba965d03033191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://188cb50b58926683ce8862d618258c1d1a6f33310608546c8651ece28f7c338f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f85438a1b57b0a590e4d005978fe70ff8c891980badb20d65e1f0e14080fae6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T01:47:25Z\\\",\\\"message\\\":\\\"dressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-api/machine-api-operator-machine-webhook_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-operator-machine-webhook\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.250\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1002 01:47:25.452713 6353 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1002 01:47:25.452745 6353 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-source-55646444c4-trplf] creating logical port openshift-network-diagnostics_network-check-source-55646444c4-trplf for pod on switch crc\\\\nF1002 01:47:25.452781 6353 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:24Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fafdd1094bd02786d745c984989852b0c690db38655f92b419733a2004f417c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0125e907d061a97286026ee823c4eed303734ed03ab4864f742592b23a1aa2f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0125e907d061a97286026ee823c4eed303734ed03ab4864f742592b23a1aa2f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dngcm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:40Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:40 crc kubenswrapper[4885]: I1002 01:47:40.736616 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-sv4md" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"724a3e49-001c-45a4-a896-82c13e0ee0e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hpqgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hpqgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-sv4md\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:40Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:40 crc kubenswrapper[4885]: I1002 01:47:40.751311 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rrn8b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52967a77-525c-4537-9b1f-d7bb754a8494\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3dee451ebe9006bf1c519fed4c65267570b6139d76841858352e087ae78d26a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtr74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e9412788d6df7332592beac48281822b26e1aa19b39addce8e1be556af9d982\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtr74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:26Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rrn8b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:40Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:40 crc kubenswrapper[4885]: I1002 01:47:40.779948 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8070bc1c-e3de-4e9a-985e-40d1a9858056\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3e634b3727c517cc48b17c2cfb38f26460f53d1bf25d631d08ec9acb8d780e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4adcc537c23697a324d82af94c31059dd444245efdf7c978db58ede256e112ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edcd1d99644a778dfb6d0a17d2a4abd906400fe8209e611ebf8d7a9abe7ccf28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://97cb59adc2ddfa365ebc9dec649600dc90466ffe59134a6d15a9d3134a7c9b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d2808bc658ea80d68469ed16422d56ef19447f4c90b4af7b2597c64406f8a44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://83acbc20919f7c1958b6d492fadb203855df02761662425df9f91a0757fe2bf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83acbc20919f7c1958b6d492fadb203855df02761662425df9f91a0757fe2bf7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e86ecd03835efc9a253b6804bc61e6ac690e75b8d1f3cc77f5647730e586574\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e86ecd03835efc9a253b6804bc61e6ac690e75b8d1f3cc77f5647730e586574\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://48f95af51feaf014a18496bdfadd59d7f1e11167c5c8b677fa7037e48368ff2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48f95af51feaf014a18496bdfadd59d7f1e11167c5c8b677fa7037e48368ff2a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:40Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:40 crc kubenswrapper[4885]: I1002 01:47:40.794427 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5c1f95f188569c9aa47e51f2916939fa0b65a5cdf4c16b798612d41f3c8411a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:40Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:40 crc kubenswrapper[4885]: I1002 01:47:40.813214 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vhgbd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c63a7d20d39b65acfb5a9be4c28b21c72111b398bea76c238e15ddb4dd63e139\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7zwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vhgbd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:40Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:40 crc kubenswrapper[4885]: I1002 01:47:40.820993 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:40 crc kubenswrapper[4885]: I1002 01:47:40.821027 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:40 crc kubenswrapper[4885]: I1002 01:47:40.821035 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:40 crc kubenswrapper[4885]: I1002 01:47:40.821051 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:40 crc kubenswrapper[4885]: I1002 01:47:40.821061 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:40Z","lastTransitionTime":"2025-10-02T01:47:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:40 crc kubenswrapper[4885]: I1002 01:47:40.924076 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:40 crc kubenswrapper[4885]: I1002 01:47:40.924117 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:40 crc kubenswrapper[4885]: I1002 01:47:40.924128 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:40 crc kubenswrapper[4885]: I1002 01:47:40.924143 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:40 crc kubenswrapper[4885]: I1002 01:47:40.924152 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:40Z","lastTransitionTime":"2025-10-02T01:47:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:41 crc kubenswrapper[4885]: I1002 01:47:41.025950 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:41 crc kubenswrapper[4885]: I1002 01:47:41.025984 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:41 crc kubenswrapper[4885]: I1002 01:47:41.025993 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:41 crc kubenswrapper[4885]: I1002 01:47:41.026007 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:41 crc kubenswrapper[4885]: I1002 01:47:41.026017 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:41Z","lastTransitionTime":"2025-10-02T01:47:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:41 crc kubenswrapper[4885]: I1002 01:47:41.128155 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:41 crc kubenswrapper[4885]: I1002 01:47:41.128186 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:41 crc kubenswrapper[4885]: I1002 01:47:41.128195 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:41 crc kubenswrapper[4885]: I1002 01:47:41.128207 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:41 crc kubenswrapper[4885]: I1002 01:47:41.128222 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:41Z","lastTransitionTime":"2025-10-02T01:47:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:41 crc kubenswrapper[4885]: I1002 01:47:41.231621 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:41 crc kubenswrapper[4885]: I1002 01:47:41.231673 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:41 crc kubenswrapper[4885]: I1002 01:47:41.231689 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:41 crc kubenswrapper[4885]: I1002 01:47:41.231709 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:41 crc kubenswrapper[4885]: I1002 01:47:41.231724 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:41Z","lastTransitionTime":"2025-10-02T01:47:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:41 crc kubenswrapper[4885]: I1002 01:47:41.336121 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:41 crc kubenswrapper[4885]: I1002 01:47:41.336206 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:41 crc kubenswrapper[4885]: I1002 01:47:41.336228 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:41 crc kubenswrapper[4885]: I1002 01:47:41.336290 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:41 crc kubenswrapper[4885]: I1002 01:47:41.336318 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:41Z","lastTransitionTime":"2025-10-02T01:47:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:41 crc kubenswrapper[4885]: I1002 01:47:41.439696 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:41 crc kubenswrapper[4885]: I1002 01:47:41.439742 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:41 crc kubenswrapper[4885]: I1002 01:47:41.439758 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:41 crc kubenswrapper[4885]: I1002 01:47:41.439782 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:41 crc kubenswrapper[4885]: I1002 01:47:41.439799 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:41Z","lastTransitionTime":"2025-10-02T01:47:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:41 crc kubenswrapper[4885]: I1002 01:47:41.495854 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dngcm_9e4679dd-f870-41e9-a32b-360f5d2ee81b/ovnkube-controller/2.log" Oct 02 01:47:41 crc kubenswrapper[4885]: I1002 01:47:41.497330 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dngcm_9e4679dd-f870-41e9-a32b-360f5d2ee81b/ovnkube-controller/1.log" Oct 02 01:47:41 crc kubenswrapper[4885]: I1002 01:47:41.503695 4885 generic.go:334] "Generic (PLEG): container finished" podID="9e4679dd-f870-41e9-a32b-360f5d2ee81b" containerID="188cb50b58926683ce8862d618258c1d1a6f33310608546c8651ece28f7c338f" exitCode=1 Oct 02 01:47:41 crc kubenswrapper[4885]: I1002 01:47:41.503798 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" event={"ID":"9e4679dd-f870-41e9-a32b-360f5d2ee81b","Type":"ContainerDied","Data":"188cb50b58926683ce8862d618258c1d1a6f33310608546c8651ece28f7c338f"} Oct 02 01:47:41 crc kubenswrapper[4885]: I1002 01:47:41.503892 4885 scope.go:117] "RemoveContainer" containerID="0f85438a1b57b0a590e4d005978fe70ff8c891980badb20d65e1f0e14080fae6" Oct 02 01:47:41 crc kubenswrapper[4885]: I1002 01:47:41.511657 4885 scope.go:117] "RemoveContainer" containerID="188cb50b58926683ce8862d618258c1d1a6f33310608546c8651ece28f7c338f" Oct 02 01:47:41 crc kubenswrapper[4885]: E1002 01:47:41.512013 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-dngcm_openshift-ovn-kubernetes(9e4679dd-f870-41e9-a32b-360f5d2ee81b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" podUID="9e4679dd-f870-41e9-a32b-360f5d2ee81b" Oct 02 01:47:41 crc kubenswrapper[4885]: I1002 01:47:41.528083 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d169c8ba-000a-4218-8e3b-4ae53035b110\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61e436b7876107ce5ea9e668c0bbc367be6e663fc804fa3a65b51adf8be8d51f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-77rzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09536e182a46e77318c10812d1777e551f834907c006cab5c726dc66f5bce88a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-77rzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rttx8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:41Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:41 crc kubenswrapper[4885]: I1002 01:47:41.543358 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:41 crc kubenswrapper[4885]: I1002 01:47:41.543417 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:41 crc kubenswrapper[4885]: I1002 01:47:41.543434 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:41 crc kubenswrapper[4885]: I1002 01:47:41.543460 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:41 crc kubenswrapper[4885]: I1002 01:47:41.543478 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:41Z","lastTransitionTime":"2025-10-02T01:47:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:41 crc kubenswrapper[4885]: I1002 01:47:41.544119 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jgdvt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81733e8d-67b0-46fa-8a41-08b40645bd3e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcef60c816ca1ba9a298da658e84c4d717bea3d352c4c2f807ebd8e56327b384\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cwbl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jgdvt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:41Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:41 crc kubenswrapper[4885]: I1002 01:47:41.566121 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a0bb885-2eb0-419c-8a7a-ef634cb3c144\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a89b49cbe08d927629564e607de13655ddf3d01ac4eb5d688b325210306c120c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06c5942805a498d9621a906088aa54538b4eed583d1b8d0e3ba2a6718378a36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b9f6f7fe44d5771a838d0b89c5bdea3ebb79023ae2675ee907ec2766b3d35a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://178b4340ee6904c60c814b99d83ee225161aeb23e438d8d02f856579a13ec4e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bb63bf30d6ecce9e71f0db0c16632bdfa6b33924b673378eb0c38ab229fa837\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T01:47:06Z\\\",\\\"message\\\":\\\"W1002 01:46:55.406027 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1002 01:46:55.406714 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759369615 cert, and key in /tmp/serving-cert-3795040473/serving-signer.crt, /tmp/serving-cert-3795040473/serving-signer.key\\\\nI1002 01:46:55.614670 1 observer_polling.go:159] Starting file observer\\\\nW1002 01:46:55.617662 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1002 01:46:55.618112 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 01:46:55.620347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3795040473/tls.crt::/tmp/serving-cert-3795040473/tls.key\\\\\\\"\\\\nF1002 01:47:06.224459 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90535f97ab2ed4626d47540ebd7baada837fd332140323bec0f49aa38d3c0009\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://592f89a326c9c1559193a71904b47ca27b1dbe1a7b30e03bc48924bb2f35f921\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://592f89a326c9c1559193a71904b47ca27b1dbe1a7b30e03bc48924bb2f35f921\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:41Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:41 crc kubenswrapper[4885]: I1002 01:47:41.584650 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b087880e05039816dcab154643ac72c7e4a0a1124817c51dc2c639b62bbee0fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:41Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:41 crc kubenswrapper[4885]: I1002 01:47:41.602958 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:41Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:41 crc kubenswrapper[4885]: I1002 01:47:41.619659 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:41Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:41 crc kubenswrapper[4885]: I1002 01:47:41.635235 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-58lhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac6a19c4-f259-4dbd-a016-d6c2c849c6cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed76c0f968c063f7795d363af194a222487b647ceda234de0dabb39b25eafe0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hzc4c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-58lhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:41Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:41 crc kubenswrapper[4885]: I1002 01:47:41.646657 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:41 crc kubenswrapper[4885]: I1002 01:47:41.646722 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:41 crc kubenswrapper[4885]: I1002 01:47:41.646747 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:41 crc kubenswrapper[4885]: I1002 01:47:41.646778 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:41 crc kubenswrapper[4885]: I1002 01:47:41.646801 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:41Z","lastTransitionTime":"2025-10-02T01:47:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:41 crc kubenswrapper[4885]: I1002 01:47:41.654828 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"506a599f-fb1c-4d14-be88-0e4833f3d624\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31c2f0193ffef54965b79a3569b0f6f4547e540bac0774155159a1c996c196c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://533e73e5f90b7258acbcdf018ee01323c118f9dc45035b25b065ba01cabecccf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c00db833a593ec40eb63a6b383455e19e960c2e98b9b66f76877ee993057e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21e2f462a0deb0d740185663b2014f4a1e57b4fdbeb82571264b002f928035bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:41Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:41 crc kubenswrapper[4885]: I1002 01:47:41.686337 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e4679dd-f870-41e9-a32b-360f5d2ee81b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4310909ad9c7b609ed6814e04a525a6afae747775598e7275080413dec8b23a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7642033b0b5995d332ccf7992d6d7e6a5e9c9369ae2d1071ccb2f536938c8f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9f99f9e26a04fb42b01a0783a5c4e360514818ce9102edc7fa08626d50b9a34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6fa72b030d309656b73121e4ac2a3924a4e449aad66683dd9de5ee5f22d5a4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef403e519d197fb37b3e6b76a14a5128ba3f30a7a5e8bf0a420cf6a39b89575f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8482ba457683ffa03100ec1f07ea9d3f918be5e490b5ba05c2ba965d03033191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://188cb50b58926683ce8862d618258c1d1a6f33310608546c8651ece28f7c338f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f85438a1b57b0a590e4d005978fe70ff8c891980badb20d65e1f0e14080fae6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T01:47:25Z\\\",\\\"message\\\":\\\"dressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-api/machine-api-operator-machine-webhook_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-operator-machine-webhook\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.250\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1002 01:47:25.452713 6353 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1002 01:47:25.452745 6353 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-source-55646444c4-trplf] creating logical port openshift-network-diagnostics_network-check-source-55646444c4-trplf for pod on switch crc\\\\nF1002 01:47:25.452781 6353 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:24Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://188cb50b58926683ce8862d618258c1d1a6f33310608546c8651ece28f7c338f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T01:47:41Z\\\",\\\"message\\\":\\\"921 6578 obj_retry.go:434] periodicallyRetryResources: Retry channel got triggered: retrying failed objects of type *v1.Pod\\\\nI1002 01:47:40.991945 6578 obj_retry.go:409] Going to retry *v1.Pod resource setup for 16 objects: [openshift-multus/network-metrics-daemon-sv4md openshift-network-operator/iptables-alerter-4ln5h openshift-kube-apiserver/kube-apiserver-crc openshift-ovn-kubernetes/ovnkube-node-dngcm openshift-dns/node-resolver-58lhj openshift-machine-config-operator/machine-config-daemon-rttx8 openshift-multus/multus-vhgbd openshift-network-console/networking-console-plugin-85b44fc459-gdk6g openshift-network-diagnostics/network-check-source-55646444c4-trplf openshift-network-operator/network-operator-58b4c7f79c-55gtf openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rrn8b openshift-image-registry/node-ca-jgdvt openshift-kube-controller-manager/kube-controller-manager-crc openshift-multus/multus-additional-cni-plugins-2ll9b openshift-network-diagnostics/network-check-target-xd92c openshift-network-node-identity/network-node-identity-vrzqb]\\\\nF1002 01:47:40.991998 6578 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller ini\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fafdd1094bd02786d745c984989852b0c690db38655f92b419733a2004f417c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0125e907d061a97286026ee823c4eed303734ed03ab4864f742592b23a1aa2f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0125e907d061a97286026ee823c4eed303734ed03ab4864f742592b23a1aa2f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dngcm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:41Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:41 crc kubenswrapper[4885]: I1002 01:47:41.702358 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-sv4md" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"724a3e49-001c-45a4-a896-82c13e0ee0e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hpqgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hpqgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-sv4md\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:41Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:41 crc kubenswrapper[4885]: I1002 01:47:41.732361 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8070bc1c-e3de-4e9a-985e-40d1a9858056\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3e634b3727c517cc48b17c2cfb38f26460f53d1bf25d631d08ec9acb8d780e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4adcc537c23697a324d82af94c31059dd444245efdf7c978db58ede256e112ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edcd1d99644a778dfb6d0a17d2a4abd906400fe8209e611ebf8d7a9abe7ccf28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://97cb59adc2ddfa365ebc9dec649600dc90466ffe59134a6d15a9d3134a7c9b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d2808bc658ea80d68469ed16422d56ef19447f4c90b4af7b2597c64406f8a44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://83acbc20919f7c1958b6d492fadb203855df02761662425df9f91a0757fe2bf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83acbc20919f7c1958b6d492fadb203855df02761662425df9f91a0757fe2bf7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e86ecd03835efc9a253b6804bc61e6ac690e75b8d1f3cc77f5647730e586574\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e86ecd03835efc9a253b6804bc61e6ac690e75b8d1f3cc77f5647730e586574\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://48f95af51feaf014a18496bdfadd59d7f1e11167c5c8b677fa7037e48368ff2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48f95af51feaf014a18496bdfadd59d7f1e11167c5c8b677fa7037e48368ff2a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:41Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:41 crc kubenswrapper[4885]: I1002 01:47:41.750321 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5c1f95f188569c9aa47e51f2916939fa0b65a5cdf4c16b798612d41f3c8411a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:41Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:41 crc kubenswrapper[4885]: I1002 01:47:41.750396 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:41 crc kubenswrapper[4885]: I1002 01:47:41.750445 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:41 crc kubenswrapper[4885]: I1002 01:47:41.750508 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:41 crc kubenswrapper[4885]: I1002 01:47:41.750536 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:41 crc kubenswrapper[4885]: I1002 01:47:41.750554 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:41Z","lastTransitionTime":"2025-10-02T01:47:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:41 crc kubenswrapper[4885]: I1002 01:47:41.768454 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vhgbd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c63a7d20d39b65acfb5a9be4c28b21c72111b398bea76c238e15ddb4dd63e139\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7zwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vhgbd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:41Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:41 crc kubenswrapper[4885]: I1002 01:47:41.786494 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rrn8b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52967a77-525c-4537-9b1f-d7bb754a8494\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3dee451ebe9006bf1c519fed4c65267570b6139d76841858352e087ae78d26a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtr74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e9412788d6df7332592beac48281822b26e1aa19b39addce8e1be556af9d982\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtr74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:26Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rrn8b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:41Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:41 crc kubenswrapper[4885]: I1002 01:47:41.805607 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:41Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:41 crc kubenswrapper[4885]: I1002 01:47:41.824134 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74adcf912ccedb6b5d4f2803fd215e3e2ef62367597aa6da528354c4511ee536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7403ef968d040ac09d722b7557438bef23892d578992826f42b700b370d495d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:41Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:41 crc kubenswrapper[4885]: I1002 01:47:41.846611 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2ll9b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52b39712-5fa1-4fe0-814b-b170a6e3938b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9a96383d0fbfd116c419054aaf1a7adba1937e8b0bad294cefd68b4ab0d2298\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c979b91ba748aa4647826a8926337662ca929c54090fb62285572819ec2fbf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c979b91ba748aa4647826a8926337662ca929c54090fb62285572819ec2fbf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97d9d693e8e60ee9c183bb44461e698e8b3cfd20a516605aa8754c9205d2bdc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://97d9d693e8e60ee9c183bb44461e698e8b3cfd20a516605aa8754c9205d2bdc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c44252676359c671d65c9588b05572d76c99cf4c90666f0ac089b2f778450a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c44252676359c671d65c9588b05572d76c99cf4c90666f0ac089b2f778450a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80c982f07874df686fe7e29cc91043eb6d58f1be5fc93cacd46d33a21dd9b622\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80c982f07874df686fe7e29cc91043eb6d58f1be5fc93cacd46d33a21dd9b622\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a354c985652c7ad7fa8944fbfdfe223cfb0f97fb11eecd9bd429020d4892b78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a354c985652c7ad7fa8944fbfdfe223cfb0f97fb11eecd9bd429020d4892b78\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2efd7283df313d086c1b8b88f289ace28764de256205bee7610d6711002d8515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2efd7283df313d086c1b8b88f289ace28764de256205bee7610d6711002d8515\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2ll9b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:41Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:41 crc kubenswrapper[4885]: I1002 01:47:41.854129 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:41 crc kubenswrapper[4885]: I1002 01:47:41.854196 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:41 crc kubenswrapper[4885]: I1002 01:47:41.854214 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:41 crc kubenswrapper[4885]: I1002 01:47:41.854238 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:41 crc kubenswrapper[4885]: I1002 01:47:41.854254 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:41Z","lastTransitionTime":"2025-10-02T01:47:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:41 crc kubenswrapper[4885]: I1002 01:47:41.956584 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:41 crc kubenswrapper[4885]: I1002 01:47:41.956643 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:41 crc kubenswrapper[4885]: I1002 01:47:41.956661 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:41 crc kubenswrapper[4885]: I1002 01:47:41.956684 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:41 crc kubenswrapper[4885]: I1002 01:47:41.956703 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:41Z","lastTransitionTime":"2025-10-02T01:47:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:42 crc kubenswrapper[4885]: I1002 01:47:42.046334 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sv4md" Oct 02 01:47:42 crc kubenswrapper[4885]: I1002 01:47:42.046415 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:47:42 crc kubenswrapper[4885]: I1002 01:47:42.046541 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:47:42 crc kubenswrapper[4885]: E1002 01:47:42.046711 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sv4md" podUID="724a3e49-001c-45a4-a896-82c13e0ee0e7" Oct 02 01:47:42 crc kubenswrapper[4885]: I1002 01:47:42.046801 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:47:42 crc kubenswrapper[4885]: E1002 01:47:42.046840 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:47:42 crc kubenswrapper[4885]: E1002 01:47:42.047068 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:47:42 crc kubenswrapper[4885]: E1002 01:47:42.047347 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:47:42 crc kubenswrapper[4885]: I1002 01:47:42.059679 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:42 crc kubenswrapper[4885]: I1002 01:47:42.059782 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:42 crc kubenswrapper[4885]: I1002 01:47:42.059837 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:42 crc kubenswrapper[4885]: I1002 01:47:42.059868 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:42 crc kubenswrapper[4885]: I1002 01:47:42.059924 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:42Z","lastTransitionTime":"2025-10-02T01:47:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:42 crc kubenswrapper[4885]: I1002 01:47:42.065192 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:42Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:42 crc kubenswrapper[4885]: I1002 01:47:42.083563 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:42Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:42 crc kubenswrapper[4885]: I1002 01:47:42.098686 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-58lhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac6a19c4-f259-4dbd-a016-d6c2c849c6cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed76c0f968c063f7795d363af194a222487b647ceda234de0dabb39b25eafe0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hzc4c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-58lhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:42Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:42 crc kubenswrapper[4885]: I1002 01:47:42.111689 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d169c8ba-000a-4218-8e3b-4ae53035b110\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61e436b7876107ce5ea9e668c0bbc367be6e663fc804fa3a65b51adf8be8d51f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-77rzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09536e182a46e77318c10812d1777e551f834907c006cab5c726dc66f5bce88a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-77rzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rttx8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:42Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:42 crc kubenswrapper[4885]: I1002 01:47:42.127056 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jgdvt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81733e8d-67b0-46fa-8a41-08b40645bd3e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcef60c816ca1ba9a298da658e84c4d717bea3d352c4c2f807ebd8e56327b384\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cwbl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jgdvt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:42Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:42 crc kubenswrapper[4885]: I1002 01:47:42.141193 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a0bb885-2eb0-419c-8a7a-ef634cb3c144\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a89b49cbe08d927629564e607de13655ddf3d01ac4eb5d688b325210306c120c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06c5942805a498d9621a906088aa54538b4eed583d1b8d0e3ba2a6718378a36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b9f6f7fe44d5771a838d0b89c5bdea3ebb79023ae2675ee907ec2766b3d35a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://178b4340ee6904c60c814b99d83ee225161aeb23e438d8d02f856579a13ec4e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bb63bf30d6ecce9e71f0db0c16632bdfa6b33924b673378eb0c38ab229fa837\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T01:47:06Z\\\",\\\"message\\\":\\\"W1002 01:46:55.406027 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1002 01:46:55.406714 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759369615 cert, and key in /tmp/serving-cert-3795040473/serving-signer.crt, /tmp/serving-cert-3795040473/serving-signer.key\\\\nI1002 01:46:55.614670 1 observer_polling.go:159] Starting file observer\\\\nW1002 01:46:55.617662 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1002 01:46:55.618112 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 01:46:55.620347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3795040473/tls.crt::/tmp/serving-cert-3795040473/tls.key\\\\\\\"\\\\nF1002 01:47:06.224459 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90535f97ab2ed4626d47540ebd7baada837fd332140323bec0f49aa38d3c0009\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://592f89a326c9c1559193a71904b47ca27b1dbe1a7b30e03bc48924bb2f35f921\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://592f89a326c9c1559193a71904b47ca27b1dbe1a7b30e03bc48924bb2f35f921\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:42Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:42 crc kubenswrapper[4885]: I1002 01:47:42.162691 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:42 crc kubenswrapper[4885]: I1002 01:47:42.162742 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:42 crc kubenswrapper[4885]: I1002 01:47:42.162757 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:42 crc kubenswrapper[4885]: I1002 01:47:42.162780 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:42 crc kubenswrapper[4885]: I1002 01:47:42.162797 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:42Z","lastTransitionTime":"2025-10-02T01:47:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:42 crc kubenswrapper[4885]: I1002 01:47:42.168990 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b087880e05039816dcab154643ac72c7e4a0a1124817c51dc2c639b62bbee0fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:42Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:42 crc kubenswrapper[4885]: I1002 01:47:42.193577 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"506a599f-fb1c-4d14-be88-0e4833f3d624\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31c2f0193ffef54965b79a3569b0f6f4547e540bac0774155159a1c996c196c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://533e73e5f90b7258acbcdf018ee01323c118f9dc45035b25b065ba01cabecccf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c00db833a593ec40eb63a6b383455e19e960c2e98b9b66f76877ee993057e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21e2f462a0deb0d740185663b2014f4a1e57b4fdbeb82571264b002f928035bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:42Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:42 crc kubenswrapper[4885]: I1002 01:47:42.223133 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e4679dd-f870-41e9-a32b-360f5d2ee81b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4310909ad9c7b609ed6814e04a525a6afae747775598e7275080413dec8b23a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7642033b0b5995d332ccf7992d6d7e6a5e9c9369ae2d1071ccb2f536938c8f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9f99f9e26a04fb42b01a0783a5c4e360514818ce9102edc7fa08626d50b9a34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6fa72b030d309656b73121e4ac2a3924a4e449aad66683dd9de5ee5f22d5a4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef403e519d197fb37b3e6b76a14a5128ba3f30a7a5e8bf0a420cf6a39b89575f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8482ba457683ffa03100ec1f07ea9d3f918be5e490b5ba05c2ba965d03033191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://188cb50b58926683ce8862d618258c1d1a6f33310608546c8651ece28f7c338f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f85438a1b57b0a590e4d005978fe70ff8c891980badb20d65e1f0e14080fae6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T01:47:25Z\\\",\\\"message\\\":\\\"dressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-api/machine-api-operator-machine-webhook_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-operator-machine-webhook\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.250\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1002 01:47:25.452713 6353 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1002 01:47:25.452745 6353 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-source-55646444c4-trplf] creating logical port openshift-network-diagnostics_network-check-source-55646444c4-trplf for pod on switch crc\\\\nF1002 01:47:25.452781 6353 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:24Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://188cb50b58926683ce8862d618258c1d1a6f33310608546c8651ece28f7c338f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T01:47:41Z\\\",\\\"message\\\":\\\"921 6578 obj_retry.go:434] periodicallyRetryResources: Retry channel got triggered: retrying failed objects of type *v1.Pod\\\\nI1002 01:47:40.991945 6578 obj_retry.go:409] Going to retry *v1.Pod resource setup for 16 objects: [openshift-multus/network-metrics-daemon-sv4md openshift-network-operator/iptables-alerter-4ln5h openshift-kube-apiserver/kube-apiserver-crc openshift-ovn-kubernetes/ovnkube-node-dngcm openshift-dns/node-resolver-58lhj openshift-machine-config-operator/machine-config-daemon-rttx8 openshift-multus/multus-vhgbd openshift-network-console/networking-console-plugin-85b44fc459-gdk6g openshift-network-diagnostics/network-check-source-55646444c4-trplf openshift-network-operator/network-operator-58b4c7f79c-55gtf openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rrn8b openshift-image-registry/node-ca-jgdvt openshift-kube-controller-manager/kube-controller-manager-crc openshift-multus/multus-additional-cni-plugins-2ll9b openshift-network-diagnostics/network-check-target-xd92c openshift-network-node-identity/network-node-identity-vrzqb]\\\\nF1002 01:47:40.991998 6578 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller ini\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fafdd1094bd02786d745c984989852b0c690db38655f92b419733a2004f417c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0125e907d061a97286026ee823c4eed303734ed03ab4864f742592b23a1aa2f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0125e907d061a97286026ee823c4eed303734ed03ab4864f742592b23a1aa2f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dngcm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:42Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:42 crc kubenswrapper[4885]: I1002 01:47:42.240327 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-sv4md" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"724a3e49-001c-45a4-a896-82c13e0ee0e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hpqgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hpqgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-sv4md\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:42Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:42 crc kubenswrapper[4885]: I1002 01:47:42.258569 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vhgbd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c63a7d20d39b65acfb5a9be4c28b21c72111b398bea76c238e15ddb4dd63e139\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7zwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vhgbd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:42Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:42 crc kubenswrapper[4885]: I1002 01:47:42.265845 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:42 crc kubenswrapper[4885]: I1002 01:47:42.265894 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:42 crc kubenswrapper[4885]: I1002 01:47:42.265913 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:42 crc kubenswrapper[4885]: I1002 01:47:42.265938 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:42 crc kubenswrapper[4885]: I1002 01:47:42.265959 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:42Z","lastTransitionTime":"2025-10-02T01:47:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:42 crc kubenswrapper[4885]: I1002 01:47:42.275530 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rrn8b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52967a77-525c-4537-9b1f-d7bb754a8494\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3dee451ebe9006bf1c519fed4c65267570b6139d76841858352e087ae78d26a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtr74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e9412788d6df7332592beac48281822b26e1aa19b39addce8e1be556af9d982\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtr74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:26Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rrn8b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:42Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:42 crc kubenswrapper[4885]: I1002 01:47:42.297157 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8070bc1c-e3de-4e9a-985e-40d1a9858056\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3e634b3727c517cc48b17c2cfb38f26460f53d1bf25d631d08ec9acb8d780e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4adcc537c23697a324d82af94c31059dd444245efdf7c978db58ede256e112ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edcd1d99644a778dfb6d0a17d2a4abd906400fe8209e611ebf8d7a9abe7ccf28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://97cb59adc2ddfa365ebc9dec649600dc90466ffe59134a6d15a9d3134a7c9b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d2808bc658ea80d68469ed16422d56ef19447f4c90b4af7b2597c64406f8a44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://83acbc20919f7c1958b6d492fadb203855df02761662425df9f91a0757fe2bf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83acbc20919f7c1958b6d492fadb203855df02761662425df9f91a0757fe2bf7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e86ecd03835efc9a253b6804bc61e6ac690e75b8d1f3cc77f5647730e586574\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e86ecd03835efc9a253b6804bc61e6ac690e75b8d1f3cc77f5647730e586574\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://48f95af51feaf014a18496bdfadd59d7f1e11167c5c8b677fa7037e48368ff2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48f95af51feaf014a18496bdfadd59d7f1e11167c5c8b677fa7037e48368ff2a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:42Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:42 crc kubenswrapper[4885]: I1002 01:47:42.313026 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5c1f95f188569c9aa47e51f2916939fa0b65a5cdf4c16b798612d41f3c8411a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:42Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:42 crc kubenswrapper[4885]: I1002 01:47:42.329704 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:42Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:42 crc kubenswrapper[4885]: I1002 01:47:42.347952 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74adcf912ccedb6b5d4f2803fd215e3e2ef62367597aa6da528354c4511ee536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7403ef968d040ac09d722b7557438bef23892d578992826f42b700b370d495d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:42Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:42 crc kubenswrapper[4885]: I1002 01:47:42.368826 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:42 crc kubenswrapper[4885]: I1002 01:47:42.368891 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:42 crc kubenswrapper[4885]: I1002 01:47:42.368910 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:42 crc kubenswrapper[4885]: I1002 01:47:42.368936 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:42 crc kubenswrapper[4885]: I1002 01:47:42.368955 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:42Z","lastTransitionTime":"2025-10-02T01:47:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:42 crc kubenswrapper[4885]: I1002 01:47:42.369256 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2ll9b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52b39712-5fa1-4fe0-814b-b170a6e3938b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9a96383d0fbfd116c419054aaf1a7adba1937e8b0bad294cefd68b4ab0d2298\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c979b91ba748aa4647826a8926337662ca929c54090fb62285572819ec2fbf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c979b91ba748aa4647826a8926337662ca929c54090fb62285572819ec2fbf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97d9d693e8e60ee9c183bb44461e698e8b3cfd20a516605aa8754c9205d2bdc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://97d9d693e8e60ee9c183bb44461e698e8b3cfd20a516605aa8754c9205d2bdc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c44252676359c671d65c9588b05572d76c99cf4c90666f0ac089b2f778450a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c44252676359c671d65c9588b05572d76c99cf4c90666f0ac089b2f778450a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80c982f07874df686fe7e29cc91043eb6d58f1be5fc93cacd46d33a21dd9b622\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80c982f07874df686fe7e29cc91043eb6d58f1be5fc93cacd46d33a21dd9b622\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a354c985652c7ad7fa8944fbfdfe223cfb0f97fb11eecd9bd429020d4892b78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a354c985652c7ad7fa8944fbfdfe223cfb0f97fb11eecd9bd429020d4892b78\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2efd7283df313d086c1b8b88f289ace28764de256205bee7610d6711002d8515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2efd7283df313d086c1b8b88f289ace28764de256205bee7610d6711002d8515\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2ll9b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:42Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:42 crc kubenswrapper[4885]: I1002 01:47:42.471914 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:42 crc kubenswrapper[4885]: I1002 01:47:42.471997 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:42 crc kubenswrapper[4885]: I1002 01:47:42.472014 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:42 crc kubenswrapper[4885]: I1002 01:47:42.472039 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:42 crc kubenswrapper[4885]: I1002 01:47:42.472060 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:42Z","lastTransitionTime":"2025-10-02T01:47:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:42 crc kubenswrapper[4885]: I1002 01:47:42.511081 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dngcm_9e4679dd-f870-41e9-a32b-360f5d2ee81b/ovnkube-controller/2.log" Oct 02 01:47:42 crc kubenswrapper[4885]: I1002 01:47:42.574615 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:42 crc kubenswrapper[4885]: I1002 01:47:42.574686 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:42 crc kubenswrapper[4885]: I1002 01:47:42.574711 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:42 crc kubenswrapper[4885]: I1002 01:47:42.574742 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:42 crc kubenswrapper[4885]: I1002 01:47:42.574765 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:42Z","lastTransitionTime":"2025-10-02T01:47:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:42 crc kubenswrapper[4885]: I1002 01:47:42.677855 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:42 crc kubenswrapper[4885]: I1002 01:47:42.677912 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:42 crc kubenswrapper[4885]: I1002 01:47:42.677938 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:42 crc kubenswrapper[4885]: I1002 01:47:42.677965 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:42 crc kubenswrapper[4885]: I1002 01:47:42.677985 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:42Z","lastTransitionTime":"2025-10-02T01:47:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:42 crc kubenswrapper[4885]: I1002 01:47:42.780474 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:42 crc kubenswrapper[4885]: I1002 01:47:42.780539 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:42 crc kubenswrapper[4885]: I1002 01:47:42.780558 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:42 crc kubenswrapper[4885]: I1002 01:47:42.780580 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:42 crc kubenswrapper[4885]: I1002 01:47:42.780596 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:42Z","lastTransitionTime":"2025-10-02T01:47:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:42 crc kubenswrapper[4885]: I1002 01:47:42.883319 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:42 crc kubenswrapper[4885]: I1002 01:47:42.883386 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:42 crc kubenswrapper[4885]: I1002 01:47:42.883404 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:42 crc kubenswrapper[4885]: I1002 01:47:42.883427 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:42 crc kubenswrapper[4885]: I1002 01:47:42.883447 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:42Z","lastTransitionTime":"2025-10-02T01:47:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:42 crc kubenswrapper[4885]: I1002 01:47:42.986785 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:42 crc kubenswrapper[4885]: I1002 01:47:42.986844 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:42 crc kubenswrapper[4885]: I1002 01:47:42.986866 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:42 crc kubenswrapper[4885]: I1002 01:47:42.986894 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:42 crc kubenswrapper[4885]: I1002 01:47:42.986917 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:42Z","lastTransitionTime":"2025-10-02T01:47:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:43 crc kubenswrapper[4885]: I1002 01:47:43.076181 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" Oct 02 01:47:43 crc kubenswrapper[4885]: I1002 01:47:43.077482 4885 scope.go:117] "RemoveContainer" containerID="188cb50b58926683ce8862d618258c1d1a6f33310608546c8651ece28f7c338f" Oct 02 01:47:43 crc kubenswrapper[4885]: E1002 01:47:43.077747 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-dngcm_openshift-ovn-kubernetes(9e4679dd-f870-41e9-a32b-360f5d2ee81b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" podUID="9e4679dd-f870-41e9-a32b-360f5d2ee81b" Oct 02 01:47:43 crc kubenswrapper[4885]: I1002 01:47:43.089729 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:43 crc kubenswrapper[4885]: I1002 01:47:43.089812 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:43 crc kubenswrapper[4885]: I1002 01:47:43.089834 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:43 crc kubenswrapper[4885]: I1002 01:47:43.089867 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:43 crc kubenswrapper[4885]: I1002 01:47:43.089891 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:43Z","lastTransitionTime":"2025-10-02T01:47:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:43 crc kubenswrapper[4885]: I1002 01:47:43.098428 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"506a599f-fb1c-4d14-be88-0e4833f3d624\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31c2f0193ffef54965b79a3569b0f6f4547e540bac0774155159a1c996c196c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://533e73e5f90b7258acbcdf018ee01323c118f9dc45035b25b065ba01cabecccf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c00db833a593ec40eb63a6b383455e19e960c2e98b9b66f76877ee993057e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21e2f462a0deb0d740185663b2014f4a1e57b4fdbeb82571264b002f928035bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:43Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:43 crc kubenswrapper[4885]: I1002 01:47:43.128821 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e4679dd-f870-41e9-a32b-360f5d2ee81b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4310909ad9c7b609ed6814e04a525a6afae747775598e7275080413dec8b23a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7642033b0b5995d332ccf7992d6d7e6a5e9c9369ae2d1071ccb2f536938c8f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9f99f9e26a04fb42b01a0783a5c4e360514818ce9102edc7fa08626d50b9a34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6fa72b030d309656b73121e4ac2a3924a4e449aad66683dd9de5ee5f22d5a4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef403e519d197fb37b3e6b76a14a5128ba3f30a7a5e8bf0a420cf6a39b89575f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8482ba457683ffa03100ec1f07ea9d3f918be5e490b5ba05c2ba965d03033191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://188cb50b58926683ce8862d618258c1d1a6f33310608546c8651ece28f7c338f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://188cb50b58926683ce8862d618258c1d1a6f33310608546c8651ece28f7c338f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T01:47:41Z\\\",\\\"message\\\":\\\"921 6578 obj_retry.go:434] periodicallyRetryResources: Retry channel got triggered: retrying failed objects of type *v1.Pod\\\\nI1002 01:47:40.991945 6578 obj_retry.go:409] Going to retry *v1.Pod resource setup for 16 objects: [openshift-multus/network-metrics-daemon-sv4md openshift-network-operator/iptables-alerter-4ln5h openshift-kube-apiserver/kube-apiserver-crc openshift-ovn-kubernetes/ovnkube-node-dngcm openshift-dns/node-resolver-58lhj openshift-machine-config-operator/machine-config-daemon-rttx8 openshift-multus/multus-vhgbd openshift-network-console/networking-console-plugin-85b44fc459-gdk6g openshift-network-diagnostics/network-check-source-55646444c4-trplf openshift-network-operator/network-operator-58b4c7f79c-55gtf openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rrn8b openshift-image-registry/node-ca-jgdvt openshift-kube-controller-manager/kube-controller-manager-crc openshift-multus/multus-additional-cni-plugins-2ll9b openshift-network-diagnostics/network-check-target-xd92c openshift-network-node-identity/network-node-identity-vrzqb]\\\\nF1002 01:47:40.991998 6578 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller ini\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:40Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-dngcm_openshift-ovn-kubernetes(9e4679dd-f870-41e9-a32b-360f5d2ee81b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fafdd1094bd02786d745c984989852b0c690db38655f92b419733a2004f417c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0125e907d061a97286026ee823c4eed303734ed03ab4864f742592b23a1aa2f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0125e907d061a97286026ee823c4eed303734ed03ab4864f742592b23a1aa2f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dngcm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:43Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:43 crc kubenswrapper[4885]: I1002 01:47:43.146725 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-sv4md" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"724a3e49-001c-45a4-a896-82c13e0ee0e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hpqgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hpqgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-sv4md\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:43Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:43 crc kubenswrapper[4885]: I1002 01:47:43.165830 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vhgbd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c63a7d20d39b65acfb5a9be4c28b21c72111b398bea76c238e15ddb4dd63e139\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7zwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vhgbd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:43Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:43 crc kubenswrapper[4885]: I1002 01:47:43.182907 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rrn8b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52967a77-525c-4537-9b1f-d7bb754a8494\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3dee451ebe9006bf1c519fed4c65267570b6139d76841858352e087ae78d26a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtr74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e9412788d6df7332592beac48281822b26e1aa19b39addce8e1be556af9d982\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtr74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:26Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rrn8b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:43Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:43 crc kubenswrapper[4885]: I1002 01:47:43.194874 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:43 crc kubenswrapper[4885]: I1002 01:47:43.194958 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:43 crc kubenswrapper[4885]: I1002 01:47:43.194985 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:43 crc kubenswrapper[4885]: I1002 01:47:43.195014 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:43 crc kubenswrapper[4885]: I1002 01:47:43.195036 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:43Z","lastTransitionTime":"2025-10-02T01:47:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:43 crc kubenswrapper[4885]: I1002 01:47:43.219163 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8070bc1c-e3de-4e9a-985e-40d1a9858056\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3e634b3727c517cc48b17c2cfb38f26460f53d1bf25d631d08ec9acb8d780e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4adcc537c23697a324d82af94c31059dd444245efdf7c978db58ede256e112ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edcd1d99644a778dfb6d0a17d2a4abd906400fe8209e611ebf8d7a9abe7ccf28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://97cb59adc2ddfa365ebc9dec649600dc90466ffe59134a6d15a9d3134a7c9b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d2808bc658ea80d68469ed16422d56ef19447f4c90b4af7b2597c64406f8a44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://83acbc20919f7c1958b6d492fadb203855df02761662425df9f91a0757fe2bf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83acbc20919f7c1958b6d492fadb203855df02761662425df9f91a0757fe2bf7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e86ecd03835efc9a253b6804bc61e6ac690e75b8d1f3cc77f5647730e586574\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e86ecd03835efc9a253b6804bc61e6ac690e75b8d1f3cc77f5647730e586574\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://48f95af51feaf014a18496bdfadd59d7f1e11167c5c8b677fa7037e48368ff2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48f95af51feaf014a18496bdfadd59d7f1e11167c5c8b677fa7037e48368ff2a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:43Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:43 crc kubenswrapper[4885]: I1002 01:47:43.238117 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5c1f95f188569c9aa47e51f2916939fa0b65a5cdf4c16b798612d41f3c8411a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:43Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:43 crc kubenswrapper[4885]: I1002 01:47:43.257770 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:43Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:43 crc kubenswrapper[4885]: I1002 01:47:43.280231 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74adcf912ccedb6b5d4f2803fd215e3e2ef62367597aa6da528354c4511ee536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7403ef968d040ac09d722b7557438bef23892d578992826f42b700b370d495d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:43Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:43 crc kubenswrapper[4885]: I1002 01:47:43.299334 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:43 crc kubenswrapper[4885]: I1002 01:47:43.299396 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:43 crc kubenswrapper[4885]: I1002 01:47:43.299419 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:43 crc kubenswrapper[4885]: I1002 01:47:43.299451 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:43 crc kubenswrapper[4885]: I1002 01:47:43.299479 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:43Z","lastTransitionTime":"2025-10-02T01:47:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:43 crc kubenswrapper[4885]: I1002 01:47:43.305866 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2ll9b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52b39712-5fa1-4fe0-814b-b170a6e3938b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9a96383d0fbfd116c419054aaf1a7adba1937e8b0bad294cefd68b4ab0d2298\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c979b91ba748aa4647826a8926337662ca929c54090fb62285572819ec2fbf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c979b91ba748aa4647826a8926337662ca929c54090fb62285572819ec2fbf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97d9d693e8e60ee9c183bb44461e698e8b3cfd20a516605aa8754c9205d2bdc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://97d9d693e8e60ee9c183bb44461e698e8b3cfd20a516605aa8754c9205d2bdc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c44252676359c671d65c9588b05572d76c99cf4c90666f0ac089b2f778450a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c44252676359c671d65c9588b05572d76c99cf4c90666f0ac089b2f778450a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80c982f07874df686fe7e29cc91043eb6d58f1be5fc93cacd46d33a21dd9b622\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80c982f07874df686fe7e29cc91043eb6d58f1be5fc93cacd46d33a21dd9b622\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a354c985652c7ad7fa8944fbfdfe223cfb0f97fb11eecd9bd429020d4892b78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a354c985652c7ad7fa8944fbfdfe223cfb0f97fb11eecd9bd429020d4892b78\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2efd7283df313d086c1b8b88f289ace28764de256205bee7610d6711002d8515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2efd7283df313d086c1b8b88f289ace28764de256205bee7610d6711002d8515\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2ll9b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:43Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:43 crc kubenswrapper[4885]: I1002 01:47:43.326124 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:43Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:43 crc kubenswrapper[4885]: I1002 01:47:43.349378 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:43Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:43 crc kubenswrapper[4885]: I1002 01:47:43.365505 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-58lhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac6a19c4-f259-4dbd-a016-d6c2c849c6cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed76c0f968c063f7795d363af194a222487b647ceda234de0dabb39b25eafe0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hzc4c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-58lhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:43Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:43 crc kubenswrapper[4885]: I1002 01:47:43.381914 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d169c8ba-000a-4218-8e3b-4ae53035b110\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61e436b7876107ce5ea9e668c0bbc367be6e663fc804fa3a65b51adf8be8d51f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-77rzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09536e182a46e77318c10812d1777e551f834907c006cab5c726dc66f5bce88a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-77rzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rttx8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:43Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:43 crc kubenswrapper[4885]: I1002 01:47:43.395852 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jgdvt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81733e8d-67b0-46fa-8a41-08b40645bd3e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcef60c816ca1ba9a298da658e84c4d717bea3d352c4c2f807ebd8e56327b384\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cwbl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jgdvt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:43Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:43 crc kubenswrapper[4885]: I1002 01:47:43.401997 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:43 crc kubenswrapper[4885]: I1002 01:47:43.402031 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:43 crc kubenswrapper[4885]: I1002 01:47:43.402045 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:43 crc kubenswrapper[4885]: I1002 01:47:43.402063 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:43 crc kubenswrapper[4885]: I1002 01:47:43.402077 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:43Z","lastTransitionTime":"2025-10-02T01:47:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:43 crc kubenswrapper[4885]: I1002 01:47:43.414417 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a0bb885-2eb0-419c-8a7a-ef634cb3c144\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a89b49cbe08d927629564e607de13655ddf3d01ac4eb5d688b325210306c120c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06c5942805a498d9621a906088aa54538b4eed583d1b8d0e3ba2a6718378a36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b9f6f7fe44d5771a838d0b89c5bdea3ebb79023ae2675ee907ec2766b3d35a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://178b4340ee6904c60c814b99d83ee225161aeb23e438d8d02f856579a13ec4e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bb63bf30d6ecce9e71f0db0c16632bdfa6b33924b673378eb0c38ab229fa837\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T01:47:06Z\\\",\\\"message\\\":\\\"W1002 01:46:55.406027 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1002 01:46:55.406714 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759369615 cert, and key in /tmp/serving-cert-3795040473/serving-signer.crt, /tmp/serving-cert-3795040473/serving-signer.key\\\\nI1002 01:46:55.614670 1 observer_polling.go:159] Starting file observer\\\\nW1002 01:46:55.617662 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1002 01:46:55.618112 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 01:46:55.620347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3795040473/tls.crt::/tmp/serving-cert-3795040473/tls.key\\\\\\\"\\\\nF1002 01:47:06.224459 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90535f97ab2ed4626d47540ebd7baada837fd332140323bec0f49aa38d3c0009\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://592f89a326c9c1559193a71904b47ca27b1dbe1a7b30e03bc48924bb2f35f921\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://592f89a326c9c1559193a71904b47ca27b1dbe1a7b30e03bc48924bb2f35f921\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:43Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:43 crc kubenswrapper[4885]: I1002 01:47:43.433897 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b087880e05039816dcab154643ac72c7e4a0a1124817c51dc2c639b62bbee0fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:43Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:43 crc kubenswrapper[4885]: I1002 01:47:43.444789 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 02 01:47:43 crc kubenswrapper[4885]: I1002 01:47:43.458519 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Oct 02 01:47:43 crc kubenswrapper[4885]: I1002 01:47:43.467066 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8070bc1c-e3de-4e9a-985e-40d1a9858056\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3e634b3727c517cc48b17c2cfb38f26460f53d1bf25d631d08ec9acb8d780e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4adcc537c23697a324d82af94c31059dd444245efdf7c978db58ede256e112ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edcd1d99644a778dfb6d0a17d2a4abd906400fe8209e611ebf8d7a9abe7ccf28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://97cb59adc2ddfa365ebc9dec649600dc90466ffe59134a6d15a9d3134a7c9b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d2808bc658ea80d68469ed16422d56ef19447f4c90b4af7b2597c64406f8a44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://83acbc20919f7c1958b6d492fadb203855df02761662425df9f91a0757fe2bf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83acbc20919f7c1958b6d492fadb203855df02761662425df9f91a0757fe2bf7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e86ecd03835efc9a253b6804bc61e6ac690e75b8d1f3cc77f5647730e586574\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e86ecd03835efc9a253b6804bc61e6ac690e75b8d1f3cc77f5647730e586574\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://48f95af51feaf014a18496bdfadd59d7f1e11167c5c8b677fa7037e48368ff2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48f95af51feaf014a18496bdfadd59d7f1e11167c5c8b677fa7037e48368ff2a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:43Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:43 crc kubenswrapper[4885]: I1002 01:47:43.484989 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5c1f95f188569c9aa47e51f2916939fa0b65a5cdf4c16b798612d41f3c8411a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:43Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:43 crc kubenswrapper[4885]: I1002 01:47:43.503633 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vhgbd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c63a7d20d39b65acfb5a9be4c28b21c72111b398bea76c238e15ddb4dd63e139\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7zwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vhgbd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:43Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:43 crc kubenswrapper[4885]: I1002 01:47:43.504789 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:43 crc kubenswrapper[4885]: I1002 01:47:43.504859 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:43 crc kubenswrapper[4885]: I1002 01:47:43.504885 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:43 crc kubenswrapper[4885]: I1002 01:47:43.504910 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:43 crc kubenswrapper[4885]: I1002 01:47:43.504928 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:43Z","lastTransitionTime":"2025-10-02T01:47:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:43 crc kubenswrapper[4885]: I1002 01:47:43.517949 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rrn8b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52967a77-525c-4537-9b1f-d7bb754a8494\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3dee451ebe9006bf1c519fed4c65267570b6139d76841858352e087ae78d26a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtr74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e9412788d6df7332592beac48281822b26e1aa19b39addce8e1be556af9d982\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtr74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:26Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rrn8b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:43Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:43 crc kubenswrapper[4885]: I1002 01:47:43.537181 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:43Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:43 crc kubenswrapper[4885]: I1002 01:47:43.558630 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74adcf912ccedb6b5d4f2803fd215e3e2ef62367597aa6da528354c4511ee536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7403ef968d040ac09d722b7557438bef23892d578992826f42b700b370d495d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:43Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:43 crc kubenswrapper[4885]: I1002 01:47:43.581934 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2ll9b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52b39712-5fa1-4fe0-814b-b170a6e3938b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9a96383d0fbfd116c419054aaf1a7adba1937e8b0bad294cefd68b4ab0d2298\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c979b91ba748aa4647826a8926337662ca929c54090fb62285572819ec2fbf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c979b91ba748aa4647826a8926337662ca929c54090fb62285572819ec2fbf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97d9d693e8e60ee9c183bb44461e698e8b3cfd20a516605aa8754c9205d2bdc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://97d9d693e8e60ee9c183bb44461e698e8b3cfd20a516605aa8754c9205d2bdc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c44252676359c671d65c9588b05572d76c99cf4c90666f0ac089b2f778450a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c44252676359c671d65c9588b05572d76c99cf4c90666f0ac089b2f778450a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80c982f07874df686fe7e29cc91043eb6d58f1be5fc93cacd46d33a21dd9b622\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80c982f07874df686fe7e29cc91043eb6d58f1be5fc93cacd46d33a21dd9b622\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a354c985652c7ad7fa8944fbfdfe223cfb0f97fb11eecd9bd429020d4892b78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a354c985652c7ad7fa8944fbfdfe223cfb0f97fb11eecd9bd429020d4892b78\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2efd7283df313d086c1b8b88f289ace28764de256205bee7610d6711002d8515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2efd7283df313d086c1b8b88f289ace28764de256205bee7610d6711002d8515\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2ll9b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:43Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:43 crc kubenswrapper[4885]: I1002 01:47:43.599900 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d169c8ba-000a-4218-8e3b-4ae53035b110\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61e436b7876107ce5ea9e668c0bbc367be6e663fc804fa3a65b51adf8be8d51f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-77rzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09536e182a46e77318c10812d1777e551f834907c006cab5c726dc66f5bce88a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-77rzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rttx8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:43Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:43 crc kubenswrapper[4885]: I1002 01:47:43.607747 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:43 crc kubenswrapper[4885]: I1002 01:47:43.607817 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:43 crc kubenswrapper[4885]: I1002 01:47:43.607843 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:43 crc kubenswrapper[4885]: I1002 01:47:43.607873 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:43 crc kubenswrapper[4885]: I1002 01:47:43.607898 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:43Z","lastTransitionTime":"2025-10-02T01:47:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:43 crc kubenswrapper[4885]: I1002 01:47:43.615217 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jgdvt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81733e8d-67b0-46fa-8a41-08b40645bd3e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcef60c816ca1ba9a298da658e84c4d717bea3d352c4c2f807ebd8e56327b384\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cwbl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jgdvt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:43Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:43 crc kubenswrapper[4885]: I1002 01:47:43.635162 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a0bb885-2eb0-419c-8a7a-ef634cb3c144\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a89b49cbe08d927629564e607de13655ddf3d01ac4eb5d688b325210306c120c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06c5942805a498d9621a906088aa54538b4eed583d1b8d0e3ba2a6718378a36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b9f6f7fe44d5771a838d0b89c5bdea3ebb79023ae2675ee907ec2766b3d35a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://178b4340ee6904c60c814b99d83ee225161aeb23e438d8d02f856579a13ec4e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bb63bf30d6ecce9e71f0db0c16632bdfa6b33924b673378eb0c38ab229fa837\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T01:47:06Z\\\",\\\"message\\\":\\\"W1002 01:46:55.406027 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1002 01:46:55.406714 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759369615 cert, and key in /tmp/serving-cert-3795040473/serving-signer.crt, /tmp/serving-cert-3795040473/serving-signer.key\\\\nI1002 01:46:55.614670 1 observer_polling.go:159] Starting file observer\\\\nW1002 01:46:55.617662 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1002 01:46:55.618112 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 01:46:55.620347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3795040473/tls.crt::/tmp/serving-cert-3795040473/tls.key\\\\\\\"\\\\nF1002 01:47:06.224459 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90535f97ab2ed4626d47540ebd7baada837fd332140323bec0f49aa38d3c0009\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://592f89a326c9c1559193a71904b47ca27b1dbe1a7b30e03bc48924bb2f35f921\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://592f89a326c9c1559193a71904b47ca27b1dbe1a7b30e03bc48924bb2f35f921\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:43Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:43 crc kubenswrapper[4885]: I1002 01:47:43.657348 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b087880e05039816dcab154643ac72c7e4a0a1124817c51dc2c639b62bbee0fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:43Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:43 crc kubenswrapper[4885]: I1002 01:47:43.676509 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:43Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:43 crc kubenswrapper[4885]: I1002 01:47:43.694297 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:43Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:43 crc kubenswrapper[4885]: I1002 01:47:43.710586 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-58lhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac6a19c4-f259-4dbd-a016-d6c2c849c6cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed76c0f968c063f7795d363af194a222487b647ceda234de0dabb39b25eafe0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hzc4c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-58lhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:43Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:43 crc kubenswrapper[4885]: I1002 01:47:43.711485 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:43 crc kubenswrapper[4885]: I1002 01:47:43.711560 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:43 crc kubenswrapper[4885]: I1002 01:47:43.711588 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:43 crc kubenswrapper[4885]: I1002 01:47:43.711620 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:43 crc kubenswrapper[4885]: I1002 01:47:43.711642 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:43Z","lastTransitionTime":"2025-10-02T01:47:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:43 crc kubenswrapper[4885]: I1002 01:47:43.732558 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"506a599f-fb1c-4d14-be88-0e4833f3d624\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31c2f0193ffef54965b79a3569b0f6f4547e540bac0774155159a1c996c196c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://533e73e5f90b7258acbcdf018ee01323c118f9dc45035b25b065ba01cabecccf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c00db833a593ec40eb63a6b383455e19e960c2e98b9b66f76877ee993057e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21e2f462a0deb0d740185663b2014f4a1e57b4fdbeb82571264b002f928035bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:43Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:43 crc kubenswrapper[4885]: I1002 01:47:43.765142 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e4679dd-f870-41e9-a32b-360f5d2ee81b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4310909ad9c7b609ed6814e04a525a6afae747775598e7275080413dec8b23a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7642033b0b5995d332ccf7992d6d7e6a5e9c9369ae2d1071ccb2f536938c8f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9f99f9e26a04fb42b01a0783a5c4e360514818ce9102edc7fa08626d50b9a34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6fa72b030d309656b73121e4ac2a3924a4e449aad66683dd9de5ee5f22d5a4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef403e519d197fb37b3e6b76a14a5128ba3f30a7a5e8bf0a420cf6a39b89575f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8482ba457683ffa03100ec1f07ea9d3f918be5e490b5ba05c2ba965d03033191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://188cb50b58926683ce8862d618258c1d1a6f33310608546c8651ece28f7c338f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://188cb50b58926683ce8862d618258c1d1a6f33310608546c8651ece28f7c338f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T01:47:41Z\\\",\\\"message\\\":\\\"921 6578 obj_retry.go:434] periodicallyRetryResources: Retry channel got triggered: retrying failed objects of type *v1.Pod\\\\nI1002 01:47:40.991945 6578 obj_retry.go:409] Going to retry *v1.Pod resource setup for 16 objects: [openshift-multus/network-metrics-daemon-sv4md openshift-network-operator/iptables-alerter-4ln5h openshift-kube-apiserver/kube-apiserver-crc openshift-ovn-kubernetes/ovnkube-node-dngcm openshift-dns/node-resolver-58lhj openshift-machine-config-operator/machine-config-daemon-rttx8 openshift-multus/multus-vhgbd openshift-network-console/networking-console-plugin-85b44fc459-gdk6g openshift-network-diagnostics/network-check-source-55646444c4-trplf openshift-network-operator/network-operator-58b4c7f79c-55gtf openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rrn8b openshift-image-registry/node-ca-jgdvt openshift-kube-controller-manager/kube-controller-manager-crc openshift-multus/multus-additional-cni-plugins-2ll9b openshift-network-diagnostics/network-check-target-xd92c openshift-network-node-identity/network-node-identity-vrzqb]\\\\nF1002 01:47:40.991998 6578 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller ini\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:40Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-dngcm_openshift-ovn-kubernetes(9e4679dd-f870-41e9-a32b-360f5d2ee81b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fafdd1094bd02786d745c984989852b0c690db38655f92b419733a2004f417c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0125e907d061a97286026ee823c4eed303734ed03ab4864f742592b23a1aa2f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0125e907d061a97286026ee823c4eed303734ed03ab4864f742592b23a1aa2f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dngcm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:43Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:43 crc kubenswrapper[4885]: I1002 01:47:43.780488 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-sv4md" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"724a3e49-001c-45a4-a896-82c13e0ee0e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hpqgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hpqgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-sv4md\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:43Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:43 crc kubenswrapper[4885]: I1002 01:47:43.814511 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:43 crc kubenswrapper[4885]: I1002 01:47:43.814587 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:43 crc kubenswrapper[4885]: I1002 01:47:43.814605 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:43 crc kubenswrapper[4885]: I1002 01:47:43.814633 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:43 crc kubenswrapper[4885]: I1002 01:47:43.814651 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:43Z","lastTransitionTime":"2025-10-02T01:47:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:43 crc kubenswrapper[4885]: I1002 01:47:43.877015 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:47:43 crc kubenswrapper[4885]: I1002 01:47:43.877198 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:47:43 crc kubenswrapper[4885]: E1002 01:47:43.877289 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:48:15.877221133 +0000 UTC m=+84.688968572 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:47:43 crc kubenswrapper[4885]: E1002 01:47:43.877363 4885 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 01:47:43 crc kubenswrapper[4885]: E1002 01:47:43.877452 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 01:48:15.87743222 +0000 UTC m=+84.689179649 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 01:47:43 crc kubenswrapper[4885]: I1002 01:47:43.922566 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:43 crc kubenswrapper[4885]: I1002 01:47:43.922619 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:43 crc kubenswrapper[4885]: I1002 01:47:43.922651 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:43 crc kubenswrapper[4885]: I1002 01:47:43.922676 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:43 crc kubenswrapper[4885]: I1002 01:47:43.922693 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:43Z","lastTransitionTime":"2025-10-02T01:47:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:43 crc kubenswrapper[4885]: I1002 01:47:43.978011 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:47:43 crc kubenswrapper[4885]: I1002 01:47:43.978144 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:47:43 crc kubenswrapper[4885]: I1002 01:47:43.978219 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:47:43 crc kubenswrapper[4885]: E1002 01:47:43.978244 4885 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 01:47:43 crc kubenswrapper[4885]: E1002 01:47:43.978312 4885 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 01:47:43 crc kubenswrapper[4885]: E1002 01:47:43.978335 4885 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 01:47:43 crc kubenswrapper[4885]: E1002 01:47:43.978406 4885 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 01:47:43 crc kubenswrapper[4885]: E1002 01:47:43.978439 4885 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 01:47:43 crc kubenswrapper[4885]: E1002 01:47:43.978418 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-02 01:48:15.978392371 +0000 UTC m=+84.790139800 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 01:47:43 crc kubenswrapper[4885]: E1002 01:47:43.978486 4885 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 01:47:43 crc kubenswrapper[4885]: E1002 01:47:43.978504 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 01:48:15.978482554 +0000 UTC m=+84.790229983 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 01:47:43 crc kubenswrapper[4885]: E1002 01:47:43.978515 4885 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 01:47:43 crc kubenswrapper[4885]: E1002 01:47:43.978635 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-02 01:48:15.978597438 +0000 UTC m=+84.790344947 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 01:47:44 crc kubenswrapper[4885]: I1002 01:47:44.026019 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:44 crc kubenswrapper[4885]: I1002 01:47:44.026075 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:44 crc kubenswrapper[4885]: I1002 01:47:44.026091 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:44 crc kubenswrapper[4885]: I1002 01:47:44.026114 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:44 crc kubenswrapper[4885]: I1002 01:47:44.026133 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:44Z","lastTransitionTime":"2025-10-02T01:47:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:44 crc kubenswrapper[4885]: I1002 01:47:44.046595 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:47:44 crc kubenswrapper[4885]: E1002 01:47:44.046802 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:47:44 crc kubenswrapper[4885]: I1002 01:47:44.046886 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:47:44 crc kubenswrapper[4885]: I1002 01:47:44.046963 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:47:44 crc kubenswrapper[4885]: I1002 01:47:44.046981 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sv4md" Oct 02 01:47:44 crc kubenswrapper[4885]: E1002 01:47:44.047117 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:47:44 crc kubenswrapper[4885]: E1002 01:47:44.047233 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:47:44 crc kubenswrapper[4885]: E1002 01:47:44.047378 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sv4md" podUID="724a3e49-001c-45a4-a896-82c13e0ee0e7" Oct 02 01:47:44 crc kubenswrapper[4885]: I1002 01:47:44.078747 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/724a3e49-001c-45a4-a896-82c13e0ee0e7-metrics-certs\") pod \"network-metrics-daemon-sv4md\" (UID: \"724a3e49-001c-45a4-a896-82c13e0ee0e7\") " pod="openshift-multus/network-metrics-daemon-sv4md" Oct 02 01:47:44 crc kubenswrapper[4885]: E1002 01:47:44.078974 4885 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 01:47:44 crc kubenswrapper[4885]: E1002 01:47:44.079067 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/724a3e49-001c-45a4-a896-82c13e0ee0e7-metrics-certs podName:724a3e49-001c-45a4-a896-82c13e0ee0e7 nodeName:}" failed. No retries permitted until 2025-10-02 01:48:00.079041203 +0000 UTC m=+68.890788642 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/724a3e49-001c-45a4-a896-82c13e0ee0e7-metrics-certs") pod "network-metrics-daemon-sv4md" (UID: "724a3e49-001c-45a4-a896-82c13e0ee0e7") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 01:47:44 crc kubenswrapper[4885]: I1002 01:47:44.129247 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:44 crc kubenswrapper[4885]: I1002 01:47:44.129344 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:44 crc kubenswrapper[4885]: I1002 01:47:44.129362 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:44 crc kubenswrapper[4885]: I1002 01:47:44.129387 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:44 crc kubenswrapper[4885]: I1002 01:47:44.129405 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:44Z","lastTransitionTime":"2025-10-02T01:47:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:44 crc kubenswrapper[4885]: I1002 01:47:44.232031 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:44 crc kubenswrapper[4885]: I1002 01:47:44.232080 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:44 crc kubenswrapper[4885]: I1002 01:47:44.232096 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:44 crc kubenswrapper[4885]: I1002 01:47:44.232116 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:44 crc kubenswrapper[4885]: I1002 01:47:44.232132 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:44Z","lastTransitionTime":"2025-10-02T01:47:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:44 crc kubenswrapper[4885]: I1002 01:47:44.334501 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:44 crc kubenswrapper[4885]: I1002 01:47:44.334575 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:44 crc kubenswrapper[4885]: I1002 01:47:44.334595 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:44 crc kubenswrapper[4885]: I1002 01:47:44.334623 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:44 crc kubenswrapper[4885]: I1002 01:47:44.334642 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:44Z","lastTransitionTime":"2025-10-02T01:47:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:44 crc kubenswrapper[4885]: I1002 01:47:44.437705 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:44 crc kubenswrapper[4885]: I1002 01:47:44.437756 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:44 crc kubenswrapper[4885]: I1002 01:47:44.437772 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:44 crc kubenswrapper[4885]: I1002 01:47:44.437797 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:44 crc kubenswrapper[4885]: I1002 01:47:44.437813 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:44Z","lastTransitionTime":"2025-10-02T01:47:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:44 crc kubenswrapper[4885]: I1002 01:47:44.540594 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:44 crc kubenswrapper[4885]: I1002 01:47:44.540654 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:44 crc kubenswrapper[4885]: I1002 01:47:44.540706 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:44 crc kubenswrapper[4885]: I1002 01:47:44.540731 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:44 crc kubenswrapper[4885]: I1002 01:47:44.540749 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:44Z","lastTransitionTime":"2025-10-02T01:47:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:44 crc kubenswrapper[4885]: I1002 01:47:44.643550 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:44 crc kubenswrapper[4885]: I1002 01:47:44.643612 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:44 crc kubenswrapper[4885]: I1002 01:47:44.643631 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:44 crc kubenswrapper[4885]: I1002 01:47:44.643654 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:44 crc kubenswrapper[4885]: I1002 01:47:44.643669 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:44Z","lastTransitionTime":"2025-10-02T01:47:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:44 crc kubenswrapper[4885]: I1002 01:47:44.746588 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:44 crc kubenswrapper[4885]: I1002 01:47:44.746647 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:44 crc kubenswrapper[4885]: I1002 01:47:44.746664 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:44 crc kubenswrapper[4885]: I1002 01:47:44.746686 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:44 crc kubenswrapper[4885]: I1002 01:47:44.746703 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:44Z","lastTransitionTime":"2025-10-02T01:47:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:44 crc kubenswrapper[4885]: I1002 01:47:44.849931 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:44 crc kubenswrapper[4885]: I1002 01:47:44.849983 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:44 crc kubenswrapper[4885]: I1002 01:47:44.850003 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:44 crc kubenswrapper[4885]: I1002 01:47:44.850028 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:44 crc kubenswrapper[4885]: I1002 01:47:44.850045 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:44Z","lastTransitionTime":"2025-10-02T01:47:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:44 crc kubenswrapper[4885]: I1002 01:47:44.953231 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:44 crc kubenswrapper[4885]: I1002 01:47:44.953317 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:44 crc kubenswrapper[4885]: I1002 01:47:44.953337 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:44 crc kubenswrapper[4885]: I1002 01:47:44.953362 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:44 crc kubenswrapper[4885]: I1002 01:47:44.953381 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:44Z","lastTransitionTime":"2025-10-02T01:47:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:45 crc kubenswrapper[4885]: I1002 01:47:45.056620 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:45 crc kubenswrapper[4885]: I1002 01:47:45.056681 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:45 crc kubenswrapper[4885]: I1002 01:47:45.056699 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:45 crc kubenswrapper[4885]: I1002 01:47:45.056723 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:45 crc kubenswrapper[4885]: I1002 01:47:45.056743 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:45Z","lastTransitionTime":"2025-10-02T01:47:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:45 crc kubenswrapper[4885]: I1002 01:47:45.159859 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:45 crc kubenswrapper[4885]: I1002 01:47:45.159924 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:45 crc kubenswrapper[4885]: I1002 01:47:45.159941 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:45 crc kubenswrapper[4885]: I1002 01:47:45.159967 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:45 crc kubenswrapper[4885]: I1002 01:47:45.159987 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:45Z","lastTransitionTime":"2025-10-02T01:47:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:45 crc kubenswrapper[4885]: I1002 01:47:45.263180 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:45 crc kubenswrapper[4885]: I1002 01:47:45.263240 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:45 crc kubenswrapper[4885]: I1002 01:47:45.263256 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:45 crc kubenswrapper[4885]: I1002 01:47:45.263304 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:45 crc kubenswrapper[4885]: I1002 01:47:45.263321 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:45Z","lastTransitionTime":"2025-10-02T01:47:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:45 crc kubenswrapper[4885]: I1002 01:47:45.366379 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:45 crc kubenswrapper[4885]: I1002 01:47:45.366447 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:45 crc kubenswrapper[4885]: I1002 01:47:45.366466 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:45 crc kubenswrapper[4885]: I1002 01:47:45.366491 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:45 crc kubenswrapper[4885]: I1002 01:47:45.366509 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:45Z","lastTransitionTime":"2025-10-02T01:47:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:45 crc kubenswrapper[4885]: I1002 01:47:45.469482 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:45 crc kubenswrapper[4885]: I1002 01:47:45.469602 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:45 crc kubenswrapper[4885]: I1002 01:47:45.469623 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:45 crc kubenswrapper[4885]: I1002 01:47:45.469644 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:45 crc kubenswrapper[4885]: I1002 01:47:45.469661 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:45Z","lastTransitionTime":"2025-10-02T01:47:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:45 crc kubenswrapper[4885]: I1002 01:47:45.572227 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:45 crc kubenswrapper[4885]: I1002 01:47:45.572315 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:45 crc kubenswrapper[4885]: I1002 01:47:45.572333 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:45 crc kubenswrapper[4885]: I1002 01:47:45.572356 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:45 crc kubenswrapper[4885]: I1002 01:47:45.572374 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:45Z","lastTransitionTime":"2025-10-02T01:47:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:45 crc kubenswrapper[4885]: I1002 01:47:45.674558 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:45 crc kubenswrapper[4885]: I1002 01:47:45.674589 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:45 crc kubenswrapper[4885]: I1002 01:47:45.674598 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:45 crc kubenswrapper[4885]: I1002 01:47:45.674609 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:45 crc kubenswrapper[4885]: I1002 01:47:45.674619 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:45Z","lastTransitionTime":"2025-10-02T01:47:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:45 crc kubenswrapper[4885]: I1002 01:47:45.777926 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:45 crc kubenswrapper[4885]: I1002 01:47:45.778000 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:45 crc kubenswrapper[4885]: I1002 01:47:45.778020 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:45 crc kubenswrapper[4885]: I1002 01:47:45.778049 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:45 crc kubenswrapper[4885]: I1002 01:47:45.778073 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:45Z","lastTransitionTime":"2025-10-02T01:47:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:45 crc kubenswrapper[4885]: I1002 01:47:45.881316 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:45 crc kubenswrapper[4885]: I1002 01:47:45.881409 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:45 crc kubenswrapper[4885]: I1002 01:47:45.881428 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:45 crc kubenswrapper[4885]: I1002 01:47:45.881456 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:45 crc kubenswrapper[4885]: I1002 01:47:45.881477 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:45Z","lastTransitionTime":"2025-10-02T01:47:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:45 crc kubenswrapper[4885]: I1002 01:47:45.984378 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:45 crc kubenswrapper[4885]: I1002 01:47:45.984443 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:45 crc kubenswrapper[4885]: I1002 01:47:45.984461 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:45 crc kubenswrapper[4885]: I1002 01:47:45.984486 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:45 crc kubenswrapper[4885]: I1002 01:47:45.984504 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:45Z","lastTransitionTime":"2025-10-02T01:47:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:46 crc kubenswrapper[4885]: I1002 01:47:46.045630 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:47:46 crc kubenswrapper[4885]: E1002 01:47:46.045801 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:47:46 crc kubenswrapper[4885]: I1002 01:47:46.045888 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:47:46 crc kubenswrapper[4885]: I1002 01:47:46.045898 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:47:46 crc kubenswrapper[4885]: I1002 01:47:46.045899 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sv4md" Oct 02 01:47:46 crc kubenswrapper[4885]: E1002 01:47:46.046357 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:47:46 crc kubenswrapper[4885]: E1002 01:47:46.046457 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:47:46 crc kubenswrapper[4885]: E1002 01:47:46.046542 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sv4md" podUID="724a3e49-001c-45a4-a896-82c13e0ee0e7" Oct 02 01:47:46 crc kubenswrapper[4885]: I1002 01:47:46.087676 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:46 crc kubenswrapper[4885]: I1002 01:47:46.087747 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:46 crc kubenswrapper[4885]: I1002 01:47:46.087765 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:46 crc kubenswrapper[4885]: I1002 01:47:46.087790 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:46 crc kubenswrapper[4885]: I1002 01:47:46.087810 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:46Z","lastTransitionTime":"2025-10-02T01:47:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:46 crc kubenswrapper[4885]: I1002 01:47:46.191570 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:46 crc kubenswrapper[4885]: I1002 01:47:46.191636 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:46 crc kubenswrapper[4885]: I1002 01:47:46.191653 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:46 crc kubenswrapper[4885]: I1002 01:47:46.191678 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:46 crc kubenswrapper[4885]: I1002 01:47:46.191697 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:46Z","lastTransitionTime":"2025-10-02T01:47:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:46 crc kubenswrapper[4885]: I1002 01:47:46.295330 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:46 crc kubenswrapper[4885]: I1002 01:47:46.295389 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:46 crc kubenswrapper[4885]: I1002 01:47:46.295405 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:46 crc kubenswrapper[4885]: I1002 01:47:46.295429 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:46 crc kubenswrapper[4885]: I1002 01:47:46.295445 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:46Z","lastTransitionTime":"2025-10-02T01:47:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:46 crc kubenswrapper[4885]: I1002 01:47:46.398292 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:46 crc kubenswrapper[4885]: I1002 01:47:46.398356 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:46 crc kubenswrapper[4885]: I1002 01:47:46.398412 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:46 crc kubenswrapper[4885]: I1002 01:47:46.398437 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:46 crc kubenswrapper[4885]: I1002 01:47:46.398457 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:46Z","lastTransitionTime":"2025-10-02T01:47:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:46 crc kubenswrapper[4885]: I1002 01:47:46.501547 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:46 crc kubenswrapper[4885]: I1002 01:47:46.501605 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:46 crc kubenswrapper[4885]: I1002 01:47:46.501622 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:46 crc kubenswrapper[4885]: I1002 01:47:46.501676 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:46 crc kubenswrapper[4885]: I1002 01:47:46.501693 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:46Z","lastTransitionTime":"2025-10-02T01:47:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:46 crc kubenswrapper[4885]: I1002 01:47:46.604602 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:46 crc kubenswrapper[4885]: I1002 01:47:46.604668 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:46 crc kubenswrapper[4885]: I1002 01:47:46.604693 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:46 crc kubenswrapper[4885]: I1002 01:47:46.604722 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:46 crc kubenswrapper[4885]: I1002 01:47:46.604746 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:46Z","lastTransitionTime":"2025-10-02T01:47:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:46 crc kubenswrapper[4885]: I1002 01:47:46.707055 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:46 crc kubenswrapper[4885]: I1002 01:47:46.707108 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:46 crc kubenswrapper[4885]: I1002 01:47:46.707125 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:46 crc kubenswrapper[4885]: I1002 01:47:46.707145 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:46 crc kubenswrapper[4885]: I1002 01:47:46.707162 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:46Z","lastTransitionTime":"2025-10-02T01:47:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:46 crc kubenswrapper[4885]: I1002 01:47:46.810347 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:46 crc kubenswrapper[4885]: I1002 01:47:46.810413 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:46 crc kubenswrapper[4885]: I1002 01:47:46.810429 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:46 crc kubenswrapper[4885]: I1002 01:47:46.810556 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:46 crc kubenswrapper[4885]: I1002 01:47:46.810575 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:46Z","lastTransitionTime":"2025-10-02T01:47:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:46 crc kubenswrapper[4885]: I1002 01:47:46.912859 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:46 crc kubenswrapper[4885]: I1002 01:47:46.912922 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:46 crc kubenswrapper[4885]: I1002 01:47:46.912939 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:46 crc kubenswrapper[4885]: I1002 01:47:46.912961 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:46 crc kubenswrapper[4885]: I1002 01:47:46.912979 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:46Z","lastTransitionTime":"2025-10-02T01:47:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:47 crc kubenswrapper[4885]: I1002 01:47:47.015664 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:47 crc kubenswrapper[4885]: I1002 01:47:47.015722 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:47 crc kubenswrapper[4885]: I1002 01:47:47.015739 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:47 crc kubenswrapper[4885]: I1002 01:47:47.015764 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:47 crc kubenswrapper[4885]: I1002 01:47:47.015780 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:47Z","lastTransitionTime":"2025-10-02T01:47:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:47 crc kubenswrapper[4885]: I1002 01:47:47.118128 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:47 crc kubenswrapper[4885]: I1002 01:47:47.118198 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:47 crc kubenswrapper[4885]: I1002 01:47:47.118216 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:47 crc kubenswrapper[4885]: I1002 01:47:47.118242 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:47 crc kubenswrapper[4885]: I1002 01:47:47.118302 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:47Z","lastTransitionTime":"2025-10-02T01:47:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:47 crc kubenswrapper[4885]: I1002 01:47:47.221592 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:47 crc kubenswrapper[4885]: I1002 01:47:47.221650 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:47 crc kubenswrapper[4885]: I1002 01:47:47.221667 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:47 crc kubenswrapper[4885]: I1002 01:47:47.221689 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:47 crc kubenswrapper[4885]: I1002 01:47:47.221709 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:47Z","lastTransitionTime":"2025-10-02T01:47:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:47 crc kubenswrapper[4885]: I1002 01:47:47.325148 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:47 crc kubenswrapper[4885]: I1002 01:47:47.325215 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:47 crc kubenswrapper[4885]: I1002 01:47:47.325241 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:47 crc kubenswrapper[4885]: I1002 01:47:47.325306 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:47 crc kubenswrapper[4885]: I1002 01:47:47.325334 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:47Z","lastTransitionTime":"2025-10-02T01:47:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:47 crc kubenswrapper[4885]: I1002 01:47:47.427939 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:47 crc kubenswrapper[4885]: I1002 01:47:47.428003 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:47 crc kubenswrapper[4885]: I1002 01:47:47.428026 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:47 crc kubenswrapper[4885]: I1002 01:47:47.428055 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:47 crc kubenswrapper[4885]: I1002 01:47:47.428079 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:47Z","lastTransitionTime":"2025-10-02T01:47:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:47 crc kubenswrapper[4885]: I1002 01:47:47.530632 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:47 crc kubenswrapper[4885]: I1002 01:47:47.530670 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:47 crc kubenswrapper[4885]: I1002 01:47:47.530683 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:47 crc kubenswrapper[4885]: I1002 01:47:47.530699 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:47 crc kubenswrapper[4885]: I1002 01:47:47.530710 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:47Z","lastTransitionTime":"2025-10-02T01:47:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:47 crc kubenswrapper[4885]: I1002 01:47:47.632960 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:47 crc kubenswrapper[4885]: I1002 01:47:47.633025 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:47 crc kubenswrapper[4885]: I1002 01:47:47.633043 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:47 crc kubenswrapper[4885]: I1002 01:47:47.633069 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:47 crc kubenswrapper[4885]: I1002 01:47:47.633088 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:47Z","lastTransitionTime":"2025-10-02T01:47:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:47 crc kubenswrapper[4885]: I1002 01:47:47.736714 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:47 crc kubenswrapper[4885]: I1002 01:47:47.736786 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:47 crc kubenswrapper[4885]: I1002 01:47:47.736805 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:47 crc kubenswrapper[4885]: I1002 01:47:47.736830 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:47 crc kubenswrapper[4885]: I1002 01:47:47.736849 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:47Z","lastTransitionTime":"2025-10-02T01:47:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:47 crc kubenswrapper[4885]: I1002 01:47:47.840195 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:47 crc kubenswrapper[4885]: I1002 01:47:47.840255 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:47 crc kubenswrapper[4885]: I1002 01:47:47.840312 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:47 crc kubenswrapper[4885]: I1002 01:47:47.840337 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:47 crc kubenswrapper[4885]: I1002 01:47:47.840356 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:47Z","lastTransitionTime":"2025-10-02T01:47:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:47 crc kubenswrapper[4885]: I1002 01:47:47.943839 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:47 crc kubenswrapper[4885]: I1002 01:47:47.943887 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:47 crc kubenswrapper[4885]: I1002 01:47:47.943908 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:47 crc kubenswrapper[4885]: I1002 01:47:47.943932 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:47 crc kubenswrapper[4885]: I1002 01:47:47.943966 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:47Z","lastTransitionTime":"2025-10-02T01:47:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:48 crc kubenswrapper[4885]: I1002 01:47:48.045603 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:47:48 crc kubenswrapper[4885]: I1002 01:47:48.045654 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sv4md" Oct 02 01:47:48 crc kubenswrapper[4885]: I1002 01:47:48.045663 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:47:48 crc kubenswrapper[4885]: I1002 01:47:48.045618 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:47:48 crc kubenswrapper[4885]: E1002 01:47:48.045769 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:47:48 crc kubenswrapper[4885]: E1002 01:47:48.045906 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sv4md" podUID="724a3e49-001c-45a4-a896-82c13e0ee0e7" Oct 02 01:47:48 crc kubenswrapper[4885]: E1002 01:47:48.046101 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:47:48 crc kubenswrapper[4885]: E1002 01:47:48.046215 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:47:48 crc kubenswrapper[4885]: I1002 01:47:48.046728 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:48 crc kubenswrapper[4885]: I1002 01:47:48.046757 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:48 crc kubenswrapper[4885]: I1002 01:47:48.046774 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:48 crc kubenswrapper[4885]: I1002 01:47:48.046795 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:48 crc kubenswrapper[4885]: I1002 01:47:48.046811 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:48Z","lastTransitionTime":"2025-10-02T01:47:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:48 crc kubenswrapper[4885]: I1002 01:47:48.149696 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:48 crc kubenswrapper[4885]: I1002 01:47:48.149759 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:48 crc kubenswrapper[4885]: I1002 01:47:48.149776 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:48 crc kubenswrapper[4885]: I1002 01:47:48.149800 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:48 crc kubenswrapper[4885]: I1002 01:47:48.149816 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:48Z","lastTransitionTime":"2025-10-02T01:47:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:48 crc kubenswrapper[4885]: I1002 01:47:48.253752 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:48 crc kubenswrapper[4885]: I1002 01:47:48.253816 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:48 crc kubenswrapper[4885]: I1002 01:47:48.253838 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:48 crc kubenswrapper[4885]: I1002 01:47:48.253867 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:48 crc kubenswrapper[4885]: I1002 01:47:48.253887 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:48Z","lastTransitionTime":"2025-10-02T01:47:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:48 crc kubenswrapper[4885]: I1002 01:47:48.356393 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:48 crc kubenswrapper[4885]: I1002 01:47:48.356477 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:48 crc kubenswrapper[4885]: I1002 01:47:48.356500 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:48 crc kubenswrapper[4885]: I1002 01:47:48.356528 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:48 crc kubenswrapper[4885]: I1002 01:47:48.356550 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:48Z","lastTransitionTime":"2025-10-02T01:47:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:48 crc kubenswrapper[4885]: I1002 01:47:48.463298 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:48 crc kubenswrapper[4885]: I1002 01:47:48.463363 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:48 crc kubenswrapper[4885]: I1002 01:47:48.463382 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:48 crc kubenswrapper[4885]: I1002 01:47:48.463418 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:48 crc kubenswrapper[4885]: I1002 01:47:48.463441 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:48Z","lastTransitionTime":"2025-10-02T01:47:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:48 crc kubenswrapper[4885]: I1002 01:47:48.566866 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:48 crc kubenswrapper[4885]: I1002 01:47:48.566949 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:48 crc kubenswrapper[4885]: I1002 01:47:48.566967 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:48 crc kubenswrapper[4885]: I1002 01:47:48.566991 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:48 crc kubenswrapper[4885]: I1002 01:47:48.567010 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:48Z","lastTransitionTime":"2025-10-02T01:47:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:48 crc kubenswrapper[4885]: I1002 01:47:48.669753 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:48 crc kubenswrapper[4885]: I1002 01:47:48.669828 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:48 crc kubenswrapper[4885]: I1002 01:47:48.669847 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:48 crc kubenswrapper[4885]: I1002 01:47:48.669876 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:48 crc kubenswrapper[4885]: I1002 01:47:48.669896 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:48Z","lastTransitionTime":"2025-10-02T01:47:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:48 crc kubenswrapper[4885]: I1002 01:47:48.772952 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:48 crc kubenswrapper[4885]: I1002 01:47:48.773013 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:48 crc kubenswrapper[4885]: I1002 01:47:48.773030 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:48 crc kubenswrapper[4885]: I1002 01:47:48.773058 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:48 crc kubenswrapper[4885]: I1002 01:47:48.773083 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:48Z","lastTransitionTime":"2025-10-02T01:47:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:48 crc kubenswrapper[4885]: I1002 01:47:48.877447 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:48 crc kubenswrapper[4885]: I1002 01:47:48.877515 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:48 crc kubenswrapper[4885]: I1002 01:47:48.877534 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:48 crc kubenswrapper[4885]: I1002 01:47:48.877559 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:48 crc kubenswrapper[4885]: I1002 01:47:48.877577 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:48Z","lastTransitionTime":"2025-10-02T01:47:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:48 crc kubenswrapper[4885]: I1002 01:47:48.980530 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:48 crc kubenswrapper[4885]: I1002 01:47:48.980603 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:48 crc kubenswrapper[4885]: I1002 01:47:48.980620 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:48 crc kubenswrapper[4885]: I1002 01:47:48.980644 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:48 crc kubenswrapper[4885]: I1002 01:47:48.980727 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:48Z","lastTransitionTime":"2025-10-02T01:47:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:48 crc kubenswrapper[4885]: I1002 01:47:48.997571 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:48 crc kubenswrapper[4885]: I1002 01:47:48.997635 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:48 crc kubenswrapper[4885]: I1002 01:47:48.997659 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:48 crc kubenswrapper[4885]: I1002 01:47:48.997690 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:48 crc kubenswrapper[4885]: I1002 01:47:48.997710 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:48Z","lastTransitionTime":"2025-10-02T01:47:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:49 crc kubenswrapper[4885]: E1002 01:47:49.018462 4885 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:47:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:47:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:47:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:47:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"92603ae5-ba38-4201-98e9-979fd56cb50d\\\",\\\"systemUUID\\\":\\\"50ddf2c2-ea95-4960-ae48-d17e9484ee2f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:49Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:49 crc kubenswrapper[4885]: I1002 01:47:49.023760 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:49 crc kubenswrapper[4885]: I1002 01:47:49.023823 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:49 crc kubenswrapper[4885]: I1002 01:47:49.023850 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:49 crc kubenswrapper[4885]: I1002 01:47:49.023880 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:49 crc kubenswrapper[4885]: I1002 01:47:49.023905 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:49Z","lastTransitionTime":"2025-10-02T01:47:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:49 crc kubenswrapper[4885]: E1002 01:47:49.045156 4885 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:47:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:47:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:47:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:47:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"92603ae5-ba38-4201-98e9-979fd56cb50d\\\",\\\"systemUUID\\\":\\\"50ddf2c2-ea95-4960-ae48-d17e9484ee2f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:49Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:49 crc kubenswrapper[4885]: I1002 01:47:49.051422 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:49 crc kubenswrapper[4885]: I1002 01:47:49.051463 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:49 crc kubenswrapper[4885]: I1002 01:47:49.051481 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:49 crc kubenswrapper[4885]: I1002 01:47:49.051504 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:49 crc kubenswrapper[4885]: I1002 01:47:49.051520 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:49Z","lastTransitionTime":"2025-10-02T01:47:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:49 crc kubenswrapper[4885]: E1002 01:47:49.074193 4885 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:47:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:47:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:47:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:47:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"92603ae5-ba38-4201-98e9-979fd56cb50d\\\",\\\"systemUUID\\\":\\\"50ddf2c2-ea95-4960-ae48-d17e9484ee2f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:49Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:49 crc kubenswrapper[4885]: I1002 01:47:49.079077 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:49 crc kubenswrapper[4885]: I1002 01:47:49.079132 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:49 crc kubenswrapper[4885]: I1002 01:47:49.079149 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:49 crc kubenswrapper[4885]: I1002 01:47:49.079173 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:49 crc kubenswrapper[4885]: I1002 01:47:49.079192 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:49Z","lastTransitionTime":"2025-10-02T01:47:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:49 crc kubenswrapper[4885]: E1002 01:47:49.102502 4885 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:47:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:47:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:47:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:47:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"92603ae5-ba38-4201-98e9-979fd56cb50d\\\",\\\"systemUUID\\\":\\\"50ddf2c2-ea95-4960-ae48-d17e9484ee2f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:49Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:49 crc kubenswrapper[4885]: I1002 01:47:49.107909 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:49 crc kubenswrapper[4885]: I1002 01:47:49.107996 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:49 crc kubenswrapper[4885]: I1002 01:47:49.108043 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:49 crc kubenswrapper[4885]: I1002 01:47:49.108069 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:49 crc kubenswrapper[4885]: I1002 01:47:49.108086 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:49Z","lastTransitionTime":"2025-10-02T01:47:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:49 crc kubenswrapper[4885]: E1002 01:47:49.130320 4885 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:47:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:47:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:47:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:47:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"92603ae5-ba38-4201-98e9-979fd56cb50d\\\",\\\"systemUUID\\\":\\\"50ddf2c2-ea95-4960-ae48-d17e9484ee2f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:49Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:49 crc kubenswrapper[4885]: E1002 01:47:49.130611 4885 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 02 01:47:49 crc kubenswrapper[4885]: I1002 01:47:49.133360 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:49 crc kubenswrapper[4885]: I1002 01:47:49.133436 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:49 crc kubenswrapper[4885]: I1002 01:47:49.133483 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:49 crc kubenswrapper[4885]: I1002 01:47:49.133507 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:49 crc kubenswrapper[4885]: I1002 01:47:49.133523 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:49Z","lastTransitionTime":"2025-10-02T01:47:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:49 crc kubenswrapper[4885]: I1002 01:47:49.237237 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:49 crc kubenswrapper[4885]: I1002 01:47:49.237332 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:49 crc kubenswrapper[4885]: I1002 01:47:49.237349 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:49 crc kubenswrapper[4885]: I1002 01:47:49.237371 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:49 crc kubenswrapper[4885]: I1002 01:47:49.237389 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:49Z","lastTransitionTime":"2025-10-02T01:47:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:49 crc kubenswrapper[4885]: I1002 01:47:49.340664 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:49 crc kubenswrapper[4885]: I1002 01:47:49.340742 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:49 crc kubenswrapper[4885]: I1002 01:47:49.340765 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:49 crc kubenswrapper[4885]: I1002 01:47:49.340794 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:49 crc kubenswrapper[4885]: I1002 01:47:49.340816 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:49Z","lastTransitionTime":"2025-10-02T01:47:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:49 crc kubenswrapper[4885]: I1002 01:47:49.443163 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:49 crc kubenswrapper[4885]: I1002 01:47:49.443236 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:49 crc kubenswrapper[4885]: I1002 01:47:49.443253 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:49 crc kubenswrapper[4885]: I1002 01:47:49.443307 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:49 crc kubenswrapper[4885]: I1002 01:47:49.443325 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:49Z","lastTransitionTime":"2025-10-02T01:47:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:49 crc kubenswrapper[4885]: I1002 01:47:49.546153 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:49 crc kubenswrapper[4885]: I1002 01:47:49.546202 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:49 crc kubenswrapper[4885]: I1002 01:47:49.546218 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:49 crc kubenswrapper[4885]: I1002 01:47:49.546241 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:49 crc kubenswrapper[4885]: I1002 01:47:49.546282 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:49Z","lastTransitionTime":"2025-10-02T01:47:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:49 crc kubenswrapper[4885]: I1002 01:47:49.649606 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:49 crc kubenswrapper[4885]: I1002 01:47:49.649735 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:49 crc kubenswrapper[4885]: I1002 01:47:49.649758 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:49 crc kubenswrapper[4885]: I1002 01:47:49.649781 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:49 crc kubenswrapper[4885]: I1002 01:47:49.649800 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:49Z","lastTransitionTime":"2025-10-02T01:47:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:49 crc kubenswrapper[4885]: I1002 01:47:49.753342 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:49 crc kubenswrapper[4885]: I1002 01:47:49.753437 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:49 crc kubenswrapper[4885]: I1002 01:47:49.753456 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:49 crc kubenswrapper[4885]: I1002 01:47:49.753481 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:49 crc kubenswrapper[4885]: I1002 01:47:49.753498 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:49Z","lastTransitionTime":"2025-10-02T01:47:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:49 crc kubenswrapper[4885]: I1002 01:47:49.855877 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:49 crc kubenswrapper[4885]: I1002 01:47:49.855953 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:49 crc kubenswrapper[4885]: I1002 01:47:49.855974 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:49 crc kubenswrapper[4885]: I1002 01:47:49.855999 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:49 crc kubenswrapper[4885]: I1002 01:47:49.856017 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:49Z","lastTransitionTime":"2025-10-02T01:47:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:49 crc kubenswrapper[4885]: I1002 01:47:49.958395 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:49 crc kubenswrapper[4885]: I1002 01:47:49.958442 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:49 crc kubenswrapper[4885]: I1002 01:47:49.958462 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:49 crc kubenswrapper[4885]: I1002 01:47:49.958486 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:49 crc kubenswrapper[4885]: I1002 01:47:49.958503 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:49Z","lastTransitionTime":"2025-10-02T01:47:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:50 crc kubenswrapper[4885]: I1002 01:47:50.045992 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sv4md" Oct 02 01:47:50 crc kubenswrapper[4885]: I1002 01:47:50.046163 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:47:50 crc kubenswrapper[4885]: E1002 01:47:50.046184 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sv4md" podUID="724a3e49-001c-45a4-a896-82c13e0ee0e7" Oct 02 01:47:50 crc kubenswrapper[4885]: E1002 01:47:50.046453 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:47:50 crc kubenswrapper[4885]: I1002 01:47:50.046498 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:47:50 crc kubenswrapper[4885]: I1002 01:47:50.046550 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:47:50 crc kubenswrapper[4885]: E1002 01:47:50.046633 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:47:50 crc kubenswrapper[4885]: E1002 01:47:50.046717 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:47:50 crc kubenswrapper[4885]: I1002 01:47:50.064510 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:50 crc kubenswrapper[4885]: I1002 01:47:50.064570 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:50 crc kubenswrapper[4885]: I1002 01:47:50.064591 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:50 crc kubenswrapper[4885]: I1002 01:47:50.064615 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:50 crc kubenswrapper[4885]: I1002 01:47:50.064635 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:50Z","lastTransitionTime":"2025-10-02T01:47:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:50 crc kubenswrapper[4885]: I1002 01:47:50.174174 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:50 crc kubenswrapper[4885]: I1002 01:47:50.174232 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:50 crc kubenswrapper[4885]: I1002 01:47:50.174249 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:50 crc kubenswrapper[4885]: I1002 01:47:50.174301 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:50 crc kubenswrapper[4885]: I1002 01:47:50.174319 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:50Z","lastTransitionTime":"2025-10-02T01:47:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:50 crc kubenswrapper[4885]: I1002 01:47:50.277827 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:50 crc kubenswrapper[4885]: I1002 01:47:50.277895 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:50 crc kubenswrapper[4885]: I1002 01:47:50.277916 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:50 crc kubenswrapper[4885]: I1002 01:47:50.277943 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:50 crc kubenswrapper[4885]: I1002 01:47:50.277963 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:50Z","lastTransitionTime":"2025-10-02T01:47:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:50 crc kubenswrapper[4885]: I1002 01:47:50.380907 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:50 crc kubenswrapper[4885]: I1002 01:47:50.380954 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:50 crc kubenswrapper[4885]: I1002 01:47:50.380971 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:50 crc kubenswrapper[4885]: I1002 01:47:50.380991 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:50 crc kubenswrapper[4885]: I1002 01:47:50.381007 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:50Z","lastTransitionTime":"2025-10-02T01:47:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:50 crc kubenswrapper[4885]: I1002 01:47:50.483419 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:50 crc kubenswrapper[4885]: I1002 01:47:50.483467 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:50 crc kubenswrapper[4885]: I1002 01:47:50.483483 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:50 crc kubenswrapper[4885]: I1002 01:47:50.483503 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:50 crc kubenswrapper[4885]: I1002 01:47:50.483520 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:50Z","lastTransitionTime":"2025-10-02T01:47:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:50 crc kubenswrapper[4885]: I1002 01:47:50.586457 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:50 crc kubenswrapper[4885]: I1002 01:47:50.586526 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:50 crc kubenswrapper[4885]: I1002 01:47:50.586544 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:50 crc kubenswrapper[4885]: I1002 01:47:50.586568 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:50 crc kubenswrapper[4885]: I1002 01:47:50.586586 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:50Z","lastTransitionTime":"2025-10-02T01:47:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:50 crc kubenswrapper[4885]: I1002 01:47:50.689729 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:50 crc kubenswrapper[4885]: I1002 01:47:50.689788 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:50 crc kubenswrapper[4885]: I1002 01:47:50.689805 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:50 crc kubenswrapper[4885]: I1002 01:47:50.689827 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:50 crc kubenswrapper[4885]: I1002 01:47:50.689843 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:50Z","lastTransitionTime":"2025-10-02T01:47:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:50 crc kubenswrapper[4885]: I1002 01:47:50.792879 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:50 crc kubenswrapper[4885]: I1002 01:47:50.792950 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:50 crc kubenswrapper[4885]: I1002 01:47:50.792970 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:50 crc kubenswrapper[4885]: I1002 01:47:50.793018 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:50 crc kubenswrapper[4885]: I1002 01:47:50.793038 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:50Z","lastTransitionTime":"2025-10-02T01:47:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:50 crc kubenswrapper[4885]: I1002 01:47:50.896695 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:50 crc kubenswrapper[4885]: I1002 01:47:50.896763 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:50 crc kubenswrapper[4885]: I1002 01:47:50.896783 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:50 crc kubenswrapper[4885]: I1002 01:47:50.896813 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:50 crc kubenswrapper[4885]: I1002 01:47:50.896832 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:50Z","lastTransitionTime":"2025-10-02T01:47:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:50 crc kubenswrapper[4885]: I1002 01:47:50.998942 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:50 crc kubenswrapper[4885]: I1002 01:47:50.999037 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:50 crc kubenswrapper[4885]: I1002 01:47:50.999055 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:50 crc kubenswrapper[4885]: I1002 01:47:50.999079 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:50 crc kubenswrapper[4885]: I1002 01:47:50.999099 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:50Z","lastTransitionTime":"2025-10-02T01:47:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:51 crc kubenswrapper[4885]: I1002 01:47:51.102791 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:51 crc kubenswrapper[4885]: I1002 01:47:51.102842 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:51 crc kubenswrapper[4885]: I1002 01:47:51.102862 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:51 crc kubenswrapper[4885]: I1002 01:47:51.102887 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:51 crc kubenswrapper[4885]: I1002 01:47:51.102907 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:51Z","lastTransitionTime":"2025-10-02T01:47:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:51 crc kubenswrapper[4885]: I1002 01:47:51.206062 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:51 crc kubenswrapper[4885]: I1002 01:47:51.206130 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:51 crc kubenswrapper[4885]: I1002 01:47:51.206149 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:51 crc kubenswrapper[4885]: I1002 01:47:51.206175 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:51 crc kubenswrapper[4885]: I1002 01:47:51.206194 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:51Z","lastTransitionTime":"2025-10-02T01:47:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:51 crc kubenswrapper[4885]: I1002 01:47:51.309806 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:51 crc kubenswrapper[4885]: I1002 01:47:51.309871 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:51 crc kubenswrapper[4885]: I1002 01:47:51.309889 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:51 crc kubenswrapper[4885]: I1002 01:47:51.309914 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:51 crc kubenswrapper[4885]: I1002 01:47:51.309936 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:51Z","lastTransitionTime":"2025-10-02T01:47:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:51 crc kubenswrapper[4885]: I1002 01:47:51.412985 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:51 crc kubenswrapper[4885]: I1002 01:47:51.413064 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:51 crc kubenswrapper[4885]: I1002 01:47:51.413090 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:51 crc kubenswrapper[4885]: I1002 01:47:51.413120 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:51 crc kubenswrapper[4885]: I1002 01:47:51.413145 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:51Z","lastTransitionTime":"2025-10-02T01:47:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:51 crc kubenswrapper[4885]: I1002 01:47:51.516956 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:51 crc kubenswrapper[4885]: I1002 01:47:51.517037 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:51 crc kubenswrapper[4885]: I1002 01:47:51.517062 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:51 crc kubenswrapper[4885]: I1002 01:47:51.517094 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:51 crc kubenswrapper[4885]: I1002 01:47:51.517118 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:51Z","lastTransitionTime":"2025-10-02T01:47:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:51 crc kubenswrapper[4885]: I1002 01:47:51.619823 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:51 crc kubenswrapper[4885]: I1002 01:47:51.619915 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:51 crc kubenswrapper[4885]: I1002 01:47:51.619946 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:51 crc kubenswrapper[4885]: I1002 01:47:51.619984 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:51 crc kubenswrapper[4885]: I1002 01:47:51.620008 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:51Z","lastTransitionTime":"2025-10-02T01:47:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:51 crc kubenswrapper[4885]: I1002 01:47:51.723384 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:51 crc kubenswrapper[4885]: I1002 01:47:51.723462 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:51 crc kubenswrapper[4885]: I1002 01:47:51.723482 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:51 crc kubenswrapper[4885]: I1002 01:47:51.723513 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:51 crc kubenswrapper[4885]: I1002 01:47:51.723539 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:51Z","lastTransitionTime":"2025-10-02T01:47:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:51 crc kubenswrapper[4885]: I1002 01:47:51.826207 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:51 crc kubenswrapper[4885]: I1002 01:47:51.826304 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:51 crc kubenswrapper[4885]: I1002 01:47:51.826322 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:51 crc kubenswrapper[4885]: I1002 01:47:51.826348 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:51 crc kubenswrapper[4885]: I1002 01:47:51.826366 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:51Z","lastTransitionTime":"2025-10-02T01:47:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:51 crc kubenswrapper[4885]: I1002 01:47:51.929545 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:51 crc kubenswrapper[4885]: I1002 01:47:51.929606 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:51 crc kubenswrapper[4885]: I1002 01:47:51.929624 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:51 crc kubenswrapper[4885]: I1002 01:47:51.929650 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:51 crc kubenswrapper[4885]: I1002 01:47:51.929671 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:51Z","lastTransitionTime":"2025-10-02T01:47:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:52 crc kubenswrapper[4885]: I1002 01:47:52.032102 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:52 crc kubenswrapper[4885]: I1002 01:47:52.032190 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:52 crc kubenswrapper[4885]: I1002 01:47:52.032219 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:52 crc kubenswrapper[4885]: I1002 01:47:52.032250 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:52 crc kubenswrapper[4885]: I1002 01:47:52.032307 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:52Z","lastTransitionTime":"2025-10-02T01:47:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:52 crc kubenswrapper[4885]: I1002 01:47:52.045641 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:47:52 crc kubenswrapper[4885]: I1002 01:47:52.045731 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:47:52 crc kubenswrapper[4885]: I1002 01:47:52.045787 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:47:52 crc kubenswrapper[4885]: I1002 01:47:52.045809 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sv4md" Oct 02 01:47:52 crc kubenswrapper[4885]: E1002 01:47:52.049154 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:47:52 crc kubenswrapper[4885]: E1002 01:47:52.049447 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:47:52 crc kubenswrapper[4885]: E1002 01:47:52.049602 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:47:52 crc kubenswrapper[4885]: E1002 01:47:52.049715 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sv4md" podUID="724a3e49-001c-45a4-a896-82c13e0ee0e7" Oct 02 01:47:52 crc kubenswrapper[4885]: I1002 01:47:52.068856 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:52Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:52 crc kubenswrapper[4885]: I1002 01:47:52.090056 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:52Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:52 crc kubenswrapper[4885]: I1002 01:47:52.107079 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-58lhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac6a19c4-f259-4dbd-a016-d6c2c849c6cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed76c0f968c063f7795d363af194a222487b647ceda234de0dabb39b25eafe0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hzc4c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-58lhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:52Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:52 crc kubenswrapper[4885]: I1002 01:47:52.124863 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d169c8ba-000a-4218-8e3b-4ae53035b110\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61e436b7876107ce5ea9e668c0bbc367be6e663fc804fa3a65b51adf8be8d51f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-77rzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09536e182a46e77318c10812d1777e551f834907c006cab5c726dc66f5bce88a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-77rzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rttx8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:52Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:52 crc kubenswrapper[4885]: I1002 01:47:52.135901 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:52 crc kubenswrapper[4885]: I1002 01:47:52.135959 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:52 crc kubenswrapper[4885]: I1002 01:47:52.135978 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:52 crc kubenswrapper[4885]: I1002 01:47:52.136004 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:52 crc kubenswrapper[4885]: I1002 01:47:52.136024 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:52Z","lastTransitionTime":"2025-10-02T01:47:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:52 crc kubenswrapper[4885]: I1002 01:47:52.140239 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jgdvt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81733e8d-67b0-46fa-8a41-08b40645bd3e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcef60c816ca1ba9a298da658e84c4d717bea3d352c4c2f807ebd8e56327b384\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cwbl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jgdvt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:52Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:52 crc kubenswrapper[4885]: I1002 01:47:52.175873 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a0bb885-2eb0-419c-8a7a-ef634cb3c144\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a89b49cbe08d927629564e607de13655ddf3d01ac4eb5d688b325210306c120c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06c5942805a498d9621a906088aa54538b4eed583d1b8d0e3ba2a6718378a36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b9f6f7fe44d5771a838d0b89c5bdea3ebb79023ae2675ee907ec2766b3d35a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://178b4340ee6904c60c814b99d83ee225161aeb23e438d8d02f856579a13ec4e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bb63bf30d6ecce9e71f0db0c16632bdfa6b33924b673378eb0c38ab229fa837\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T01:47:06Z\\\",\\\"message\\\":\\\"W1002 01:46:55.406027 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1002 01:46:55.406714 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759369615 cert, and key in /tmp/serving-cert-3795040473/serving-signer.crt, /tmp/serving-cert-3795040473/serving-signer.key\\\\nI1002 01:46:55.614670 1 observer_polling.go:159] Starting file observer\\\\nW1002 01:46:55.617662 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1002 01:46:55.618112 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 01:46:55.620347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3795040473/tls.crt::/tmp/serving-cert-3795040473/tls.key\\\\\\\"\\\\nF1002 01:47:06.224459 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90535f97ab2ed4626d47540ebd7baada837fd332140323bec0f49aa38d3c0009\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://592f89a326c9c1559193a71904b47ca27b1dbe1a7b30e03bc48924bb2f35f921\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://592f89a326c9c1559193a71904b47ca27b1dbe1a7b30e03bc48924bb2f35f921\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:52Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:52 crc kubenswrapper[4885]: I1002 01:47:52.199878 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b087880e05039816dcab154643ac72c7e4a0a1124817c51dc2c639b62bbee0fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:52Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:52 crc kubenswrapper[4885]: I1002 01:47:52.222981 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"506a599f-fb1c-4d14-be88-0e4833f3d624\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31c2f0193ffef54965b79a3569b0f6f4547e540bac0774155159a1c996c196c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://533e73e5f90b7258acbcdf018ee01323c118f9dc45035b25b065ba01cabecccf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c00db833a593ec40eb63a6b383455e19e960c2e98b9b66f76877ee993057e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21e2f462a0deb0d740185663b2014f4a1e57b4fdbeb82571264b002f928035bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:52Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:52 crc kubenswrapper[4885]: I1002 01:47:52.238815 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:52 crc kubenswrapper[4885]: I1002 01:47:52.238905 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:52 crc kubenswrapper[4885]: I1002 01:47:52.238925 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:52 crc kubenswrapper[4885]: I1002 01:47:52.238957 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:52 crc kubenswrapper[4885]: I1002 01:47:52.238976 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:52Z","lastTransitionTime":"2025-10-02T01:47:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:52 crc kubenswrapper[4885]: I1002 01:47:52.255715 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e4679dd-f870-41e9-a32b-360f5d2ee81b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4310909ad9c7b609ed6814e04a525a6afae747775598e7275080413dec8b23a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7642033b0b5995d332ccf7992d6d7e6a5e9c9369ae2d1071ccb2f536938c8f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9f99f9e26a04fb42b01a0783a5c4e360514818ce9102edc7fa08626d50b9a34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6fa72b030d309656b73121e4ac2a3924a4e449aad66683dd9de5ee5f22d5a4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef403e519d197fb37b3e6b76a14a5128ba3f30a7a5e8bf0a420cf6a39b89575f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8482ba457683ffa03100ec1f07ea9d3f918be5e490b5ba05c2ba965d03033191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://188cb50b58926683ce8862d618258c1d1a6f33310608546c8651ece28f7c338f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://188cb50b58926683ce8862d618258c1d1a6f33310608546c8651ece28f7c338f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T01:47:41Z\\\",\\\"message\\\":\\\"921 6578 obj_retry.go:434] periodicallyRetryResources: Retry channel got triggered: retrying failed objects of type *v1.Pod\\\\nI1002 01:47:40.991945 6578 obj_retry.go:409] Going to retry *v1.Pod resource setup for 16 objects: [openshift-multus/network-metrics-daemon-sv4md openshift-network-operator/iptables-alerter-4ln5h openshift-kube-apiserver/kube-apiserver-crc openshift-ovn-kubernetes/ovnkube-node-dngcm openshift-dns/node-resolver-58lhj openshift-machine-config-operator/machine-config-daemon-rttx8 openshift-multus/multus-vhgbd openshift-network-console/networking-console-plugin-85b44fc459-gdk6g openshift-network-diagnostics/network-check-source-55646444c4-trplf openshift-network-operator/network-operator-58b4c7f79c-55gtf openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rrn8b openshift-image-registry/node-ca-jgdvt openshift-kube-controller-manager/kube-controller-manager-crc openshift-multus/multus-additional-cni-plugins-2ll9b openshift-network-diagnostics/network-check-target-xd92c openshift-network-node-identity/network-node-identity-vrzqb]\\\\nF1002 01:47:40.991998 6578 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller ini\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:40Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-dngcm_openshift-ovn-kubernetes(9e4679dd-f870-41e9-a32b-360f5d2ee81b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fafdd1094bd02786d745c984989852b0c690db38655f92b419733a2004f417c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0125e907d061a97286026ee823c4eed303734ed03ab4864f742592b23a1aa2f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0125e907d061a97286026ee823c4eed303734ed03ab4864f742592b23a1aa2f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dngcm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:52Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:52 crc kubenswrapper[4885]: I1002 01:47:52.274297 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-sv4md" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"724a3e49-001c-45a4-a896-82c13e0ee0e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hpqgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hpqgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-sv4md\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:52Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:52 crc kubenswrapper[4885]: I1002 01:47:52.295400 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vhgbd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c63a7d20d39b65acfb5a9be4c28b21c72111b398bea76c238e15ddb4dd63e139\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7zwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vhgbd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:52Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:52 crc kubenswrapper[4885]: I1002 01:47:52.315473 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rrn8b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52967a77-525c-4537-9b1f-d7bb754a8494\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3dee451ebe9006bf1c519fed4c65267570b6139d76841858352e087ae78d26a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtr74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e9412788d6df7332592beac48281822b26e1aa19b39addce8e1be556af9d982\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtr74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:26Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rrn8b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:52Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:52 crc kubenswrapper[4885]: I1002 01:47:52.335138 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18466979-7c7c-4d73-86c6-b8b6eaf20484\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b742e4b288bb50569d2fd23f3f9857d649a46063998170eee582ecbbba64d190\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://298c2ace58c4a9d09cfdc864316317e370bb0ac5735d631af6d85c6298f686d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://620f54e02c463f6e603808e5e57e9e5770d125578cc258ba16117a2416e57d3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b4cd17b64e5657ae05e71f3ca2340cb0d7965b1b49d1f7b4bfe4494292293ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b4cd17b64e5657ae05e71f3ca2340cb0d7965b1b49d1f7b4bfe4494292293ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:52Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:52 crc kubenswrapper[4885]: I1002 01:47:52.342563 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:52 crc kubenswrapper[4885]: I1002 01:47:52.342617 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:52 crc kubenswrapper[4885]: I1002 01:47:52.342637 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:52 crc kubenswrapper[4885]: I1002 01:47:52.342664 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:52 crc kubenswrapper[4885]: I1002 01:47:52.342683 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:52Z","lastTransitionTime":"2025-10-02T01:47:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:52 crc kubenswrapper[4885]: I1002 01:47:52.370230 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8070bc1c-e3de-4e9a-985e-40d1a9858056\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3e634b3727c517cc48b17c2cfb38f26460f53d1bf25d631d08ec9acb8d780e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4adcc537c23697a324d82af94c31059dd444245efdf7c978db58ede256e112ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edcd1d99644a778dfb6d0a17d2a4abd906400fe8209e611ebf8d7a9abe7ccf28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://97cb59adc2ddfa365ebc9dec649600dc90466ffe59134a6d15a9d3134a7c9b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d2808bc658ea80d68469ed16422d56ef19447f4c90b4af7b2597c64406f8a44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://83acbc20919f7c1958b6d492fadb203855df02761662425df9f91a0757fe2bf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83acbc20919f7c1958b6d492fadb203855df02761662425df9f91a0757fe2bf7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e86ecd03835efc9a253b6804bc61e6ac690e75b8d1f3cc77f5647730e586574\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e86ecd03835efc9a253b6804bc61e6ac690e75b8d1f3cc77f5647730e586574\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://48f95af51feaf014a18496bdfadd59d7f1e11167c5c8b677fa7037e48368ff2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48f95af51feaf014a18496bdfadd59d7f1e11167c5c8b677fa7037e48368ff2a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:52Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:52 crc kubenswrapper[4885]: I1002 01:47:52.391004 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5c1f95f188569c9aa47e51f2916939fa0b65a5cdf4c16b798612d41f3c8411a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:52Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:52 crc kubenswrapper[4885]: I1002 01:47:52.411103 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:52Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:52 crc kubenswrapper[4885]: I1002 01:47:52.431554 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74adcf912ccedb6b5d4f2803fd215e3e2ef62367597aa6da528354c4511ee536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7403ef968d040ac09d722b7557438bef23892d578992826f42b700b370d495d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:52Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:52 crc kubenswrapper[4885]: I1002 01:47:52.445393 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:52 crc kubenswrapper[4885]: I1002 01:47:52.445619 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:52 crc kubenswrapper[4885]: I1002 01:47:52.445772 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:52 crc kubenswrapper[4885]: I1002 01:47:52.445915 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:52 crc kubenswrapper[4885]: I1002 01:47:52.446206 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:52Z","lastTransitionTime":"2025-10-02T01:47:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:52 crc kubenswrapper[4885]: I1002 01:47:52.455313 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2ll9b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52b39712-5fa1-4fe0-814b-b170a6e3938b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9a96383d0fbfd116c419054aaf1a7adba1937e8b0bad294cefd68b4ab0d2298\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c979b91ba748aa4647826a8926337662ca929c54090fb62285572819ec2fbf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c979b91ba748aa4647826a8926337662ca929c54090fb62285572819ec2fbf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97d9d693e8e60ee9c183bb44461e698e8b3cfd20a516605aa8754c9205d2bdc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://97d9d693e8e60ee9c183bb44461e698e8b3cfd20a516605aa8754c9205d2bdc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c44252676359c671d65c9588b05572d76c99cf4c90666f0ac089b2f778450a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c44252676359c671d65c9588b05572d76c99cf4c90666f0ac089b2f778450a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80c982f07874df686fe7e29cc91043eb6d58f1be5fc93cacd46d33a21dd9b622\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80c982f07874df686fe7e29cc91043eb6d58f1be5fc93cacd46d33a21dd9b622\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a354c985652c7ad7fa8944fbfdfe223cfb0f97fb11eecd9bd429020d4892b78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a354c985652c7ad7fa8944fbfdfe223cfb0f97fb11eecd9bd429020d4892b78\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2efd7283df313d086c1b8b88f289ace28764de256205bee7610d6711002d8515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2efd7283df313d086c1b8b88f289ace28764de256205bee7610d6711002d8515\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2ll9b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:52Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:52 crc kubenswrapper[4885]: I1002 01:47:52.548808 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:52 crc kubenswrapper[4885]: I1002 01:47:52.548864 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:52 crc kubenswrapper[4885]: I1002 01:47:52.548883 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:52 crc kubenswrapper[4885]: I1002 01:47:52.548908 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:52 crc kubenswrapper[4885]: I1002 01:47:52.548926 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:52Z","lastTransitionTime":"2025-10-02T01:47:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:52 crc kubenswrapper[4885]: I1002 01:47:52.652625 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:52 crc kubenswrapper[4885]: I1002 01:47:52.652729 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:52 crc kubenswrapper[4885]: I1002 01:47:52.652748 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:52 crc kubenswrapper[4885]: I1002 01:47:52.652776 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:52 crc kubenswrapper[4885]: I1002 01:47:52.652795 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:52Z","lastTransitionTime":"2025-10-02T01:47:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:52 crc kubenswrapper[4885]: I1002 01:47:52.755639 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:52 crc kubenswrapper[4885]: I1002 01:47:52.755683 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:52 crc kubenswrapper[4885]: I1002 01:47:52.755700 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:52 crc kubenswrapper[4885]: I1002 01:47:52.755722 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:52 crc kubenswrapper[4885]: I1002 01:47:52.755739 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:52Z","lastTransitionTime":"2025-10-02T01:47:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:52 crc kubenswrapper[4885]: I1002 01:47:52.858957 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:52 crc kubenswrapper[4885]: I1002 01:47:52.859021 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:52 crc kubenswrapper[4885]: I1002 01:47:52.859040 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:52 crc kubenswrapper[4885]: I1002 01:47:52.859063 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:52 crc kubenswrapper[4885]: I1002 01:47:52.859081 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:52Z","lastTransitionTime":"2025-10-02T01:47:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:52 crc kubenswrapper[4885]: I1002 01:47:52.961374 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:52 crc kubenswrapper[4885]: I1002 01:47:52.961415 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:52 crc kubenswrapper[4885]: I1002 01:47:52.961431 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:52 crc kubenswrapper[4885]: I1002 01:47:52.961452 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:52 crc kubenswrapper[4885]: I1002 01:47:52.961470 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:52Z","lastTransitionTime":"2025-10-02T01:47:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:53 crc kubenswrapper[4885]: I1002 01:47:53.064572 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:53 crc kubenswrapper[4885]: I1002 01:47:53.064620 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:53 crc kubenswrapper[4885]: I1002 01:47:53.064636 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:53 crc kubenswrapper[4885]: I1002 01:47:53.064658 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:53 crc kubenswrapper[4885]: I1002 01:47:53.064675 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:53Z","lastTransitionTime":"2025-10-02T01:47:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:53 crc kubenswrapper[4885]: I1002 01:47:53.167925 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:53 crc kubenswrapper[4885]: I1002 01:47:53.167976 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:53 crc kubenswrapper[4885]: I1002 01:47:53.167993 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:53 crc kubenswrapper[4885]: I1002 01:47:53.168017 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:53 crc kubenswrapper[4885]: I1002 01:47:53.168034 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:53Z","lastTransitionTime":"2025-10-02T01:47:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:53 crc kubenswrapper[4885]: I1002 01:47:53.271226 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:53 crc kubenswrapper[4885]: I1002 01:47:53.271306 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:53 crc kubenswrapper[4885]: I1002 01:47:53.271327 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:53 crc kubenswrapper[4885]: I1002 01:47:53.271353 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:53 crc kubenswrapper[4885]: I1002 01:47:53.271370 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:53Z","lastTransitionTime":"2025-10-02T01:47:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:53 crc kubenswrapper[4885]: I1002 01:47:53.374840 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:53 crc kubenswrapper[4885]: I1002 01:47:53.374895 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:53 crc kubenswrapper[4885]: I1002 01:47:53.374913 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:53 crc kubenswrapper[4885]: I1002 01:47:53.374934 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:53 crc kubenswrapper[4885]: I1002 01:47:53.374951 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:53Z","lastTransitionTime":"2025-10-02T01:47:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:53 crc kubenswrapper[4885]: I1002 01:47:53.478750 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:53 crc kubenswrapper[4885]: I1002 01:47:53.478816 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:53 crc kubenswrapper[4885]: I1002 01:47:53.478834 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:53 crc kubenswrapper[4885]: I1002 01:47:53.478858 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:53 crc kubenswrapper[4885]: I1002 01:47:53.478876 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:53Z","lastTransitionTime":"2025-10-02T01:47:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:53 crc kubenswrapper[4885]: I1002 01:47:53.582021 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:53 crc kubenswrapper[4885]: I1002 01:47:53.582105 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:53 crc kubenswrapper[4885]: I1002 01:47:53.582130 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:53 crc kubenswrapper[4885]: I1002 01:47:53.582166 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:53 crc kubenswrapper[4885]: I1002 01:47:53.582189 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:53Z","lastTransitionTime":"2025-10-02T01:47:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:53 crc kubenswrapper[4885]: I1002 01:47:53.685230 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:53 crc kubenswrapper[4885]: I1002 01:47:53.685323 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:53 crc kubenswrapper[4885]: I1002 01:47:53.685343 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:53 crc kubenswrapper[4885]: I1002 01:47:53.685367 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:53 crc kubenswrapper[4885]: I1002 01:47:53.685386 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:53Z","lastTransitionTime":"2025-10-02T01:47:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:53 crc kubenswrapper[4885]: I1002 01:47:53.788446 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:53 crc kubenswrapper[4885]: I1002 01:47:53.788509 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:53 crc kubenswrapper[4885]: I1002 01:47:53.788722 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:53 crc kubenswrapper[4885]: I1002 01:47:53.788749 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:53 crc kubenswrapper[4885]: I1002 01:47:53.788774 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:53Z","lastTransitionTime":"2025-10-02T01:47:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:53 crc kubenswrapper[4885]: I1002 01:47:53.892053 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:53 crc kubenswrapper[4885]: I1002 01:47:53.892112 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:53 crc kubenswrapper[4885]: I1002 01:47:53.892128 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:53 crc kubenswrapper[4885]: I1002 01:47:53.892151 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:53 crc kubenswrapper[4885]: I1002 01:47:53.892168 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:53Z","lastTransitionTime":"2025-10-02T01:47:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:53 crc kubenswrapper[4885]: I1002 01:47:53.995411 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:53 crc kubenswrapper[4885]: I1002 01:47:53.995483 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:53 crc kubenswrapper[4885]: I1002 01:47:53.995503 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:53 crc kubenswrapper[4885]: I1002 01:47:53.995532 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:53 crc kubenswrapper[4885]: I1002 01:47:53.995557 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:53Z","lastTransitionTime":"2025-10-02T01:47:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:54 crc kubenswrapper[4885]: I1002 01:47:54.046738 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:47:54 crc kubenswrapper[4885]: I1002 01:47:54.046818 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:47:54 crc kubenswrapper[4885]: E1002 01:47:54.046957 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:47:54 crc kubenswrapper[4885]: I1002 01:47:54.047041 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:47:54 crc kubenswrapper[4885]: I1002 01:47:54.047072 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sv4md" Oct 02 01:47:54 crc kubenswrapper[4885]: E1002 01:47:54.047220 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:47:54 crc kubenswrapper[4885]: E1002 01:47:54.047351 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sv4md" podUID="724a3e49-001c-45a4-a896-82c13e0ee0e7" Oct 02 01:47:54 crc kubenswrapper[4885]: E1002 01:47:54.047473 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:47:54 crc kubenswrapper[4885]: I1002 01:47:54.103340 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:54 crc kubenswrapper[4885]: I1002 01:47:54.103403 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:54 crc kubenswrapper[4885]: I1002 01:47:54.103426 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:54 crc kubenswrapper[4885]: I1002 01:47:54.103460 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:54 crc kubenswrapper[4885]: I1002 01:47:54.103483 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:54Z","lastTransitionTime":"2025-10-02T01:47:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:54 crc kubenswrapper[4885]: I1002 01:47:54.206408 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:54 crc kubenswrapper[4885]: I1002 01:47:54.206484 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:54 crc kubenswrapper[4885]: I1002 01:47:54.206505 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:54 crc kubenswrapper[4885]: I1002 01:47:54.206539 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:54 crc kubenswrapper[4885]: I1002 01:47:54.206566 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:54Z","lastTransitionTime":"2025-10-02T01:47:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:54 crc kubenswrapper[4885]: I1002 01:47:54.309545 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:54 crc kubenswrapper[4885]: I1002 01:47:54.309601 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:54 crc kubenswrapper[4885]: I1002 01:47:54.309624 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:54 crc kubenswrapper[4885]: I1002 01:47:54.309648 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:54 crc kubenswrapper[4885]: I1002 01:47:54.309667 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:54Z","lastTransitionTime":"2025-10-02T01:47:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:54 crc kubenswrapper[4885]: I1002 01:47:54.412638 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:54 crc kubenswrapper[4885]: I1002 01:47:54.412758 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:54 crc kubenswrapper[4885]: I1002 01:47:54.412780 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:54 crc kubenswrapper[4885]: I1002 01:47:54.412815 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:54 crc kubenswrapper[4885]: I1002 01:47:54.412841 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:54Z","lastTransitionTime":"2025-10-02T01:47:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:54 crc kubenswrapper[4885]: I1002 01:47:54.552773 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:54 crc kubenswrapper[4885]: I1002 01:47:54.552832 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:54 crc kubenswrapper[4885]: I1002 01:47:54.552862 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:54 crc kubenswrapper[4885]: I1002 01:47:54.552891 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:54 crc kubenswrapper[4885]: I1002 01:47:54.552916 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:54Z","lastTransitionTime":"2025-10-02T01:47:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:54 crc kubenswrapper[4885]: I1002 01:47:54.655807 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:54 crc kubenswrapper[4885]: I1002 01:47:54.655854 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:54 crc kubenswrapper[4885]: I1002 01:47:54.655873 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:54 crc kubenswrapper[4885]: I1002 01:47:54.655899 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:54 crc kubenswrapper[4885]: I1002 01:47:54.655918 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:54Z","lastTransitionTime":"2025-10-02T01:47:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:54 crc kubenswrapper[4885]: I1002 01:47:54.759800 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:54 crc kubenswrapper[4885]: I1002 01:47:54.759857 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:54 crc kubenswrapper[4885]: I1002 01:47:54.759873 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:54 crc kubenswrapper[4885]: I1002 01:47:54.759895 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:54 crc kubenswrapper[4885]: I1002 01:47:54.759912 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:54Z","lastTransitionTime":"2025-10-02T01:47:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:54 crc kubenswrapper[4885]: I1002 01:47:54.863979 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:54 crc kubenswrapper[4885]: I1002 01:47:54.864059 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:54 crc kubenswrapper[4885]: I1002 01:47:54.864082 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:54 crc kubenswrapper[4885]: I1002 01:47:54.864115 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:54 crc kubenswrapper[4885]: I1002 01:47:54.864135 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:54Z","lastTransitionTime":"2025-10-02T01:47:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:54 crc kubenswrapper[4885]: I1002 01:47:54.968674 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:54 crc kubenswrapper[4885]: I1002 01:47:54.968764 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:54 crc kubenswrapper[4885]: I1002 01:47:54.968787 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:54 crc kubenswrapper[4885]: I1002 01:47:54.968823 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:54 crc kubenswrapper[4885]: I1002 01:47:54.968851 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:54Z","lastTransitionTime":"2025-10-02T01:47:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:55 crc kubenswrapper[4885]: I1002 01:47:55.072472 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:55 crc kubenswrapper[4885]: I1002 01:47:55.072560 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:55 crc kubenswrapper[4885]: I1002 01:47:55.072582 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:55 crc kubenswrapper[4885]: I1002 01:47:55.072613 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:55 crc kubenswrapper[4885]: I1002 01:47:55.072635 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:55Z","lastTransitionTime":"2025-10-02T01:47:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:55 crc kubenswrapper[4885]: I1002 01:47:55.176248 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:55 crc kubenswrapper[4885]: I1002 01:47:55.176366 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:55 crc kubenswrapper[4885]: I1002 01:47:55.176389 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:55 crc kubenswrapper[4885]: I1002 01:47:55.176423 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:55 crc kubenswrapper[4885]: I1002 01:47:55.176445 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:55Z","lastTransitionTime":"2025-10-02T01:47:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:55 crc kubenswrapper[4885]: I1002 01:47:55.289152 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:55 crc kubenswrapper[4885]: I1002 01:47:55.289217 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:55 crc kubenswrapper[4885]: I1002 01:47:55.289234 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:55 crc kubenswrapper[4885]: I1002 01:47:55.289295 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:55 crc kubenswrapper[4885]: I1002 01:47:55.289324 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:55Z","lastTransitionTime":"2025-10-02T01:47:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:55 crc kubenswrapper[4885]: I1002 01:47:55.393336 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:55 crc kubenswrapper[4885]: I1002 01:47:55.393394 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:55 crc kubenswrapper[4885]: I1002 01:47:55.393412 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:55 crc kubenswrapper[4885]: I1002 01:47:55.393439 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:55 crc kubenswrapper[4885]: I1002 01:47:55.393460 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:55Z","lastTransitionTime":"2025-10-02T01:47:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:55 crc kubenswrapper[4885]: I1002 01:47:55.496586 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:55 crc kubenswrapper[4885]: I1002 01:47:55.496652 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:55 crc kubenswrapper[4885]: I1002 01:47:55.496669 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:55 crc kubenswrapper[4885]: I1002 01:47:55.496696 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:55 crc kubenswrapper[4885]: I1002 01:47:55.496715 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:55Z","lastTransitionTime":"2025-10-02T01:47:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:55 crc kubenswrapper[4885]: I1002 01:47:55.600089 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:55 crc kubenswrapper[4885]: I1002 01:47:55.600143 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:55 crc kubenswrapper[4885]: I1002 01:47:55.600161 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:55 crc kubenswrapper[4885]: I1002 01:47:55.600188 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:55 crc kubenswrapper[4885]: I1002 01:47:55.600207 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:55Z","lastTransitionTime":"2025-10-02T01:47:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:55 crc kubenswrapper[4885]: I1002 01:47:55.702820 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:55 crc kubenswrapper[4885]: I1002 01:47:55.702889 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:55 crc kubenswrapper[4885]: I1002 01:47:55.702904 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:55 crc kubenswrapper[4885]: I1002 01:47:55.702929 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:55 crc kubenswrapper[4885]: I1002 01:47:55.702945 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:55Z","lastTransitionTime":"2025-10-02T01:47:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:55 crc kubenswrapper[4885]: I1002 01:47:55.806715 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:55 crc kubenswrapper[4885]: I1002 01:47:55.806790 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:55 crc kubenswrapper[4885]: I1002 01:47:55.806810 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:55 crc kubenswrapper[4885]: I1002 01:47:55.806836 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:55 crc kubenswrapper[4885]: I1002 01:47:55.806858 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:55Z","lastTransitionTime":"2025-10-02T01:47:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:55 crc kubenswrapper[4885]: I1002 01:47:55.910403 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:55 crc kubenswrapper[4885]: I1002 01:47:55.910476 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:55 crc kubenswrapper[4885]: I1002 01:47:55.910496 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:55 crc kubenswrapper[4885]: I1002 01:47:55.910527 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:55 crc kubenswrapper[4885]: I1002 01:47:55.910546 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:55Z","lastTransitionTime":"2025-10-02T01:47:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:56 crc kubenswrapper[4885]: I1002 01:47:56.013200 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:56 crc kubenswrapper[4885]: I1002 01:47:56.013335 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:56 crc kubenswrapper[4885]: I1002 01:47:56.013356 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:56 crc kubenswrapper[4885]: I1002 01:47:56.013383 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:56 crc kubenswrapper[4885]: I1002 01:47:56.013399 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:56Z","lastTransitionTime":"2025-10-02T01:47:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:56 crc kubenswrapper[4885]: I1002 01:47:56.046213 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sv4md" Oct 02 01:47:56 crc kubenswrapper[4885]: I1002 01:47:56.046366 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:47:56 crc kubenswrapper[4885]: I1002 01:47:56.046447 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:47:56 crc kubenswrapper[4885]: E1002 01:47:56.046637 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sv4md" podUID="724a3e49-001c-45a4-a896-82c13e0ee0e7" Oct 02 01:47:56 crc kubenswrapper[4885]: I1002 01:47:56.046661 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:47:56 crc kubenswrapper[4885]: E1002 01:47:56.046768 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:47:56 crc kubenswrapper[4885]: E1002 01:47:56.046977 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:47:56 crc kubenswrapper[4885]: E1002 01:47:56.047158 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:47:56 crc kubenswrapper[4885]: I1002 01:47:56.116968 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:56 crc kubenswrapper[4885]: I1002 01:47:56.117016 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:56 crc kubenswrapper[4885]: I1002 01:47:56.117102 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:56 crc kubenswrapper[4885]: I1002 01:47:56.117127 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:56 crc kubenswrapper[4885]: I1002 01:47:56.117145 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:56Z","lastTransitionTime":"2025-10-02T01:47:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:56 crc kubenswrapper[4885]: I1002 01:47:56.220236 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:56 crc kubenswrapper[4885]: I1002 01:47:56.220324 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:56 crc kubenswrapper[4885]: I1002 01:47:56.220343 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:56 crc kubenswrapper[4885]: I1002 01:47:56.220365 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:56 crc kubenswrapper[4885]: I1002 01:47:56.220382 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:56Z","lastTransitionTime":"2025-10-02T01:47:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:56 crc kubenswrapper[4885]: I1002 01:47:56.324056 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:56 crc kubenswrapper[4885]: I1002 01:47:56.324159 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:56 crc kubenswrapper[4885]: I1002 01:47:56.324186 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:56 crc kubenswrapper[4885]: I1002 01:47:56.324216 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:56 crc kubenswrapper[4885]: I1002 01:47:56.324237 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:56Z","lastTransitionTime":"2025-10-02T01:47:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:56 crc kubenswrapper[4885]: I1002 01:47:56.427978 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:56 crc kubenswrapper[4885]: I1002 01:47:56.428017 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:56 crc kubenswrapper[4885]: I1002 01:47:56.428035 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:56 crc kubenswrapper[4885]: I1002 01:47:56.428057 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:56 crc kubenswrapper[4885]: I1002 01:47:56.428074 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:56Z","lastTransitionTime":"2025-10-02T01:47:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:56 crc kubenswrapper[4885]: I1002 01:47:56.531819 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:56 crc kubenswrapper[4885]: I1002 01:47:56.531879 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:56 crc kubenswrapper[4885]: I1002 01:47:56.531898 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:56 crc kubenswrapper[4885]: I1002 01:47:56.531928 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:56 crc kubenswrapper[4885]: I1002 01:47:56.531950 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:56Z","lastTransitionTime":"2025-10-02T01:47:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:56 crc kubenswrapper[4885]: I1002 01:47:56.634690 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:56 crc kubenswrapper[4885]: I1002 01:47:56.634752 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:56 crc kubenswrapper[4885]: I1002 01:47:56.634769 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:56 crc kubenswrapper[4885]: I1002 01:47:56.634796 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:56 crc kubenswrapper[4885]: I1002 01:47:56.634816 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:56Z","lastTransitionTime":"2025-10-02T01:47:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:56 crc kubenswrapper[4885]: I1002 01:47:56.737186 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:56 crc kubenswrapper[4885]: I1002 01:47:56.737247 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:56 crc kubenswrapper[4885]: I1002 01:47:56.737294 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:56 crc kubenswrapper[4885]: I1002 01:47:56.737320 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:56 crc kubenswrapper[4885]: I1002 01:47:56.737342 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:56Z","lastTransitionTime":"2025-10-02T01:47:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:56 crc kubenswrapper[4885]: I1002 01:47:56.840050 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:56 crc kubenswrapper[4885]: I1002 01:47:56.840091 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:56 crc kubenswrapper[4885]: I1002 01:47:56.840105 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:56 crc kubenswrapper[4885]: I1002 01:47:56.840128 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:56 crc kubenswrapper[4885]: I1002 01:47:56.840142 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:56Z","lastTransitionTime":"2025-10-02T01:47:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:56 crc kubenswrapper[4885]: I1002 01:47:56.943652 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:56 crc kubenswrapper[4885]: I1002 01:47:56.943736 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:56 crc kubenswrapper[4885]: I1002 01:47:56.943763 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:56 crc kubenswrapper[4885]: I1002 01:47:56.943798 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:56 crc kubenswrapper[4885]: I1002 01:47:56.943825 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:56Z","lastTransitionTime":"2025-10-02T01:47:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:57 crc kubenswrapper[4885]: I1002 01:47:57.046789 4885 scope.go:117] "RemoveContainer" containerID="188cb50b58926683ce8862d618258c1d1a6f33310608546c8651ece28f7c338f" Oct 02 01:47:57 crc kubenswrapper[4885]: I1002 01:47:57.047059 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:57 crc kubenswrapper[4885]: I1002 01:47:57.047125 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:57 crc kubenswrapper[4885]: I1002 01:47:57.047144 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:57 crc kubenswrapper[4885]: I1002 01:47:57.047170 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:57 crc kubenswrapper[4885]: I1002 01:47:57.047191 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:57Z","lastTransitionTime":"2025-10-02T01:47:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:57 crc kubenswrapper[4885]: E1002 01:47:57.047294 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-dngcm_openshift-ovn-kubernetes(9e4679dd-f870-41e9-a32b-360f5d2ee81b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" podUID="9e4679dd-f870-41e9-a32b-360f5d2ee81b" Oct 02 01:47:57 crc kubenswrapper[4885]: I1002 01:47:57.151165 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:57 crc kubenswrapper[4885]: I1002 01:47:57.151225 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:57 crc kubenswrapper[4885]: I1002 01:47:57.151244 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:57 crc kubenswrapper[4885]: I1002 01:47:57.151297 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:57 crc kubenswrapper[4885]: I1002 01:47:57.151319 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:57Z","lastTransitionTime":"2025-10-02T01:47:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:57 crc kubenswrapper[4885]: I1002 01:47:57.254659 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:57 crc kubenswrapper[4885]: I1002 01:47:57.254712 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:57 crc kubenswrapper[4885]: I1002 01:47:57.254726 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:57 crc kubenswrapper[4885]: I1002 01:47:57.254747 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:57 crc kubenswrapper[4885]: I1002 01:47:57.254761 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:57Z","lastTransitionTime":"2025-10-02T01:47:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:57 crc kubenswrapper[4885]: I1002 01:47:57.358012 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:57 crc kubenswrapper[4885]: I1002 01:47:57.358099 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:57 crc kubenswrapper[4885]: I1002 01:47:57.358113 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:57 crc kubenswrapper[4885]: I1002 01:47:57.358133 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:57 crc kubenswrapper[4885]: I1002 01:47:57.358148 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:57Z","lastTransitionTime":"2025-10-02T01:47:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:57 crc kubenswrapper[4885]: I1002 01:47:57.460471 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:57 crc kubenswrapper[4885]: I1002 01:47:57.460533 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:57 crc kubenswrapper[4885]: I1002 01:47:57.460551 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:57 crc kubenswrapper[4885]: I1002 01:47:57.460577 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:57 crc kubenswrapper[4885]: I1002 01:47:57.460596 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:57Z","lastTransitionTime":"2025-10-02T01:47:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:57 crc kubenswrapper[4885]: I1002 01:47:57.563377 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:57 crc kubenswrapper[4885]: I1002 01:47:57.563450 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:57 crc kubenswrapper[4885]: I1002 01:47:57.563470 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:57 crc kubenswrapper[4885]: I1002 01:47:57.563501 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:57 crc kubenswrapper[4885]: I1002 01:47:57.563520 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:57Z","lastTransitionTime":"2025-10-02T01:47:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:57 crc kubenswrapper[4885]: I1002 01:47:57.666568 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:57 crc kubenswrapper[4885]: I1002 01:47:57.666632 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:57 crc kubenswrapper[4885]: I1002 01:47:57.666651 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:57 crc kubenswrapper[4885]: I1002 01:47:57.666677 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:57 crc kubenswrapper[4885]: I1002 01:47:57.666697 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:57Z","lastTransitionTime":"2025-10-02T01:47:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:57 crc kubenswrapper[4885]: I1002 01:47:57.776305 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:57 crc kubenswrapper[4885]: I1002 01:47:57.776364 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:57 crc kubenswrapper[4885]: I1002 01:47:57.776383 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:57 crc kubenswrapper[4885]: I1002 01:47:57.776410 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:57 crc kubenswrapper[4885]: I1002 01:47:57.776429 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:57Z","lastTransitionTime":"2025-10-02T01:47:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:57 crc kubenswrapper[4885]: I1002 01:47:57.879360 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:57 crc kubenswrapper[4885]: I1002 01:47:57.879406 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:57 crc kubenswrapper[4885]: I1002 01:47:57.879425 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:57 crc kubenswrapper[4885]: I1002 01:47:57.879448 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:57 crc kubenswrapper[4885]: I1002 01:47:57.879466 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:57Z","lastTransitionTime":"2025-10-02T01:47:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:57 crc kubenswrapper[4885]: I1002 01:47:57.982751 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:57 crc kubenswrapper[4885]: I1002 01:47:57.982816 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:57 crc kubenswrapper[4885]: I1002 01:47:57.982836 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:57 crc kubenswrapper[4885]: I1002 01:47:57.982862 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:57 crc kubenswrapper[4885]: I1002 01:47:57.982881 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:57Z","lastTransitionTime":"2025-10-02T01:47:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:58 crc kubenswrapper[4885]: I1002 01:47:58.045667 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sv4md" Oct 02 01:47:58 crc kubenswrapper[4885]: I1002 01:47:58.045768 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:47:58 crc kubenswrapper[4885]: I1002 01:47:58.045825 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:47:58 crc kubenswrapper[4885]: E1002 01:47:58.045883 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sv4md" podUID="724a3e49-001c-45a4-a896-82c13e0ee0e7" Oct 02 01:47:58 crc kubenswrapper[4885]: E1002 01:47:58.045931 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:47:58 crc kubenswrapper[4885]: I1002 01:47:58.046038 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:47:58 crc kubenswrapper[4885]: E1002 01:47:58.046103 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:47:58 crc kubenswrapper[4885]: E1002 01:47:58.046315 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:47:58 crc kubenswrapper[4885]: I1002 01:47:58.084695 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:58 crc kubenswrapper[4885]: I1002 01:47:58.084757 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:58 crc kubenswrapper[4885]: I1002 01:47:58.084776 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:58 crc kubenswrapper[4885]: I1002 01:47:58.084801 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:58 crc kubenswrapper[4885]: I1002 01:47:58.084820 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:58Z","lastTransitionTime":"2025-10-02T01:47:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:58 crc kubenswrapper[4885]: I1002 01:47:58.187603 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:58 crc kubenswrapper[4885]: I1002 01:47:58.187654 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:58 crc kubenswrapper[4885]: I1002 01:47:58.187673 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:58 crc kubenswrapper[4885]: I1002 01:47:58.187697 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:58 crc kubenswrapper[4885]: I1002 01:47:58.187710 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:58Z","lastTransitionTime":"2025-10-02T01:47:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:58 crc kubenswrapper[4885]: I1002 01:47:58.290849 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:58 crc kubenswrapper[4885]: I1002 01:47:58.290917 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:58 crc kubenswrapper[4885]: I1002 01:47:58.290941 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:58 crc kubenswrapper[4885]: I1002 01:47:58.290972 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:58 crc kubenswrapper[4885]: I1002 01:47:58.290995 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:58Z","lastTransitionTime":"2025-10-02T01:47:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:58 crc kubenswrapper[4885]: I1002 01:47:58.393530 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:58 crc kubenswrapper[4885]: I1002 01:47:58.393615 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:58 crc kubenswrapper[4885]: I1002 01:47:58.393637 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:58 crc kubenswrapper[4885]: I1002 01:47:58.393663 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:58 crc kubenswrapper[4885]: I1002 01:47:58.393682 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:58Z","lastTransitionTime":"2025-10-02T01:47:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:58 crc kubenswrapper[4885]: I1002 01:47:58.496579 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:58 crc kubenswrapper[4885]: I1002 01:47:58.496657 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:58 crc kubenswrapper[4885]: I1002 01:47:58.496669 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:58 crc kubenswrapper[4885]: I1002 01:47:58.496691 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:58 crc kubenswrapper[4885]: I1002 01:47:58.496705 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:58Z","lastTransitionTime":"2025-10-02T01:47:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:58 crc kubenswrapper[4885]: I1002 01:47:58.599035 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:58 crc kubenswrapper[4885]: I1002 01:47:58.599101 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:58 crc kubenswrapper[4885]: I1002 01:47:58.599113 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:58 crc kubenswrapper[4885]: I1002 01:47:58.599130 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:58 crc kubenswrapper[4885]: I1002 01:47:58.599143 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:58Z","lastTransitionTime":"2025-10-02T01:47:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:58 crc kubenswrapper[4885]: I1002 01:47:58.701766 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:58 crc kubenswrapper[4885]: I1002 01:47:58.701806 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:58 crc kubenswrapper[4885]: I1002 01:47:58.701814 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:58 crc kubenswrapper[4885]: I1002 01:47:58.701829 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:58 crc kubenswrapper[4885]: I1002 01:47:58.701843 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:58Z","lastTransitionTime":"2025-10-02T01:47:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:58 crc kubenswrapper[4885]: I1002 01:47:58.804537 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:58 crc kubenswrapper[4885]: I1002 01:47:58.804590 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:58 crc kubenswrapper[4885]: I1002 01:47:58.804603 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:58 crc kubenswrapper[4885]: I1002 01:47:58.804623 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:58 crc kubenswrapper[4885]: I1002 01:47:58.804637 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:58Z","lastTransitionTime":"2025-10-02T01:47:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:58 crc kubenswrapper[4885]: I1002 01:47:58.907129 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:58 crc kubenswrapper[4885]: I1002 01:47:58.907169 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:58 crc kubenswrapper[4885]: I1002 01:47:58.907179 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:58 crc kubenswrapper[4885]: I1002 01:47:58.907197 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:58 crc kubenswrapper[4885]: I1002 01:47:58.907211 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:58Z","lastTransitionTime":"2025-10-02T01:47:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:59 crc kubenswrapper[4885]: I1002 01:47:59.009535 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:59 crc kubenswrapper[4885]: I1002 01:47:59.009586 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:59 crc kubenswrapper[4885]: I1002 01:47:59.009597 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:59 crc kubenswrapper[4885]: I1002 01:47:59.009612 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:59 crc kubenswrapper[4885]: I1002 01:47:59.009626 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:59Z","lastTransitionTime":"2025-10-02T01:47:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:59 crc kubenswrapper[4885]: I1002 01:47:59.112604 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:59 crc kubenswrapper[4885]: I1002 01:47:59.112668 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:59 crc kubenswrapper[4885]: I1002 01:47:59.112688 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:59 crc kubenswrapper[4885]: I1002 01:47:59.112716 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:59 crc kubenswrapper[4885]: I1002 01:47:59.112736 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:59Z","lastTransitionTime":"2025-10-02T01:47:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:59 crc kubenswrapper[4885]: I1002 01:47:59.216668 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:59 crc kubenswrapper[4885]: I1002 01:47:59.216805 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:59 crc kubenswrapper[4885]: I1002 01:47:59.216832 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:59 crc kubenswrapper[4885]: I1002 01:47:59.216872 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:59 crc kubenswrapper[4885]: I1002 01:47:59.216897 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:59Z","lastTransitionTime":"2025-10-02T01:47:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:59 crc kubenswrapper[4885]: I1002 01:47:59.320240 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:59 crc kubenswrapper[4885]: I1002 01:47:59.320322 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:59 crc kubenswrapper[4885]: I1002 01:47:59.320336 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:59 crc kubenswrapper[4885]: I1002 01:47:59.320354 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:59 crc kubenswrapper[4885]: I1002 01:47:59.320367 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:59Z","lastTransitionTime":"2025-10-02T01:47:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:59 crc kubenswrapper[4885]: I1002 01:47:59.417324 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:59 crc kubenswrapper[4885]: I1002 01:47:59.417394 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:59 crc kubenswrapper[4885]: I1002 01:47:59.417417 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:59 crc kubenswrapper[4885]: I1002 01:47:59.417450 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:59 crc kubenswrapper[4885]: I1002 01:47:59.417472 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:59Z","lastTransitionTime":"2025-10-02T01:47:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:59 crc kubenswrapper[4885]: E1002 01:47:59.432462 4885 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:47:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:47:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:47:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:47:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"92603ae5-ba38-4201-98e9-979fd56cb50d\\\",\\\"systemUUID\\\":\\\"50ddf2c2-ea95-4960-ae48-d17e9484ee2f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:59Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:59 crc kubenswrapper[4885]: I1002 01:47:59.438999 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:59 crc kubenswrapper[4885]: I1002 01:47:59.439115 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:59 crc kubenswrapper[4885]: I1002 01:47:59.439192 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:59 crc kubenswrapper[4885]: I1002 01:47:59.439293 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:59 crc kubenswrapper[4885]: I1002 01:47:59.439343 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:59Z","lastTransitionTime":"2025-10-02T01:47:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:59 crc kubenswrapper[4885]: E1002 01:47:59.461839 4885 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:47:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:47:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:47:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:47:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"92603ae5-ba38-4201-98e9-979fd56cb50d\\\",\\\"systemUUID\\\":\\\"50ddf2c2-ea95-4960-ae48-d17e9484ee2f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:59Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:59 crc kubenswrapper[4885]: I1002 01:47:59.468426 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:59 crc kubenswrapper[4885]: I1002 01:47:59.468502 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:59 crc kubenswrapper[4885]: I1002 01:47:59.468522 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:59 crc kubenswrapper[4885]: I1002 01:47:59.468571 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:59 crc kubenswrapper[4885]: I1002 01:47:59.468590 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:59Z","lastTransitionTime":"2025-10-02T01:47:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:59 crc kubenswrapper[4885]: E1002 01:47:59.505944 4885 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:47:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:47:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:47:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:47:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"92603ae5-ba38-4201-98e9-979fd56cb50d\\\",\\\"systemUUID\\\":\\\"50ddf2c2-ea95-4960-ae48-d17e9484ee2f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:59Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:59 crc kubenswrapper[4885]: I1002 01:47:59.513206 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:59 crc kubenswrapper[4885]: I1002 01:47:59.513289 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:59 crc kubenswrapper[4885]: I1002 01:47:59.513310 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:59 crc kubenswrapper[4885]: I1002 01:47:59.513337 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:59 crc kubenswrapper[4885]: I1002 01:47:59.513357 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:59Z","lastTransitionTime":"2025-10-02T01:47:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:59 crc kubenswrapper[4885]: E1002 01:47:59.542417 4885 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:47:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:47:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:47:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:47:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"92603ae5-ba38-4201-98e9-979fd56cb50d\\\",\\\"systemUUID\\\":\\\"50ddf2c2-ea95-4960-ae48-d17e9484ee2f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:59Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:59 crc kubenswrapper[4885]: I1002 01:47:59.547410 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:59 crc kubenswrapper[4885]: I1002 01:47:59.547444 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:59 crc kubenswrapper[4885]: I1002 01:47:59.547456 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:59 crc kubenswrapper[4885]: I1002 01:47:59.547475 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:59 crc kubenswrapper[4885]: I1002 01:47:59.547489 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:59Z","lastTransitionTime":"2025-10-02T01:47:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:59 crc kubenswrapper[4885]: E1002 01:47:59.574599 4885 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:47:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:47:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:47:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:47:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"92603ae5-ba38-4201-98e9-979fd56cb50d\\\",\\\"systemUUID\\\":\\\"50ddf2c2-ea95-4960-ae48-d17e9484ee2f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:47:59Z is after 2025-08-24T17:21:41Z" Oct 02 01:47:59 crc kubenswrapper[4885]: E1002 01:47:59.574719 4885 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 02 01:47:59 crc kubenswrapper[4885]: I1002 01:47:59.577064 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:59 crc kubenswrapper[4885]: I1002 01:47:59.577109 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:59 crc kubenswrapper[4885]: I1002 01:47:59.577122 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:59 crc kubenswrapper[4885]: I1002 01:47:59.577140 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:59 crc kubenswrapper[4885]: I1002 01:47:59.577153 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:59Z","lastTransitionTime":"2025-10-02T01:47:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:59 crc kubenswrapper[4885]: I1002 01:47:59.679688 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:59 crc kubenswrapper[4885]: I1002 01:47:59.679728 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:59 crc kubenswrapper[4885]: I1002 01:47:59.679747 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:59 crc kubenswrapper[4885]: I1002 01:47:59.679770 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:59 crc kubenswrapper[4885]: I1002 01:47:59.679788 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:59Z","lastTransitionTime":"2025-10-02T01:47:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:59 crc kubenswrapper[4885]: I1002 01:47:59.782063 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:59 crc kubenswrapper[4885]: I1002 01:47:59.782124 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:59 crc kubenswrapper[4885]: I1002 01:47:59.782142 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:59 crc kubenswrapper[4885]: I1002 01:47:59.782169 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:59 crc kubenswrapper[4885]: I1002 01:47:59.782190 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:59Z","lastTransitionTime":"2025-10-02T01:47:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:59 crc kubenswrapper[4885]: I1002 01:47:59.884817 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:59 crc kubenswrapper[4885]: I1002 01:47:59.884872 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:59 crc kubenswrapper[4885]: I1002 01:47:59.884889 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:59 crc kubenswrapper[4885]: I1002 01:47:59.884918 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:59 crc kubenswrapper[4885]: I1002 01:47:59.884936 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:59Z","lastTransitionTime":"2025-10-02T01:47:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:47:59 crc kubenswrapper[4885]: I1002 01:47:59.987238 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:47:59 crc kubenswrapper[4885]: I1002 01:47:59.987335 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:47:59 crc kubenswrapper[4885]: I1002 01:47:59.987396 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:47:59 crc kubenswrapper[4885]: I1002 01:47:59.987427 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:47:59 crc kubenswrapper[4885]: I1002 01:47:59.987447 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:47:59Z","lastTransitionTime":"2025-10-02T01:47:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:00 crc kubenswrapper[4885]: I1002 01:48:00.046242 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:48:00 crc kubenswrapper[4885]: E1002 01:48:00.046519 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:48:00 crc kubenswrapper[4885]: I1002 01:48:00.046577 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sv4md" Oct 02 01:48:00 crc kubenswrapper[4885]: I1002 01:48:00.046541 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:48:00 crc kubenswrapper[4885]: E1002 01:48:00.046727 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sv4md" podUID="724a3e49-001c-45a4-a896-82c13e0ee0e7" Oct 02 01:48:00 crc kubenswrapper[4885]: I1002 01:48:00.046550 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:48:00 crc kubenswrapper[4885]: E1002 01:48:00.046878 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:48:00 crc kubenswrapper[4885]: E1002 01:48:00.047007 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:48:00 crc kubenswrapper[4885]: I1002 01:48:00.089613 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:00 crc kubenswrapper[4885]: I1002 01:48:00.089654 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:00 crc kubenswrapper[4885]: I1002 01:48:00.089665 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:00 crc kubenswrapper[4885]: I1002 01:48:00.089678 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:00 crc kubenswrapper[4885]: I1002 01:48:00.089690 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:00Z","lastTransitionTime":"2025-10-02T01:48:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:00 crc kubenswrapper[4885]: I1002 01:48:00.171686 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/724a3e49-001c-45a4-a896-82c13e0ee0e7-metrics-certs\") pod \"network-metrics-daemon-sv4md\" (UID: \"724a3e49-001c-45a4-a896-82c13e0ee0e7\") " pod="openshift-multus/network-metrics-daemon-sv4md" Oct 02 01:48:00 crc kubenswrapper[4885]: E1002 01:48:00.171825 4885 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 01:48:00 crc kubenswrapper[4885]: E1002 01:48:00.171891 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/724a3e49-001c-45a4-a896-82c13e0ee0e7-metrics-certs podName:724a3e49-001c-45a4-a896-82c13e0ee0e7 nodeName:}" failed. No retries permitted until 2025-10-02 01:48:32.171872187 +0000 UTC m=+100.983619586 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/724a3e49-001c-45a4-a896-82c13e0ee0e7-metrics-certs") pod "network-metrics-daemon-sv4md" (UID: "724a3e49-001c-45a4-a896-82c13e0ee0e7") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 01:48:00 crc kubenswrapper[4885]: I1002 01:48:00.192764 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:00 crc kubenswrapper[4885]: I1002 01:48:00.192816 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:00 crc kubenswrapper[4885]: I1002 01:48:00.192837 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:00 crc kubenswrapper[4885]: I1002 01:48:00.192865 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:00 crc kubenswrapper[4885]: I1002 01:48:00.192883 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:00Z","lastTransitionTime":"2025-10-02T01:48:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:00 crc kubenswrapper[4885]: I1002 01:48:00.296594 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:00 crc kubenswrapper[4885]: I1002 01:48:00.296660 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:00 crc kubenswrapper[4885]: I1002 01:48:00.296679 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:00 crc kubenswrapper[4885]: I1002 01:48:00.296708 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:00 crc kubenswrapper[4885]: I1002 01:48:00.296734 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:00Z","lastTransitionTime":"2025-10-02T01:48:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:00 crc kubenswrapper[4885]: I1002 01:48:00.400307 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:00 crc kubenswrapper[4885]: I1002 01:48:00.400406 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:00 crc kubenswrapper[4885]: I1002 01:48:00.400427 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:00 crc kubenswrapper[4885]: I1002 01:48:00.400526 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:00 crc kubenswrapper[4885]: I1002 01:48:00.400550 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:00Z","lastTransitionTime":"2025-10-02T01:48:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:00 crc kubenswrapper[4885]: I1002 01:48:00.503559 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:00 crc kubenswrapper[4885]: I1002 01:48:00.503603 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:00 crc kubenswrapper[4885]: I1002 01:48:00.503617 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:00 crc kubenswrapper[4885]: I1002 01:48:00.503637 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:00 crc kubenswrapper[4885]: I1002 01:48:00.503648 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:00Z","lastTransitionTime":"2025-10-02T01:48:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:00 crc kubenswrapper[4885]: I1002 01:48:00.606156 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:00 crc kubenswrapper[4885]: I1002 01:48:00.606241 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:00 crc kubenswrapper[4885]: I1002 01:48:00.606290 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:00 crc kubenswrapper[4885]: I1002 01:48:00.606324 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:00 crc kubenswrapper[4885]: I1002 01:48:00.606346 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:00Z","lastTransitionTime":"2025-10-02T01:48:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:00 crc kubenswrapper[4885]: I1002 01:48:00.709432 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:00 crc kubenswrapper[4885]: I1002 01:48:00.709500 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:00 crc kubenswrapper[4885]: I1002 01:48:00.709524 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:00 crc kubenswrapper[4885]: I1002 01:48:00.709552 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:00 crc kubenswrapper[4885]: I1002 01:48:00.709574 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:00Z","lastTransitionTime":"2025-10-02T01:48:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:00 crc kubenswrapper[4885]: I1002 01:48:00.812931 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:00 crc kubenswrapper[4885]: I1002 01:48:00.813001 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:00 crc kubenswrapper[4885]: I1002 01:48:00.813018 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:00 crc kubenswrapper[4885]: I1002 01:48:00.813044 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:00 crc kubenswrapper[4885]: I1002 01:48:00.813115 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:00Z","lastTransitionTime":"2025-10-02T01:48:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:00 crc kubenswrapper[4885]: I1002 01:48:00.915479 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:00 crc kubenswrapper[4885]: I1002 01:48:00.915563 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:00 crc kubenswrapper[4885]: I1002 01:48:00.915587 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:00 crc kubenswrapper[4885]: I1002 01:48:00.916106 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:00 crc kubenswrapper[4885]: I1002 01:48:00.916438 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:00Z","lastTransitionTime":"2025-10-02T01:48:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:01 crc kubenswrapper[4885]: I1002 01:48:01.018901 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:01 crc kubenswrapper[4885]: I1002 01:48:01.018941 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:01 crc kubenswrapper[4885]: I1002 01:48:01.018962 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:01 crc kubenswrapper[4885]: I1002 01:48:01.018987 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:01 crc kubenswrapper[4885]: I1002 01:48:01.019009 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:01Z","lastTransitionTime":"2025-10-02T01:48:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:01 crc kubenswrapper[4885]: I1002 01:48:01.121663 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:01 crc kubenswrapper[4885]: I1002 01:48:01.121727 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:01 crc kubenswrapper[4885]: I1002 01:48:01.121740 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:01 crc kubenswrapper[4885]: I1002 01:48:01.121757 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:01 crc kubenswrapper[4885]: I1002 01:48:01.121770 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:01Z","lastTransitionTime":"2025-10-02T01:48:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:01 crc kubenswrapper[4885]: I1002 01:48:01.224586 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:01 crc kubenswrapper[4885]: I1002 01:48:01.224631 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:01 crc kubenswrapper[4885]: I1002 01:48:01.224639 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:01 crc kubenswrapper[4885]: I1002 01:48:01.224650 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:01 crc kubenswrapper[4885]: I1002 01:48:01.224658 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:01Z","lastTransitionTime":"2025-10-02T01:48:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:01 crc kubenswrapper[4885]: I1002 01:48:01.326700 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:01 crc kubenswrapper[4885]: I1002 01:48:01.326761 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:01 crc kubenswrapper[4885]: I1002 01:48:01.326778 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:01 crc kubenswrapper[4885]: I1002 01:48:01.326802 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:01 crc kubenswrapper[4885]: I1002 01:48:01.326820 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:01Z","lastTransitionTime":"2025-10-02T01:48:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:01 crc kubenswrapper[4885]: I1002 01:48:01.428403 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:01 crc kubenswrapper[4885]: I1002 01:48:01.428432 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:01 crc kubenswrapper[4885]: I1002 01:48:01.428440 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:01 crc kubenswrapper[4885]: I1002 01:48:01.428454 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:01 crc kubenswrapper[4885]: I1002 01:48:01.428463 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:01Z","lastTransitionTime":"2025-10-02T01:48:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:01 crc kubenswrapper[4885]: I1002 01:48:01.530433 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:01 crc kubenswrapper[4885]: I1002 01:48:01.530481 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:01 crc kubenswrapper[4885]: I1002 01:48:01.530500 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:01 crc kubenswrapper[4885]: I1002 01:48:01.530519 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:01 crc kubenswrapper[4885]: I1002 01:48:01.530535 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:01Z","lastTransitionTime":"2025-10-02T01:48:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:01 crc kubenswrapper[4885]: I1002 01:48:01.586034 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-vhgbd_ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f/kube-multus/0.log" Oct 02 01:48:01 crc kubenswrapper[4885]: I1002 01:48:01.586105 4885 generic.go:334] "Generic (PLEG): container finished" podID="ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f" containerID="c63a7d20d39b65acfb5a9be4c28b21c72111b398bea76c238e15ddb4dd63e139" exitCode=1 Oct 02 01:48:01 crc kubenswrapper[4885]: I1002 01:48:01.586139 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-vhgbd" event={"ID":"ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f","Type":"ContainerDied","Data":"c63a7d20d39b65acfb5a9be4c28b21c72111b398bea76c238e15ddb4dd63e139"} Oct 02 01:48:01 crc kubenswrapper[4885]: I1002 01:48:01.586637 4885 scope.go:117] "RemoveContainer" containerID="c63a7d20d39b65acfb5a9be4c28b21c72111b398bea76c238e15ddb4dd63e139" Oct 02 01:48:01 crc kubenswrapper[4885]: I1002 01:48:01.600536 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-sv4md" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"724a3e49-001c-45a4-a896-82c13e0ee0e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hpqgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hpqgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-sv4md\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:01Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:01 crc kubenswrapper[4885]: I1002 01:48:01.617012 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"506a599f-fb1c-4d14-be88-0e4833f3d624\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31c2f0193ffef54965b79a3569b0f6f4547e540bac0774155159a1c996c196c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://533e73e5f90b7258acbcdf018ee01323c118f9dc45035b25b065ba01cabecccf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c00db833a593ec40eb63a6b383455e19e960c2e98b9b66f76877ee993057e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21e2f462a0deb0d740185663b2014f4a1e57b4fdbeb82571264b002f928035bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:01Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:01 crc kubenswrapper[4885]: I1002 01:48:01.637238 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:01 crc kubenswrapper[4885]: I1002 01:48:01.637333 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:01 crc kubenswrapper[4885]: I1002 01:48:01.637357 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:01 crc kubenswrapper[4885]: I1002 01:48:01.637435 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:01 crc kubenswrapper[4885]: I1002 01:48:01.637469 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:01Z","lastTransitionTime":"2025-10-02T01:48:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:01 crc kubenswrapper[4885]: I1002 01:48:01.638847 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e4679dd-f870-41e9-a32b-360f5d2ee81b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4310909ad9c7b609ed6814e04a525a6afae747775598e7275080413dec8b23a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7642033b0b5995d332ccf7992d6d7e6a5e9c9369ae2d1071ccb2f536938c8f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9f99f9e26a04fb42b01a0783a5c4e360514818ce9102edc7fa08626d50b9a34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6fa72b030d309656b73121e4ac2a3924a4e449aad66683dd9de5ee5f22d5a4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef403e519d197fb37b3e6b76a14a5128ba3f30a7a5e8bf0a420cf6a39b89575f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8482ba457683ffa03100ec1f07ea9d3f918be5e490b5ba05c2ba965d03033191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://188cb50b58926683ce8862d618258c1d1a6f33310608546c8651ece28f7c338f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://188cb50b58926683ce8862d618258c1d1a6f33310608546c8651ece28f7c338f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T01:47:41Z\\\",\\\"message\\\":\\\"921 6578 obj_retry.go:434] periodicallyRetryResources: Retry channel got triggered: retrying failed objects of type *v1.Pod\\\\nI1002 01:47:40.991945 6578 obj_retry.go:409] Going to retry *v1.Pod resource setup for 16 objects: [openshift-multus/network-metrics-daemon-sv4md openshift-network-operator/iptables-alerter-4ln5h openshift-kube-apiserver/kube-apiserver-crc openshift-ovn-kubernetes/ovnkube-node-dngcm openshift-dns/node-resolver-58lhj openshift-machine-config-operator/machine-config-daemon-rttx8 openshift-multus/multus-vhgbd openshift-network-console/networking-console-plugin-85b44fc459-gdk6g openshift-network-diagnostics/network-check-source-55646444c4-trplf openshift-network-operator/network-operator-58b4c7f79c-55gtf openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rrn8b openshift-image-registry/node-ca-jgdvt openshift-kube-controller-manager/kube-controller-manager-crc openshift-multus/multus-additional-cni-plugins-2ll9b openshift-network-diagnostics/network-check-target-xd92c openshift-network-node-identity/network-node-identity-vrzqb]\\\\nF1002 01:47:40.991998 6578 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller ini\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:40Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-dngcm_openshift-ovn-kubernetes(9e4679dd-f870-41e9-a32b-360f5d2ee81b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fafdd1094bd02786d745c984989852b0c690db38655f92b419733a2004f417c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0125e907d061a97286026ee823c4eed303734ed03ab4864f742592b23a1aa2f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0125e907d061a97286026ee823c4eed303734ed03ab4864f742592b23a1aa2f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dngcm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:01Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:01 crc kubenswrapper[4885]: I1002 01:48:01.656571 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5c1f95f188569c9aa47e51f2916939fa0b65a5cdf4c16b798612d41f3c8411a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:01Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:01 crc kubenswrapper[4885]: I1002 01:48:01.677512 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vhgbd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:48:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:48:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c63a7d20d39b65acfb5a9be4c28b21c72111b398bea76c238e15ddb4dd63e139\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c63a7d20d39b65acfb5a9be4c28b21c72111b398bea76c238e15ddb4dd63e139\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T01:48:00Z\\\",\\\"message\\\":\\\"2025-10-02T01:47:15+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_9f42b0ef-8651-49db-8a31-d21c0c1c47e7\\\\n2025-10-02T01:47:15+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_9f42b0ef-8651-49db-8a31-d21c0c1c47e7 to /host/opt/cni/bin/\\\\n2025-10-02T01:47:15Z [verbose] multus-daemon started\\\\n2025-10-02T01:47:15Z [verbose] Readiness Indicator file check\\\\n2025-10-02T01:48:00Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7zwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vhgbd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:01Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:01 crc kubenswrapper[4885]: I1002 01:48:01.694905 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rrn8b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52967a77-525c-4537-9b1f-d7bb754a8494\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3dee451ebe9006bf1c519fed4c65267570b6139d76841858352e087ae78d26a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtr74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e9412788d6df7332592beac48281822b26e1aa19b39addce8e1be556af9d982\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtr74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:26Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rrn8b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:01Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:01 crc kubenswrapper[4885]: I1002 01:48:01.705207 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18466979-7c7c-4d73-86c6-b8b6eaf20484\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b742e4b288bb50569d2fd23f3f9857d649a46063998170eee582ecbbba64d190\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://298c2ace58c4a9d09cfdc864316317e370bb0ac5735d631af6d85c6298f686d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://620f54e02c463f6e603808e5e57e9e5770d125578cc258ba16117a2416e57d3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b4cd17b64e5657ae05e71f3ca2340cb0d7965b1b49d1f7b4bfe4494292293ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b4cd17b64e5657ae05e71f3ca2340cb0d7965b1b49d1f7b4bfe4494292293ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:01Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:01 crc kubenswrapper[4885]: I1002 01:48:01.734682 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8070bc1c-e3de-4e9a-985e-40d1a9858056\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3e634b3727c517cc48b17c2cfb38f26460f53d1bf25d631d08ec9acb8d780e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4adcc537c23697a324d82af94c31059dd444245efdf7c978db58ede256e112ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edcd1d99644a778dfb6d0a17d2a4abd906400fe8209e611ebf8d7a9abe7ccf28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://97cb59adc2ddfa365ebc9dec649600dc90466ffe59134a6d15a9d3134a7c9b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d2808bc658ea80d68469ed16422d56ef19447f4c90b4af7b2597c64406f8a44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://83acbc20919f7c1958b6d492fadb203855df02761662425df9f91a0757fe2bf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83acbc20919f7c1958b6d492fadb203855df02761662425df9f91a0757fe2bf7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e86ecd03835efc9a253b6804bc61e6ac690e75b8d1f3cc77f5647730e586574\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e86ecd03835efc9a253b6804bc61e6ac690e75b8d1f3cc77f5647730e586574\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://48f95af51feaf014a18496bdfadd59d7f1e11167c5c8b677fa7037e48368ff2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48f95af51feaf014a18496bdfadd59d7f1e11167c5c8b677fa7037e48368ff2a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:01Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:01 crc kubenswrapper[4885]: I1002 01:48:01.741212 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:01 crc kubenswrapper[4885]: I1002 01:48:01.741297 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:01 crc kubenswrapper[4885]: I1002 01:48:01.741317 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:01 crc kubenswrapper[4885]: I1002 01:48:01.741339 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:01 crc kubenswrapper[4885]: I1002 01:48:01.741356 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:01Z","lastTransitionTime":"2025-10-02T01:48:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:01 crc kubenswrapper[4885]: I1002 01:48:01.757154 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2ll9b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52b39712-5fa1-4fe0-814b-b170a6e3938b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9a96383d0fbfd116c419054aaf1a7adba1937e8b0bad294cefd68b4ab0d2298\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c979b91ba748aa4647826a8926337662ca929c54090fb62285572819ec2fbf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c979b91ba748aa4647826a8926337662ca929c54090fb62285572819ec2fbf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97d9d693e8e60ee9c183bb44461e698e8b3cfd20a516605aa8754c9205d2bdc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://97d9d693e8e60ee9c183bb44461e698e8b3cfd20a516605aa8754c9205d2bdc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c44252676359c671d65c9588b05572d76c99cf4c90666f0ac089b2f778450a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c44252676359c671d65c9588b05572d76c99cf4c90666f0ac089b2f778450a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80c982f07874df686fe7e29cc91043eb6d58f1be5fc93cacd46d33a21dd9b622\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80c982f07874df686fe7e29cc91043eb6d58f1be5fc93cacd46d33a21dd9b622\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a354c985652c7ad7fa8944fbfdfe223cfb0f97fb11eecd9bd429020d4892b78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a354c985652c7ad7fa8944fbfdfe223cfb0f97fb11eecd9bd429020d4892b78\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2efd7283df313d086c1b8b88f289ace28764de256205bee7610d6711002d8515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2efd7283df313d086c1b8b88f289ace28764de256205bee7610d6711002d8515\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2ll9b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:01Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:01 crc kubenswrapper[4885]: I1002 01:48:01.776092 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:01Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:01 crc kubenswrapper[4885]: I1002 01:48:01.790715 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74adcf912ccedb6b5d4f2803fd215e3e2ef62367597aa6da528354c4511ee536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7403ef968d040ac09d722b7557438bef23892d578992826f42b700b370d495d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:01Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:01 crc kubenswrapper[4885]: I1002 01:48:01.807057 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b087880e05039816dcab154643ac72c7e4a0a1124817c51dc2c639b62bbee0fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:01Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:01 crc kubenswrapper[4885]: I1002 01:48:01.822015 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:01Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:01 crc kubenswrapper[4885]: I1002 01:48:01.838040 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:01Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:01 crc kubenswrapper[4885]: I1002 01:48:01.843625 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:01 crc kubenswrapper[4885]: I1002 01:48:01.843675 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:01 crc kubenswrapper[4885]: I1002 01:48:01.843688 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:01 crc kubenswrapper[4885]: I1002 01:48:01.843709 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:01 crc kubenswrapper[4885]: I1002 01:48:01.843722 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:01Z","lastTransitionTime":"2025-10-02T01:48:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:01 crc kubenswrapper[4885]: I1002 01:48:01.852578 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-58lhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac6a19c4-f259-4dbd-a016-d6c2c849c6cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed76c0f968c063f7795d363af194a222487b647ceda234de0dabb39b25eafe0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hzc4c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-58lhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:01Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:01 crc kubenswrapper[4885]: I1002 01:48:01.867444 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d169c8ba-000a-4218-8e3b-4ae53035b110\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61e436b7876107ce5ea9e668c0bbc367be6e663fc804fa3a65b51adf8be8d51f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-77rzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09536e182a46e77318c10812d1777e551f834907c006cab5c726dc66f5bce88a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-77rzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rttx8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:01Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:01 crc kubenswrapper[4885]: I1002 01:48:01.880980 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jgdvt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81733e8d-67b0-46fa-8a41-08b40645bd3e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcef60c816ca1ba9a298da658e84c4d717bea3d352c4c2f807ebd8e56327b384\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cwbl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jgdvt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:01Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:01 crc kubenswrapper[4885]: I1002 01:48:01.902791 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a0bb885-2eb0-419c-8a7a-ef634cb3c144\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a89b49cbe08d927629564e607de13655ddf3d01ac4eb5d688b325210306c120c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06c5942805a498d9621a906088aa54538b4eed583d1b8d0e3ba2a6718378a36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b9f6f7fe44d5771a838d0b89c5bdea3ebb79023ae2675ee907ec2766b3d35a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://178b4340ee6904c60c814b99d83ee225161aeb23e438d8d02f856579a13ec4e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bb63bf30d6ecce9e71f0db0c16632bdfa6b33924b673378eb0c38ab229fa837\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T01:47:06Z\\\",\\\"message\\\":\\\"W1002 01:46:55.406027 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1002 01:46:55.406714 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759369615 cert, and key in /tmp/serving-cert-3795040473/serving-signer.crt, /tmp/serving-cert-3795040473/serving-signer.key\\\\nI1002 01:46:55.614670 1 observer_polling.go:159] Starting file observer\\\\nW1002 01:46:55.617662 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1002 01:46:55.618112 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 01:46:55.620347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3795040473/tls.crt::/tmp/serving-cert-3795040473/tls.key\\\\\\\"\\\\nF1002 01:47:06.224459 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90535f97ab2ed4626d47540ebd7baada837fd332140323bec0f49aa38d3c0009\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://592f89a326c9c1559193a71904b47ca27b1dbe1a7b30e03bc48924bb2f35f921\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://592f89a326c9c1559193a71904b47ca27b1dbe1a7b30e03bc48924bb2f35f921\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:01Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:01 crc kubenswrapper[4885]: I1002 01:48:01.945616 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:01 crc kubenswrapper[4885]: I1002 01:48:01.945805 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:01 crc kubenswrapper[4885]: I1002 01:48:01.945963 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:01 crc kubenswrapper[4885]: I1002 01:48:01.946118 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:01 crc kubenswrapper[4885]: I1002 01:48:01.946255 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:01Z","lastTransitionTime":"2025-10-02T01:48:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:02 crc kubenswrapper[4885]: I1002 01:48:02.046345 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:48:02 crc kubenswrapper[4885]: I1002 01:48:02.046455 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:48:02 crc kubenswrapper[4885]: I1002 01:48:02.046568 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:48:02 crc kubenswrapper[4885]: E1002 01:48:02.046562 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:48:02 crc kubenswrapper[4885]: E1002 01:48:02.046660 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:48:02 crc kubenswrapper[4885]: E1002 01:48:02.046729 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:48:02 crc kubenswrapper[4885]: I1002 01:48:02.047007 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sv4md" Oct 02 01:48:02 crc kubenswrapper[4885]: E1002 01:48:02.047367 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sv4md" podUID="724a3e49-001c-45a4-a896-82c13e0ee0e7" Oct 02 01:48:02 crc kubenswrapper[4885]: I1002 01:48:02.047897 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:02 crc kubenswrapper[4885]: I1002 01:48:02.047939 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:02 crc kubenswrapper[4885]: I1002 01:48:02.047949 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:02 crc kubenswrapper[4885]: I1002 01:48:02.047963 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:02 crc kubenswrapper[4885]: I1002 01:48:02.047972 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:02Z","lastTransitionTime":"2025-10-02T01:48:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:02 crc kubenswrapper[4885]: I1002 01:48:02.065884 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"506a599f-fb1c-4d14-be88-0e4833f3d624\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31c2f0193ffef54965b79a3569b0f6f4547e540bac0774155159a1c996c196c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://533e73e5f90b7258acbcdf018ee01323c118f9dc45035b25b065ba01cabecccf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c00db833a593ec40eb63a6b383455e19e960c2e98b9b66f76877ee993057e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21e2f462a0deb0d740185663b2014f4a1e57b4fdbeb82571264b002f928035bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:02Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:02 crc kubenswrapper[4885]: I1002 01:48:02.094580 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e4679dd-f870-41e9-a32b-360f5d2ee81b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4310909ad9c7b609ed6814e04a525a6afae747775598e7275080413dec8b23a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7642033b0b5995d332ccf7992d6d7e6a5e9c9369ae2d1071ccb2f536938c8f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9f99f9e26a04fb42b01a0783a5c4e360514818ce9102edc7fa08626d50b9a34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6fa72b030d309656b73121e4ac2a3924a4e449aad66683dd9de5ee5f22d5a4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef403e519d197fb37b3e6b76a14a5128ba3f30a7a5e8bf0a420cf6a39b89575f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8482ba457683ffa03100ec1f07ea9d3f918be5e490b5ba05c2ba965d03033191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://188cb50b58926683ce8862d618258c1d1a6f33310608546c8651ece28f7c338f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://188cb50b58926683ce8862d618258c1d1a6f33310608546c8651ece28f7c338f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T01:47:41Z\\\",\\\"message\\\":\\\"921 6578 obj_retry.go:434] periodicallyRetryResources: Retry channel got triggered: retrying failed objects of type *v1.Pod\\\\nI1002 01:47:40.991945 6578 obj_retry.go:409] Going to retry *v1.Pod resource setup for 16 objects: [openshift-multus/network-metrics-daemon-sv4md openshift-network-operator/iptables-alerter-4ln5h openshift-kube-apiserver/kube-apiserver-crc openshift-ovn-kubernetes/ovnkube-node-dngcm openshift-dns/node-resolver-58lhj openshift-machine-config-operator/machine-config-daemon-rttx8 openshift-multus/multus-vhgbd openshift-network-console/networking-console-plugin-85b44fc459-gdk6g openshift-network-diagnostics/network-check-source-55646444c4-trplf openshift-network-operator/network-operator-58b4c7f79c-55gtf openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rrn8b openshift-image-registry/node-ca-jgdvt openshift-kube-controller-manager/kube-controller-manager-crc openshift-multus/multus-additional-cni-plugins-2ll9b openshift-network-diagnostics/network-check-target-xd92c openshift-network-node-identity/network-node-identity-vrzqb]\\\\nF1002 01:47:40.991998 6578 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller ini\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:40Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-dngcm_openshift-ovn-kubernetes(9e4679dd-f870-41e9-a32b-360f5d2ee81b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fafdd1094bd02786d745c984989852b0c690db38655f92b419733a2004f417c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0125e907d061a97286026ee823c4eed303734ed03ab4864f742592b23a1aa2f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0125e907d061a97286026ee823c4eed303734ed03ab4864f742592b23a1aa2f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dngcm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:02Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:02 crc kubenswrapper[4885]: I1002 01:48:02.110639 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-sv4md" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"724a3e49-001c-45a4-a896-82c13e0ee0e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hpqgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hpqgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-sv4md\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:02Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:02 crc kubenswrapper[4885]: I1002 01:48:02.127001 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18466979-7c7c-4d73-86c6-b8b6eaf20484\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b742e4b288bb50569d2fd23f3f9857d649a46063998170eee582ecbbba64d190\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://298c2ace58c4a9d09cfdc864316317e370bb0ac5735d631af6d85c6298f686d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://620f54e02c463f6e603808e5e57e9e5770d125578cc258ba16117a2416e57d3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b4cd17b64e5657ae05e71f3ca2340cb0d7965b1b49d1f7b4bfe4494292293ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b4cd17b64e5657ae05e71f3ca2340cb0d7965b1b49d1f7b4bfe4494292293ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:02Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:02 crc kubenswrapper[4885]: I1002 01:48:02.150426 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:02 crc kubenswrapper[4885]: I1002 01:48:02.150473 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:02 crc kubenswrapper[4885]: I1002 01:48:02.150483 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:02 crc kubenswrapper[4885]: I1002 01:48:02.150500 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:02 crc kubenswrapper[4885]: I1002 01:48:02.150509 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:02Z","lastTransitionTime":"2025-10-02T01:48:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:02 crc kubenswrapper[4885]: I1002 01:48:02.161460 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8070bc1c-e3de-4e9a-985e-40d1a9858056\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3e634b3727c517cc48b17c2cfb38f26460f53d1bf25d631d08ec9acb8d780e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4adcc537c23697a324d82af94c31059dd444245efdf7c978db58ede256e112ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edcd1d99644a778dfb6d0a17d2a4abd906400fe8209e611ebf8d7a9abe7ccf28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://97cb59adc2ddfa365ebc9dec649600dc90466ffe59134a6d15a9d3134a7c9b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d2808bc658ea80d68469ed16422d56ef19447f4c90b4af7b2597c64406f8a44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://83acbc20919f7c1958b6d492fadb203855df02761662425df9f91a0757fe2bf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83acbc20919f7c1958b6d492fadb203855df02761662425df9f91a0757fe2bf7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e86ecd03835efc9a253b6804bc61e6ac690e75b8d1f3cc77f5647730e586574\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e86ecd03835efc9a253b6804bc61e6ac690e75b8d1f3cc77f5647730e586574\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://48f95af51feaf014a18496bdfadd59d7f1e11167c5c8b677fa7037e48368ff2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48f95af51feaf014a18496bdfadd59d7f1e11167c5c8b677fa7037e48368ff2a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:02Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:02 crc kubenswrapper[4885]: I1002 01:48:02.177996 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5c1f95f188569c9aa47e51f2916939fa0b65a5cdf4c16b798612d41f3c8411a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:02Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:02 crc kubenswrapper[4885]: I1002 01:48:02.191244 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vhgbd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:48:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:48:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c63a7d20d39b65acfb5a9be4c28b21c72111b398bea76c238e15ddb4dd63e139\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c63a7d20d39b65acfb5a9be4c28b21c72111b398bea76c238e15ddb4dd63e139\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T01:48:00Z\\\",\\\"message\\\":\\\"2025-10-02T01:47:15+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_9f42b0ef-8651-49db-8a31-d21c0c1c47e7\\\\n2025-10-02T01:47:15+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_9f42b0ef-8651-49db-8a31-d21c0c1c47e7 to /host/opt/cni/bin/\\\\n2025-10-02T01:47:15Z [verbose] multus-daemon started\\\\n2025-10-02T01:47:15Z [verbose] Readiness Indicator file check\\\\n2025-10-02T01:48:00Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7zwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vhgbd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:02Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:02 crc kubenswrapper[4885]: I1002 01:48:02.204453 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rrn8b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52967a77-525c-4537-9b1f-d7bb754a8494\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3dee451ebe9006bf1c519fed4c65267570b6139d76841858352e087ae78d26a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtr74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e9412788d6df7332592beac48281822b26e1aa19b39addce8e1be556af9d982\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtr74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:26Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rrn8b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:02Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:02 crc kubenswrapper[4885]: I1002 01:48:02.219754 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:02Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:02 crc kubenswrapper[4885]: I1002 01:48:02.238736 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74adcf912ccedb6b5d4f2803fd215e3e2ef62367597aa6da528354c4511ee536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7403ef968d040ac09d722b7557438bef23892d578992826f42b700b370d495d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:02Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:02 crc kubenswrapper[4885]: I1002 01:48:02.253024 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:02 crc kubenswrapper[4885]: I1002 01:48:02.253054 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:02 crc kubenswrapper[4885]: I1002 01:48:02.253065 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:02 crc kubenswrapper[4885]: I1002 01:48:02.253078 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:02 crc kubenswrapper[4885]: I1002 01:48:02.253088 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:02Z","lastTransitionTime":"2025-10-02T01:48:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:02 crc kubenswrapper[4885]: I1002 01:48:02.263612 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2ll9b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52b39712-5fa1-4fe0-814b-b170a6e3938b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9a96383d0fbfd116c419054aaf1a7adba1937e8b0bad294cefd68b4ab0d2298\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c979b91ba748aa4647826a8926337662ca929c54090fb62285572819ec2fbf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c979b91ba748aa4647826a8926337662ca929c54090fb62285572819ec2fbf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97d9d693e8e60ee9c183bb44461e698e8b3cfd20a516605aa8754c9205d2bdc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://97d9d693e8e60ee9c183bb44461e698e8b3cfd20a516605aa8754c9205d2bdc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c44252676359c671d65c9588b05572d76c99cf4c90666f0ac089b2f778450a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c44252676359c671d65c9588b05572d76c99cf4c90666f0ac089b2f778450a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80c982f07874df686fe7e29cc91043eb6d58f1be5fc93cacd46d33a21dd9b622\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80c982f07874df686fe7e29cc91043eb6d58f1be5fc93cacd46d33a21dd9b622\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a354c985652c7ad7fa8944fbfdfe223cfb0f97fb11eecd9bd429020d4892b78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a354c985652c7ad7fa8944fbfdfe223cfb0f97fb11eecd9bd429020d4892b78\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2efd7283df313d086c1b8b88f289ace28764de256205bee7610d6711002d8515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2efd7283df313d086c1b8b88f289ace28764de256205bee7610d6711002d8515\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2ll9b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:02Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:02 crc kubenswrapper[4885]: I1002 01:48:02.277724 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-58lhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac6a19c4-f259-4dbd-a016-d6c2c849c6cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed76c0f968c063f7795d363af194a222487b647ceda234de0dabb39b25eafe0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hzc4c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-58lhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:02Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:02 crc kubenswrapper[4885]: I1002 01:48:02.290341 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d169c8ba-000a-4218-8e3b-4ae53035b110\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61e436b7876107ce5ea9e668c0bbc367be6e663fc804fa3a65b51adf8be8d51f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-77rzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09536e182a46e77318c10812d1777e551f834907c006cab5c726dc66f5bce88a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-77rzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rttx8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:02Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:02 crc kubenswrapper[4885]: I1002 01:48:02.302133 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jgdvt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81733e8d-67b0-46fa-8a41-08b40645bd3e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcef60c816ca1ba9a298da658e84c4d717bea3d352c4c2f807ebd8e56327b384\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cwbl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jgdvt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:02Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:02 crc kubenswrapper[4885]: I1002 01:48:02.321147 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a0bb885-2eb0-419c-8a7a-ef634cb3c144\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a89b49cbe08d927629564e607de13655ddf3d01ac4eb5d688b325210306c120c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06c5942805a498d9621a906088aa54538b4eed583d1b8d0e3ba2a6718378a36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b9f6f7fe44d5771a838d0b89c5bdea3ebb79023ae2675ee907ec2766b3d35a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://178b4340ee6904c60c814b99d83ee225161aeb23e438d8d02f856579a13ec4e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bb63bf30d6ecce9e71f0db0c16632bdfa6b33924b673378eb0c38ab229fa837\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T01:47:06Z\\\",\\\"message\\\":\\\"W1002 01:46:55.406027 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1002 01:46:55.406714 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759369615 cert, and key in /tmp/serving-cert-3795040473/serving-signer.crt, /tmp/serving-cert-3795040473/serving-signer.key\\\\nI1002 01:46:55.614670 1 observer_polling.go:159] Starting file observer\\\\nW1002 01:46:55.617662 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1002 01:46:55.618112 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 01:46:55.620347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3795040473/tls.crt::/tmp/serving-cert-3795040473/tls.key\\\\\\\"\\\\nF1002 01:47:06.224459 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90535f97ab2ed4626d47540ebd7baada837fd332140323bec0f49aa38d3c0009\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://592f89a326c9c1559193a71904b47ca27b1dbe1a7b30e03bc48924bb2f35f921\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://592f89a326c9c1559193a71904b47ca27b1dbe1a7b30e03bc48924bb2f35f921\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:02Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:02 crc kubenswrapper[4885]: I1002 01:48:02.336498 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b087880e05039816dcab154643ac72c7e4a0a1124817c51dc2c639b62bbee0fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:02Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:02 crc kubenswrapper[4885]: I1002 01:48:02.351798 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:02Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:02 crc kubenswrapper[4885]: I1002 01:48:02.355562 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:02 crc kubenswrapper[4885]: I1002 01:48:02.355595 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:02 crc kubenswrapper[4885]: I1002 01:48:02.355604 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:02 crc kubenswrapper[4885]: I1002 01:48:02.355619 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:02 crc kubenswrapper[4885]: I1002 01:48:02.355630 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:02Z","lastTransitionTime":"2025-10-02T01:48:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:02 crc kubenswrapper[4885]: I1002 01:48:02.366667 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:02Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:02 crc kubenswrapper[4885]: I1002 01:48:02.458754 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:02 crc kubenswrapper[4885]: I1002 01:48:02.458809 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:02 crc kubenswrapper[4885]: I1002 01:48:02.458818 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:02 crc kubenswrapper[4885]: I1002 01:48:02.458835 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:02 crc kubenswrapper[4885]: I1002 01:48:02.458846 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:02Z","lastTransitionTime":"2025-10-02T01:48:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:02 crc kubenswrapper[4885]: I1002 01:48:02.561451 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:02 crc kubenswrapper[4885]: I1002 01:48:02.561517 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:02 crc kubenswrapper[4885]: I1002 01:48:02.561534 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:02 crc kubenswrapper[4885]: I1002 01:48:02.561558 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:02 crc kubenswrapper[4885]: I1002 01:48:02.561575 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:02Z","lastTransitionTime":"2025-10-02T01:48:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:02 crc kubenswrapper[4885]: I1002 01:48:02.590970 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-vhgbd_ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f/kube-multus/0.log" Oct 02 01:48:02 crc kubenswrapper[4885]: I1002 01:48:02.591055 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-vhgbd" event={"ID":"ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f","Type":"ContainerStarted","Data":"df9792d80c807ed98c8442822d01f619327361c40425abfe26f11e5c27ba2584"} Oct 02 01:48:02 crc kubenswrapper[4885]: I1002 01:48:02.605610 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-58lhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac6a19c4-f259-4dbd-a016-d6c2c849c6cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed76c0f968c063f7795d363af194a222487b647ceda234de0dabb39b25eafe0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hzc4c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-58lhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:02Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:02 crc kubenswrapper[4885]: I1002 01:48:02.621981 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d169c8ba-000a-4218-8e3b-4ae53035b110\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61e436b7876107ce5ea9e668c0bbc367be6e663fc804fa3a65b51adf8be8d51f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-77rzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09536e182a46e77318c10812d1777e551f834907c006cab5c726dc66f5bce88a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-77rzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rttx8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:02Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:02 crc kubenswrapper[4885]: I1002 01:48:02.637064 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jgdvt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81733e8d-67b0-46fa-8a41-08b40645bd3e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcef60c816ca1ba9a298da658e84c4d717bea3d352c4c2f807ebd8e56327b384\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cwbl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jgdvt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:02Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:02 crc kubenswrapper[4885]: I1002 01:48:02.652506 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a0bb885-2eb0-419c-8a7a-ef634cb3c144\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a89b49cbe08d927629564e607de13655ddf3d01ac4eb5d688b325210306c120c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06c5942805a498d9621a906088aa54538b4eed583d1b8d0e3ba2a6718378a36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b9f6f7fe44d5771a838d0b89c5bdea3ebb79023ae2675ee907ec2766b3d35a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://178b4340ee6904c60c814b99d83ee225161aeb23e438d8d02f856579a13ec4e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bb63bf30d6ecce9e71f0db0c16632bdfa6b33924b673378eb0c38ab229fa837\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T01:47:06Z\\\",\\\"message\\\":\\\"W1002 01:46:55.406027 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1002 01:46:55.406714 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759369615 cert, and key in /tmp/serving-cert-3795040473/serving-signer.crt, /tmp/serving-cert-3795040473/serving-signer.key\\\\nI1002 01:46:55.614670 1 observer_polling.go:159] Starting file observer\\\\nW1002 01:46:55.617662 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1002 01:46:55.618112 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 01:46:55.620347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3795040473/tls.crt::/tmp/serving-cert-3795040473/tls.key\\\\\\\"\\\\nF1002 01:47:06.224459 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90535f97ab2ed4626d47540ebd7baada837fd332140323bec0f49aa38d3c0009\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://592f89a326c9c1559193a71904b47ca27b1dbe1a7b30e03bc48924bb2f35f921\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://592f89a326c9c1559193a71904b47ca27b1dbe1a7b30e03bc48924bb2f35f921\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:02Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:02 crc kubenswrapper[4885]: I1002 01:48:02.663498 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:02 crc kubenswrapper[4885]: I1002 01:48:02.663536 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:02 crc kubenswrapper[4885]: I1002 01:48:02.663545 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:02 crc kubenswrapper[4885]: I1002 01:48:02.663560 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:02 crc kubenswrapper[4885]: I1002 01:48:02.663570 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:02Z","lastTransitionTime":"2025-10-02T01:48:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:02 crc kubenswrapper[4885]: I1002 01:48:02.669935 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b087880e05039816dcab154643ac72c7e4a0a1124817c51dc2c639b62bbee0fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:02Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:02 crc kubenswrapper[4885]: I1002 01:48:02.683678 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:02Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:02 crc kubenswrapper[4885]: I1002 01:48:02.700029 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:02Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:02 crc kubenswrapper[4885]: I1002 01:48:02.713465 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"506a599f-fb1c-4d14-be88-0e4833f3d624\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31c2f0193ffef54965b79a3569b0f6f4547e540bac0774155159a1c996c196c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://533e73e5f90b7258acbcdf018ee01323c118f9dc45035b25b065ba01cabecccf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c00db833a593ec40eb63a6b383455e19e960c2e98b9b66f76877ee993057e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21e2f462a0deb0d740185663b2014f4a1e57b4fdbeb82571264b002f928035bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:02Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:02 crc kubenswrapper[4885]: I1002 01:48:02.738867 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e4679dd-f870-41e9-a32b-360f5d2ee81b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4310909ad9c7b609ed6814e04a525a6afae747775598e7275080413dec8b23a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7642033b0b5995d332ccf7992d6d7e6a5e9c9369ae2d1071ccb2f536938c8f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9f99f9e26a04fb42b01a0783a5c4e360514818ce9102edc7fa08626d50b9a34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6fa72b030d309656b73121e4ac2a3924a4e449aad66683dd9de5ee5f22d5a4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef403e519d197fb37b3e6b76a14a5128ba3f30a7a5e8bf0a420cf6a39b89575f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8482ba457683ffa03100ec1f07ea9d3f918be5e490b5ba05c2ba965d03033191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://188cb50b58926683ce8862d618258c1d1a6f33310608546c8651ece28f7c338f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://188cb50b58926683ce8862d618258c1d1a6f33310608546c8651ece28f7c338f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T01:47:41Z\\\",\\\"message\\\":\\\"921 6578 obj_retry.go:434] periodicallyRetryResources: Retry channel got triggered: retrying failed objects of type *v1.Pod\\\\nI1002 01:47:40.991945 6578 obj_retry.go:409] Going to retry *v1.Pod resource setup for 16 objects: [openshift-multus/network-metrics-daemon-sv4md openshift-network-operator/iptables-alerter-4ln5h openshift-kube-apiserver/kube-apiserver-crc openshift-ovn-kubernetes/ovnkube-node-dngcm openshift-dns/node-resolver-58lhj openshift-machine-config-operator/machine-config-daemon-rttx8 openshift-multus/multus-vhgbd openshift-network-console/networking-console-plugin-85b44fc459-gdk6g openshift-network-diagnostics/network-check-source-55646444c4-trplf openshift-network-operator/network-operator-58b4c7f79c-55gtf openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rrn8b openshift-image-registry/node-ca-jgdvt openshift-kube-controller-manager/kube-controller-manager-crc openshift-multus/multus-additional-cni-plugins-2ll9b openshift-network-diagnostics/network-check-target-xd92c openshift-network-node-identity/network-node-identity-vrzqb]\\\\nF1002 01:47:40.991998 6578 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller ini\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:40Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-dngcm_openshift-ovn-kubernetes(9e4679dd-f870-41e9-a32b-360f5d2ee81b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fafdd1094bd02786d745c984989852b0c690db38655f92b419733a2004f417c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0125e907d061a97286026ee823c4eed303734ed03ab4864f742592b23a1aa2f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0125e907d061a97286026ee823c4eed303734ed03ab4864f742592b23a1aa2f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dngcm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:02Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:02 crc kubenswrapper[4885]: I1002 01:48:02.753632 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-sv4md" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"724a3e49-001c-45a4-a896-82c13e0ee0e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hpqgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hpqgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-sv4md\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:02Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:02 crc kubenswrapper[4885]: I1002 01:48:02.767248 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:02 crc kubenswrapper[4885]: I1002 01:48:02.767355 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:02 crc kubenswrapper[4885]: I1002 01:48:02.767373 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:02 crc kubenswrapper[4885]: I1002 01:48:02.767462 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:02 crc kubenswrapper[4885]: I1002 01:48:02.767483 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:02Z","lastTransitionTime":"2025-10-02T01:48:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:02 crc kubenswrapper[4885]: I1002 01:48:02.768649 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18466979-7c7c-4d73-86c6-b8b6eaf20484\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b742e4b288bb50569d2fd23f3f9857d649a46063998170eee582ecbbba64d190\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://298c2ace58c4a9d09cfdc864316317e370bb0ac5735d631af6d85c6298f686d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://620f54e02c463f6e603808e5e57e9e5770d125578cc258ba16117a2416e57d3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b4cd17b64e5657ae05e71f3ca2340cb0d7965b1b49d1f7b4bfe4494292293ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b4cd17b64e5657ae05e71f3ca2340cb0d7965b1b49d1f7b4bfe4494292293ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:02Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:02 crc kubenswrapper[4885]: I1002 01:48:02.793117 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8070bc1c-e3de-4e9a-985e-40d1a9858056\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3e634b3727c517cc48b17c2cfb38f26460f53d1bf25d631d08ec9acb8d780e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4adcc537c23697a324d82af94c31059dd444245efdf7c978db58ede256e112ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edcd1d99644a778dfb6d0a17d2a4abd906400fe8209e611ebf8d7a9abe7ccf28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://97cb59adc2ddfa365ebc9dec649600dc90466ffe59134a6d15a9d3134a7c9b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d2808bc658ea80d68469ed16422d56ef19447f4c90b4af7b2597c64406f8a44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://83acbc20919f7c1958b6d492fadb203855df02761662425df9f91a0757fe2bf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83acbc20919f7c1958b6d492fadb203855df02761662425df9f91a0757fe2bf7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e86ecd03835efc9a253b6804bc61e6ac690e75b8d1f3cc77f5647730e586574\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e86ecd03835efc9a253b6804bc61e6ac690e75b8d1f3cc77f5647730e586574\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://48f95af51feaf014a18496bdfadd59d7f1e11167c5c8b677fa7037e48368ff2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48f95af51feaf014a18496bdfadd59d7f1e11167c5c8b677fa7037e48368ff2a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:02Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:02 crc kubenswrapper[4885]: I1002 01:48:02.809789 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5c1f95f188569c9aa47e51f2916939fa0b65a5cdf4c16b798612d41f3c8411a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:02Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:02 crc kubenswrapper[4885]: I1002 01:48:02.827744 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vhgbd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df9792d80c807ed98c8442822d01f619327361c40425abfe26f11e5c27ba2584\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c63a7d20d39b65acfb5a9be4c28b21c72111b398bea76c238e15ddb4dd63e139\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T01:48:00Z\\\",\\\"message\\\":\\\"2025-10-02T01:47:15+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_9f42b0ef-8651-49db-8a31-d21c0c1c47e7\\\\n2025-10-02T01:47:15+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_9f42b0ef-8651-49db-8a31-d21c0c1c47e7 to /host/opt/cni/bin/\\\\n2025-10-02T01:47:15Z [verbose] multus-daemon started\\\\n2025-10-02T01:47:15Z [verbose] Readiness Indicator file check\\\\n2025-10-02T01:48:00Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7zwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vhgbd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:02Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:02 crc kubenswrapper[4885]: I1002 01:48:02.842077 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rrn8b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52967a77-525c-4537-9b1f-d7bb754a8494\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3dee451ebe9006bf1c519fed4c65267570b6139d76841858352e087ae78d26a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtr74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e9412788d6df7332592beac48281822b26e1aa19b39addce8e1be556af9d982\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtr74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:26Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rrn8b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:02Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:02 crc kubenswrapper[4885]: I1002 01:48:02.859509 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:02Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:02 crc kubenswrapper[4885]: I1002 01:48:02.871107 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:02 crc kubenswrapper[4885]: I1002 01:48:02.871139 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:02 crc kubenswrapper[4885]: I1002 01:48:02.871149 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:02 crc kubenswrapper[4885]: I1002 01:48:02.871173 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:02 crc kubenswrapper[4885]: I1002 01:48:02.871185 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:02Z","lastTransitionTime":"2025-10-02T01:48:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:02 crc kubenswrapper[4885]: I1002 01:48:02.876414 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74adcf912ccedb6b5d4f2803fd215e3e2ef62367597aa6da528354c4511ee536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7403ef968d040ac09d722b7557438bef23892d578992826f42b700b370d495d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:02Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:02 crc kubenswrapper[4885]: I1002 01:48:02.895845 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2ll9b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52b39712-5fa1-4fe0-814b-b170a6e3938b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9a96383d0fbfd116c419054aaf1a7adba1937e8b0bad294cefd68b4ab0d2298\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c979b91ba748aa4647826a8926337662ca929c54090fb62285572819ec2fbf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c979b91ba748aa4647826a8926337662ca929c54090fb62285572819ec2fbf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97d9d693e8e60ee9c183bb44461e698e8b3cfd20a516605aa8754c9205d2bdc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://97d9d693e8e60ee9c183bb44461e698e8b3cfd20a516605aa8754c9205d2bdc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c44252676359c671d65c9588b05572d76c99cf4c90666f0ac089b2f778450a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c44252676359c671d65c9588b05572d76c99cf4c90666f0ac089b2f778450a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80c982f07874df686fe7e29cc91043eb6d58f1be5fc93cacd46d33a21dd9b622\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80c982f07874df686fe7e29cc91043eb6d58f1be5fc93cacd46d33a21dd9b622\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a354c985652c7ad7fa8944fbfdfe223cfb0f97fb11eecd9bd429020d4892b78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a354c985652c7ad7fa8944fbfdfe223cfb0f97fb11eecd9bd429020d4892b78\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2efd7283df313d086c1b8b88f289ace28764de256205bee7610d6711002d8515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2efd7283df313d086c1b8b88f289ace28764de256205bee7610d6711002d8515\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2ll9b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:02Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:02 crc kubenswrapper[4885]: I1002 01:48:02.973223 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:02 crc kubenswrapper[4885]: I1002 01:48:02.973297 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:02 crc kubenswrapper[4885]: I1002 01:48:02.973309 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:02 crc kubenswrapper[4885]: I1002 01:48:02.973329 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:02 crc kubenswrapper[4885]: I1002 01:48:02.973340 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:02Z","lastTransitionTime":"2025-10-02T01:48:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:03 crc kubenswrapper[4885]: I1002 01:48:03.076077 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:03 crc kubenswrapper[4885]: I1002 01:48:03.076128 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:03 crc kubenswrapper[4885]: I1002 01:48:03.076141 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:03 crc kubenswrapper[4885]: I1002 01:48:03.076176 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:03 crc kubenswrapper[4885]: I1002 01:48:03.076187 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:03Z","lastTransitionTime":"2025-10-02T01:48:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:03 crc kubenswrapper[4885]: I1002 01:48:03.179334 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:03 crc kubenswrapper[4885]: I1002 01:48:03.179388 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:03 crc kubenswrapper[4885]: I1002 01:48:03.179406 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:03 crc kubenswrapper[4885]: I1002 01:48:03.179431 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:03 crc kubenswrapper[4885]: I1002 01:48:03.179450 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:03Z","lastTransitionTime":"2025-10-02T01:48:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:03 crc kubenswrapper[4885]: I1002 01:48:03.282103 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:03 crc kubenswrapper[4885]: I1002 01:48:03.282162 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:03 crc kubenswrapper[4885]: I1002 01:48:03.282181 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:03 crc kubenswrapper[4885]: I1002 01:48:03.282201 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:03 crc kubenswrapper[4885]: I1002 01:48:03.282217 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:03Z","lastTransitionTime":"2025-10-02T01:48:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:03 crc kubenswrapper[4885]: I1002 01:48:03.385290 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:03 crc kubenswrapper[4885]: I1002 01:48:03.385354 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:03 crc kubenswrapper[4885]: I1002 01:48:03.385375 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:03 crc kubenswrapper[4885]: I1002 01:48:03.385403 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:03 crc kubenswrapper[4885]: I1002 01:48:03.385423 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:03Z","lastTransitionTime":"2025-10-02T01:48:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:03 crc kubenswrapper[4885]: I1002 01:48:03.488312 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:03 crc kubenswrapper[4885]: I1002 01:48:03.488363 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:03 crc kubenswrapper[4885]: I1002 01:48:03.488379 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:03 crc kubenswrapper[4885]: I1002 01:48:03.488403 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:03 crc kubenswrapper[4885]: I1002 01:48:03.488423 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:03Z","lastTransitionTime":"2025-10-02T01:48:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:03 crc kubenswrapper[4885]: I1002 01:48:03.590531 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:03 crc kubenswrapper[4885]: I1002 01:48:03.590596 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:03 crc kubenswrapper[4885]: I1002 01:48:03.590613 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:03 crc kubenswrapper[4885]: I1002 01:48:03.590639 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:03 crc kubenswrapper[4885]: I1002 01:48:03.590656 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:03Z","lastTransitionTime":"2025-10-02T01:48:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:03 crc kubenswrapper[4885]: I1002 01:48:03.693150 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:03 crc kubenswrapper[4885]: I1002 01:48:03.693209 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:03 crc kubenswrapper[4885]: I1002 01:48:03.693228 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:03 crc kubenswrapper[4885]: I1002 01:48:03.693251 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:03 crc kubenswrapper[4885]: I1002 01:48:03.693301 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:03Z","lastTransitionTime":"2025-10-02T01:48:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:03 crc kubenswrapper[4885]: I1002 01:48:03.796103 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:03 crc kubenswrapper[4885]: I1002 01:48:03.796195 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:03 crc kubenswrapper[4885]: I1002 01:48:03.796214 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:03 crc kubenswrapper[4885]: I1002 01:48:03.796240 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:03 crc kubenswrapper[4885]: I1002 01:48:03.796319 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:03Z","lastTransitionTime":"2025-10-02T01:48:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:03 crc kubenswrapper[4885]: I1002 01:48:03.899420 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:03 crc kubenswrapper[4885]: I1002 01:48:03.899490 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:03 crc kubenswrapper[4885]: I1002 01:48:03.899508 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:03 crc kubenswrapper[4885]: I1002 01:48:03.899534 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:03 crc kubenswrapper[4885]: I1002 01:48:03.899553 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:03Z","lastTransitionTime":"2025-10-02T01:48:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:04 crc kubenswrapper[4885]: I1002 01:48:04.001932 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:04 crc kubenswrapper[4885]: I1002 01:48:04.001992 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:04 crc kubenswrapper[4885]: I1002 01:48:04.002010 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:04 crc kubenswrapper[4885]: I1002 01:48:04.002035 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:04 crc kubenswrapper[4885]: I1002 01:48:04.002054 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:04Z","lastTransitionTime":"2025-10-02T01:48:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:04 crc kubenswrapper[4885]: I1002 01:48:04.045935 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:48:04 crc kubenswrapper[4885]: I1002 01:48:04.045993 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:48:04 crc kubenswrapper[4885]: E1002 01:48:04.046097 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:48:04 crc kubenswrapper[4885]: I1002 01:48:04.046109 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sv4md" Oct 02 01:48:04 crc kubenswrapper[4885]: E1002 01:48:04.046190 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:48:04 crc kubenswrapper[4885]: I1002 01:48:04.046202 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:48:04 crc kubenswrapper[4885]: E1002 01:48:04.046323 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sv4md" podUID="724a3e49-001c-45a4-a896-82c13e0ee0e7" Oct 02 01:48:04 crc kubenswrapper[4885]: E1002 01:48:04.046524 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:48:04 crc kubenswrapper[4885]: I1002 01:48:04.104426 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:04 crc kubenswrapper[4885]: I1002 01:48:04.104500 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:04 crc kubenswrapper[4885]: I1002 01:48:04.104523 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:04 crc kubenswrapper[4885]: I1002 01:48:04.104553 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:04 crc kubenswrapper[4885]: I1002 01:48:04.104575 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:04Z","lastTransitionTime":"2025-10-02T01:48:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:04 crc kubenswrapper[4885]: I1002 01:48:04.207513 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:04 crc kubenswrapper[4885]: I1002 01:48:04.207558 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:04 crc kubenswrapper[4885]: I1002 01:48:04.207568 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:04 crc kubenswrapper[4885]: I1002 01:48:04.207586 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:04 crc kubenswrapper[4885]: I1002 01:48:04.207598 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:04Z","lastTransitionTime":"2025-10-02T01:48:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:04 crc kubenswrapper[4885]: I1002 01:48:04.309380 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:04 crc kubenswrapper[4885]: I1002 01:48:04.309422 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:04 crc kubenswrapper[4885]: I1002 01:48:04.309431 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:04 crc kubenswrapper[4885]: I1002 01:48:04.309445 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:04 crc kubenswrapper[4885]: I1002 01:48:04.309455 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:04Z","lastTransitionTime":"2025-10-02T01:48:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:04 crc kubenswrapper[4885]: I1002 01:48:04.411769 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:04 crc kubenswrapper[4885]: I1002 01:48:04.411818 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:04 crc kubenswrapper[4885]: I1002 01:48:04.411835 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:04 crc kubenswrapper[4885]: I1002 01:48:04.411858 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:04 crc kubenswrapper[4885]: I1002 01:48:04.411876 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:04Z","lastTransitionTime":"2025-10-02T01:48:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:04 crc kubenswrapper[4885]: I1002 01:48:04.514005 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:04 crc kubenswrapper[4885]: I1002 01:48:04.514058 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:04 crc kubenswrapper[4885]: I1002 01:48:04.514081 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:04 crc kubenswrapper[4885]: I1002 01:48:04.514107 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:04 crc kubenswrapper[4885]: I1002 01:48:04.514169 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:04Z","lastTransitionTime":"2025-10-02T01:48:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:04 crc kubenswrapper[4885]: I1002 01:48:04.616768 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:04 crc kubenswrapper[4885]: I1002 01:48:04.616816 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:04 crc kubenswrapper[4885]: I1002 01:48:04.616836 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:04 crc kubenswrapper[4885]: I1002 01:48:04.616856 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:04 crc kubenswrapper[4885]: I1002 01:48:04.616873 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:04Z","lastTransitionTime":"2025-10-02T01:48:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:04 crc kubenswrapper[4885]: I1002 01:48:04.719466 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:04 crc kubenswrapper[4885]: I1002 01:48:04.719518 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:04 crc kubenswrapper[4885]: I1002 01:48:04.719536 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:04 crc kubenswrapper[4885]: I1002 01:48:04.719559 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:04 crc kubenswrapper[4885]: I1002 01:48:04.719575 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:04Z","lastTransitionTime":"2025-10-02T01:48:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:04 crc kubenswrapper[4885]: I1002 01:48:04.822011 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:04 crc kubenswrapper[4885]: I1002 01:48:04.822115 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:04 crc kubenswrapper[4885]: I1002 01:48:04.822141 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:04 crc kubenswrapper[4885]: I1002 01:48:04.822168 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:04 crc kubenswrapper[4885]: I1002 01:48:04.822189 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:04Z","lastTransitionTime":"2025-10-02T01:48:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:04 crc kubenswrapper[4885]: I1002 01:48:04.924316 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:04 crc kubenswrapper[4885]: I1002 01:48:04.924388 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:04 crc kubenswrapper[4885]: I1002 01:48:04.924410 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:04 crc kubenswrapper[4885]: I1002 01:48:04.924436 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:04 crc kubenswrapper[4885]: I1002 01:48:04.924458 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:04Z","lastTransitionTime":"2025-10-02T01:48:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:05 crc kubenswrapper[4885]: I1002 01:48:05.026429 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:05 crc kubenswrapper[4885]: I1002 01:48:05.026485 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:05 crc kubenswrapper[4885]: I1002 01:48:05.026556 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:05 crc kubenswrapper[4885]: I1002 01:48:05.026585 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:05 crc kubenswrapper[4885]: I1002 01:48:05.026659 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:05Z","lastTransitionTime":"2025-10-02T01:48:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:05 crc kubenswrapper[4885]: I1002 01:48:05.129225 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:05 crc kubenswrapper[4885]: I1002 01:48:05.129283 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:05 crc kubenswrapper[4885]: I1002 01:48:05.129293 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:05 crc kubenswrapper[4885]: I1002 01:48:05.129307 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:05 crc kubenswrapper[4885]: I1002 01:48:05.129318 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:05Z","lastTransitionTime":"2025-10-02T01:48:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:05 crc kubenswrapper[4885]: I1002 01:48:05.231764 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:05 crc kubenswrapper[4885]: I1002 01:48:05.231824 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:05 crc kubenswrapper[4885]: I1002 01:48:05.231841 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:05 crc kubenswrapper[4885]: I1002 01:48:05.231866 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:05 crc kubenswrapper[4885]: I1002 01:48:05.231883 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:05Z","lastTransitionTime":"2025-10-02T01:48:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:05 crc kubenswrapper[4885]: I1002 01:48:05.334038 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:05 crc kubenswrapper[4885]: I1002 01:48:05.334098 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:05 crc kubenswrapper[4885]: I1002 01:48:05.334115 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:05 crc kubenswrapper[4885]: I1002 01:48:05.334139 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:05 crc kubenswrapper[4885]: I1002 01:48:05.334156 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:05Z","lastTransitionTime":"2025-10-02T01:48:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:05 crc kubenswrapper[4885]: I1002 01:48:05.436477 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:05 crc kubenswrapper[4885]: I1002 01:48:05.436553 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:05 crc kubenswrapper[4885]: I1002 01:48:05.436571 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:05 crc kubenswrapper[4885]: I1002 01:48:05.436599 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:05 crc kubenswrapper[4885]: I1002 01:48:05.436616 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:05Z","lastTransitionTime":"2025-10-02T01:48:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:05 crc kubenswrapper[4885]: I1002 01:48:05.539479 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:05 crc kubenswrapper[4885]: I1002 01:48:05.539523 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:05 crc kubenswrapper[4885]: I1002 01:48:05.539541 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:05 crc kubenswrapper[4885]: I1002 01:48:05.539566 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:05 crc kubenswrapper[4885]: I1002 01:48:05.539583 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:05Z","lastTransitionTime":"2025-10-02T01:48:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:05 crc kubenswrapper[4885]: I1002 01:48:05.642610 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:05 crc kubenswrapper[4885]: I1002 01:48:05.642675 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:05 crc kubenswrapper[4885]: I1002 01:48:05.642694 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:05 crc kubenswrapper[4885]: I1002 01:48:05.642719 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:05 crc kubenswrapper[4885]: I1002 01:48:05.642737 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:05Z","lastTransitionTime":"2025-10-02T01:48:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:05 crc kubenswrapper[4885]: I1002 01:48:05.744969 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:05 crc kubenswrapper[4885]: I1002 01:48:05.745011 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:05 crc kubenswrapper[4885]: I1002 01:48:05.745028 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:05 crc kubenswrapper[4885]: I1002 01:48:05.745050 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:05 crc kubenswrapper[4885]: I1002 01:48:05.745066 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:05Z","lastTransitionTime":"2025-10-02T01:48:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:05 crc kubenswrapper[4885]: I1002 01:48:05.848318 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:05 crc kubenswrapper[4885]: I1002 01:48:05.848371 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:05 crc kubenswrapper[4885]: I1002 01:48:05.848387 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:05 crc kubenswrapper[4885]: I1002 01:48:05.848412 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:05 crc kubenswrapper[4885]: I1002 01:48:05.848428 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:05Z","lastTransitionTime":"2025-10-02T01:48:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:05 crc kubenswrapper[4885]: I1002 01:48:05.951389 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:05 crc kubenswrapper[4885]: I1002 01:48:05.951443 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:05 crc kubenswrapper[4885]: I1002 01:48:05.951456 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:05 crc kubenswrapper[4885]: I1002 01:48:05.951476 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:05 crc kubenswrapper[4885]: I1002 01:48:05.951491 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:05Z","lastTransitionTime":"2025-10-02T01:48:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:06 crc kubenswrapper[4885]: I1002 01:48:06.045669 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:48:06 crc kubenswrapper[4885]: I1002 01:48:06.045716 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:48:06 crc kubenswrapper[4885]: I1002 01:48:06.045740 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sv4md" Oct 02 01:48:06 crc kubenswrapper[4885]: E1002 01:48:06.045857 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:48:06 crc kubenswrapper[4885]: I1002 01:48:06.045666 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:48:06 crc kubenswrapper[4885]: E1002 01:48:06.046056 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:48:06 crc kubenswrapper[4885]: E1002 01:48:06.046117 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:48:06 crc kubenswrapper[4885]: E1002 01:48:06.046251 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sv4md" podUID="724a3e49-001c-45a4-a896-82c13e0ee0e7" Oct 02 01:48:06 crc kubenswrapper[4885]: I1002 01:48:06.053337 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:06 crc kubenswrapper[4885]: I1002 01:48:06.053369 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:06 crc kubenswrapper[4885]: I1002 01:48:06.053379 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:06 crc kubenswrapper[4885]: I1002 01:48:06.053392 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:06 crc kubenswrapper[4885]: I1002 01:48:06.053403 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:06Z","lastTransitionTime":"2025-10-02T01:48:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:06 crc kubenswrapper[4885]: I1002 01:48:06.155753 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:06 crc kubenswrapper[4885]: I1002 01:48:06.155790 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:06 crc kubenswrapper[4885]: I1002 01:48:06.155803 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:06 crc kubenswrapper[4885]: I1002 01:48:06.155819 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:06 crc kubenswrapper[4885]: I1002 01:48:06.155830 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:06Z","lastTransitionTime":"2025-10-02T01:48:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:06 crc kubenswrapper[4885]: I1002 01:48:06.261643 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:06 crc kubenswrapper[4885]: I1002 01:48:06.261714 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:06 crc kubenswrapper[4885]: I1002 01:48:06.261732 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:06 crc kubenswrapper[4885]: I1002 01:48:06.261761 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:06 crc kubenswrapper[4885]: I1002 01:48:06.261786 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:06Z","lastTransitionTime":"2025-10-02T01:48:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:06 crc kubenswrapper[4885]: I1002 01:48:06.364604 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:06 crc kubenswrapper[4885]: I1002 01:48:06.364707 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:06 crc kubenswrapper[4885]: I1002 01:48:06.364734 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:06 crc kubenswrapper[4885]: I1002 01:48:06.364771 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:06 crc kubenswrapper[4885]: I1002 01:48:06.364794 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:06Z","lastTransitionTime":"2025-10-02T01:48:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:06 crc kubenswrapper[4885]: I1002 01:48:06.467345 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:06 crc kubenswrapper[4885]: I1002 01:48:06.467426 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:06 crc kubenswrapper[4885]: I1002 01:48:06.467446 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:06 crc kubenswrapper[4885]: I1002 01:48:06.467476 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:06 crc kubenswrapper[4885]: I1002 01:48:06.467500 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:06Z","lastTransitionTime":"2025-10-02T01:48:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:06 crc kubenswrapper[4885]: I1002 01:48:06.570653 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:06 crc kubenswrapper[4885]: I1002 01:48:06.570715 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:06 crc kubenswrapper[4885]: I1002 01:48:06.570734 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:06 crc kubenswrapper[4885]: I1002 01:48:06.570762 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:06 crc kubenswrapper[4885]: I1002 01:48:06.570783 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:06Z","lastTransitionTime":"2025-10-02T01:48:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:06 crc kubenswrapper[4885]: I1002 01:48:06.674400 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:06 crc kubenswrapper[4885]: I1002 01:48:06.674470 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:06 crc kubenswrapper[4885]: I1002 01:48:06.674490 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:06 crc kubenswrapper[4885]: I1002 01:48:06.674515 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:06 crc kubenswrapper[4885]: I1002 01:48:06.674539 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:06Z","lastTransitionTime":"2025-10-02T01:48:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:06 crc kubenswrapper[4885]: I1002 01:48:06.778345 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:06 crc kubenswrapper[4885]: I1002 01:48:06.778376 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:06 crc kubenswrapper[4885]: I1002 01:48:06.778386 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:06 crc kubenswrapper[4885]: I1002 01:48:06.778400 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:06 crc kubenswrapper[4885]: I1002 01:48:06.778411 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:06Z","lastTransitionTime":"2025-10-02T01:48:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:06 crc kubenswrapper[4885]: I1002 01:48:06.880786 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:06 crc kubenswrapper[4885]: I1002 01:48:06.880890 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:06 crc kubenswrapper[4885]: I1002 01:48:06.880909 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:06 crc kubenswrapper[4885]: I1002 01:48:06.880934 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:06 crc kubenswrapper[4885]: I1002 01:48:06.880953 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:06Z","lastTransitionTime":"2025-10-02T01:48:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:06 crc kubenswrapper[4885]: I1002 01:48:06.983791 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:06 crc kubenswrapper[4885]: I1002 01:48:06.983844 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:06 crc kubenswrapper[4885]: I1002 01:48:06.983856 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:06 crc kubenswrapper[4885]: I1002 01:48:06.983871 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:06 crc kubenswrapper[4885]: I1002 01:48:06.983880 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:06Z","lastTransitionTime":"2025-10-02T01:48:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:07 crc kubenswrapper[4885]: I1002 01:48:07.086292 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:07 crc kubenswrapper[4885]: I1002 01:48:07.086316 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:07 crc kubenswrapper[4885]: I1002 01:48:07.086325 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:07 crc kubenswrapper[4885]: I1002 01:48:07.086337 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:07 crc kubenswrapper[4885]: I1002 01:48:07.086345 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:07Z","lastTransitionTime":"2025-10-02T01:48:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:07 crc kubenswrapper[4885]: I1002 01:48:07.189298 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:07 crc kubenswrapper[4885]: I1002 01:48:07.189356 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:07 crc kubenswrapper[4885]: I1002 01:48:07.189373 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:07 crc kubenswrapper[4885]: I1002 01:48:07.189399 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:07 crc kubenswrapper[4885]: I1002 01:48:07.189416 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:07Z","lastTransitionTime":"2025-10-02T01:48:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:07 crc kubenswrapper[4885]: I1002 01:48:07.291988 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:07 crc kubenswrapper[4885]: I1002 01:48:07.292058 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:07 crc kubenswrapper[4885]: I1002 01:48:07.292078 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:07 crc kubenswrapper[4885]: I1002 01:48:07.292104 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:07 crc kubenswrapper[4885]: I1002 01:48:07.292123 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:07Z","lastTransitionTime":"2025-10-02T01:48:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:07 crc kubenswrapper[4885]: I1002 01:48:07.394582 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:07 crc kubenswrapper[4885]: I1002 01:48:07.394650 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:07 crc kubenswrapper[4885]: I1002 01:48:07.394668 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:07 crc kubenswrapper[4885]: I1002 01:48:07.394694 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:07 crc kubenswrapper[4885]: I1002 01:48:07.394715 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:07Z","lastTransitionTime":"2025-10-02T01:48:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:07 crc kubenswrapper[4885]: I1002 01:48:07.497185 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:07 crc kubenswrapper[4885]: I1002 01:48:07.497237 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:07 crc kubenswrapper[4885]: I1002 01:48:07.497249 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:07 crc kubenswrapper[4885]: I1002 01:48:07.497293 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:07 crc kubenswrapper[4885]: I1002 01:48:07.497308 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:07Z","lastTransitionTime":"2025-10-02T01:48:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:07 crc kubenswrapper[4885]: I1002 01:48:07.598985 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:07 crc kubenswrapper[4885]: I1002 01:48:07.599052 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:07 crc kubenswrapper[4885]: I1002 01:48:07.599071 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:07 crc kubenswrapper[4885]: I1002 01:48:07.599100 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:07 crc kubenswrapper[4885]: I1002 01:48:07.599119 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:07Z","lastTransitionTime":"2025-10-02T01:48:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:07 crc kubenswrapper[4885]: I1002 01:48:07.702057 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:07 crc kubenswrapper[4885]: I1002 01:48:07.702113 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:07 crc kubenswrapper[4885]: I1002 01:48:07.702132 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:07 crc kubenswrapper[4885]: I1002 01:48:07.702155 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:07 crc kubenswrapper[4885]: I1002 01:48:07.702170 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:07Z","lastTransitionTime":"2025-10-02T01:48:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:07 crc kubenswrapper[4885]: I1002 01:48:07.804745 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:07 crc kubenswrapper[4885]: I1002 01:48:07.804787 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:07 crc kubenswrapper[4885]: I1002 01:48:07.804799 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:07 crc kubenswrapper[4885]: I1002 01:48:07.804816 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:07 crc kubenswrapper[4885]: I1002 01:48:07.804829 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:07Z","lastTransitionTime":"2025-10-02T01:48:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:07 crc kubenswrapper[4885]: I1002 01:48:07.910725 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:07 crc kubenswrapper[4885]: I1002 01:48:07.910788 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:07 crc kubenswrapper[4885]: I1002 01:48:07.910807 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:07 crc kubenswrapper[4885]: I1002 01:48:07.910833 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:07 crc kubenswrapper[4885]: I1002 01:48:07.910850 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:07Z","lastTransitionTime":"2025-10-02T01:48:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:08 crc kubenswrapper[4885]: I1002 01:48:08.013122 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:08 crc kubenswrapper[4885]: I1002 01:48:08.013199 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:08 crc kubenswrapper[4885]: I1002 01:48:08.013222 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:08 crc kubenswrapper[4885]: I1002 01:48:08.013289 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:08 crc kubenswrapper[4885]: I1002 01:48:08.013319 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:08Z","lastTransitionTime":"2025-10-02T01:48:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:08 crc kubenswrapper[4885]: I1002 01:48:08.045614 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:48:08 crc kubenswrapper[4885]: I1002 01:48:08.045674 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:48:08 crc kubenswrapper[4885]: I1002 01:48:08.045703 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sv4md" Oct 02 01:48:08 crc kubenswrapper[4885]: I1002 01:48:08.045632 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:48:08 crc kubenswrapper[4885]: E1002 01:48:08.045783 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:48:08 crc kubenswrapper[4885]: E1002 01:48:08.046079 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sv4md" podUID="724a3e49-001c-45a4-a896-82c13e0ee0e7" Oct 02 01:48:08 crc kubenswrapper[4885]: E1002 01:48:08.046177 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:48:08 crc kubenswrapper[4885]: E1002 01:48:08.046428 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:48:08 crc kubenswrapper[4885]: I1002 01:48:08.116557 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:08 crc kubenswrapper[4885]: I1002 01:48:08.116610 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:08 crc kubenswrapper[4885]: I1002 01:48:08.116623 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:08 crc kubenswrapper[4885]: I1002 01:48:08.116642 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:08 crc kubenswrapper[4885]: I1002 01:48:08.116656 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:08Z","lastTransitionTime":"2025-10-02T01:48:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:08 crc kubenswrapper[4885]: I1002 01:48:08.220350 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:08 crc kubenswrapper[4885]: I1002 01:48:08.220448 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:08 crc kubenswrapper[4885]: I1002 01:48:08.220465 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:08 crc kubenswrapper[4885]: I1002 01:48:08.220491 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:08 crc kubenswrapper[4885]: I1002 01:48:08.220510 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:08Z","lastTransitionTime":"2025-10-02T01:48:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:08 crc kubenswrapper[4885]: I1002 01:48:08.323836 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:08 crc kubenswrapper[4885]: I1002 01:48:08.323891 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:08 crc kubenswrapper[4885]: I1002 01:48:08.323952 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:08 crc kubenswrapper[4885]: I1002 01:48:08.323988 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:08 crc kubenswrapper[4885]: I1002 01:48:08.324013 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:08Z","lastTransitionTime":"2025-10-02T01:48:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:08 crc kubenswrapper[4885]: I1002 01:48:08.427227 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:08 crc kubenswrapper[4885]: I1002 01:48:08.427333 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:08 crc kubenswrapper[4885]: I1002 01:48:08.427359 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:08 crc kubenswrapper[4885]: I1002 01:48:08.427387 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:08 crc kubenswrapper[4885]: I1002 01:48:08.427407 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:08Z","lastTransitionTime":"2025-10-02T01:48:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:08 crc kubenswrapper[4885]: I1002 01:48:08.530443 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:08 crc kubenswrapper[4885]: I1002 01:48:08.530512 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:08 crc kubenswrapper[4885]: I1002 01:48:08.530530 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:08 crc kubenswrapper[4885]: I1002 01:48:08.530557 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:08 crc kubenswrapper[4885]: I1002 01:48:08.530578 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:08Z","lastTransitionTime":"2025-10-02T01:48:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:08 crc kubenswrapper[4885]: I1002 01:48:08.634103 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:08 crc kubenswrapper[4885]: I1002 01:48:08.634203 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:08 crc kubenswrapper[4885]: I1002 01:48:08.634247 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:08 crc kubenswrapper[4885]: I1002 01:48:08.634316 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:08 crc kubenswrapper[4885]: I1002 01:48:08.634335 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:08Z","lastTransitionTime":"2025-10-02T01:48:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:08 crc kubenswrapper[4885]: I1002 01:48:08.737910 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:08 crc kubenswrapper[4885]: I1002 01:48:08.737981 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:08 crc kubenswrapper[4885]: I1002 01:48:08.738003 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:08 crc kubenswrapper[4885]: I1002 01:48:08.738031 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:08 crc kubenswrapper[4885]: I1002 01:48:08.738054 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:08Z","lastTransitionTime":"2025-10-02T01:48:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:08 crc kubenswrapper[4885]: I1002 01:48:08.841370 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:08 crc kubenswrapper[4885]: I1002 01:48:08.841483 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:08 crc kubenswrapper[4885]: I1002 01:48:08.841514 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:08 crc kubenswrapper[4885]: I1002 01:48:08.841546 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:08 crc kubenswrapper[4885]: I1002 01:48:08.841567 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:08Z","lastTransitionTime":"2025-10-02T01:48:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:08 crc kubenswrapper[4885]: I1002 01:48:08.944614 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:08 crc kubenswrapper[4885]: I1002 01:48:08.944745 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:08 crc kubenswrapper[4885]: I1002 01:48:08.944781 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:08 crc kubenswrapper[4885]: I1002 01:48:08.944853 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:08 crc kubenswrapper[4885]: I1002 01:48:08.944877 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:08Z","lastTransitionTime":"2025-10-02T01:48:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:09 crc kubenswrapper[4885]: I1002 01:48:09.047364 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:09 crc kubenswrapper[4885]: I1002 01:48:09.047435 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:09 crc kubenswrapper[4885]: I1002 01:48:09.047450 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:09 crc kubenswrapper[4885]: I1002 01:48:09.047477 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:09 crc kubenswrapper[4885]: I1002 01:48:09.047495 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:09Z","lastTransitionTime":"2025-10-02T01:48:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:09 crc kubenswrapper[4885]: I1002 01:48:09.155015 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:09 crc kubenswrapper[4885]: I1002 01:48:09.155095 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:09 crc kubenswrapper[4885]: I1002 01:48:09.155114 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:09 crc kubenswrapper[4885]: I1002 01:48:09.155141 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:09 crc kubenswrapper[4885]: I1002 01:48:09.155160 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:09Z","lastTransitionTime":"2025-10-02T01:48:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:09 crc kubenswrapper[4885]: I1002 01:48:09.258294 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:09 crc kubenswrapper[4885]: I1002 01:48:09.258380 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:09 crc kubenswrapper[4885]: I1002 01:48:09.258400 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:09 crc kubenswrapper[4885]: I1002 01:48:09.258427 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:09 crc kubenswrapper[4885]: I1002 01:48:09.258444 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:09Z","lastTransitionTime":"2025-10-02T01:48:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:09 crc kubenswrapper[4885]: I1002 01:48:09.361161 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:09 crc kubenswrapper[4885]: I1002 01:48:09.361205 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:09 crc kubenswrapper[4885]: I1002 01:48:09.361223 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:09 crc kubenswrapper[4885]: I1002 01:48:09.361245 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:09 crc kubenswrapper[4885]: I1002 01:48:09.361296 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:09Z","lastTransitionTime":"2025-10-02T01:48:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:09 crc kubenswrapper[4885]: I1002 01:48:09.464185 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:09 crc kubenswrapper[4885]: I1002 01:48:09.464304 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:09 crc kubenswrapper[4885]: I1002 01:48:09.464334 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:09 crc kubenswrapper[4885]: I1002 01:48:09.464373 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:09 crc kubenswrapper[4885]: I1002 01:48:09.464395 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:09Z","lastTransitionTime":"2025-10-02T01:48:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:09 crc kubenswrapper[4885]: I1002 01:48:09.567331 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:09 crc kubenswrapper[4885]: I1002 01:48:09.567396 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:09 crc kubenswrapper[4885]: I1002 01:48:09.567417 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:09 crc kubenswrapper[4885]: I1002 01:48:09.567443 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:09 crc kubenswrapper[4885]: I1002 01:48:09.567462 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:09Z","lastTransitionTime":"2025-10-02T01:48:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:09 crc kubenswrapper[4885]: I1002 01:48:09.670756 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:09 crc kubenswrapper[4885]: I1002 01:48:09.670819 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:09 crc kubenswrapper[4885]: I1002 01:48:09.670836 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:09 crc kubenswrapper[4885]: I1002 01:48:09.670860 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:09 crc kubenswrapper[4885]: I1002 01:48:09.670879 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:09Z","lastTransitionTime":"2025-10-02T01:48:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:09 crc kubenswrapper[4885]: I1002 01:48:09.773687 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:09 crc kubenswrapper[4885]: I1002 01:48:09.773757 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:09 crc kubenswrapper[4885]: I1002 01:48:09.773776 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:09 crc kubenswrapper[4885]: I1002 01:48:09.773804 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:09 crc kubenswrapper[4885]: I1002 01:48:09.773823 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:09Z","lastTransitionTime":"2025-10-02T01:48:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:09 crc kubenswrapper[4885]: I1002 01:48:09.786432 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:09 crc kubenswrapper[4885]: I1002 01:48:09.786494 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:09 crc kubenswrapper[4885]: I1002 01:48:09.786516 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:09 crc kubenswrapper[4885]: I1002 01:48:09.786547 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:09 crc kubenswrapper[4885]: I1002 01:48:09.786571 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:09Z","lastTransitionTime":"2025-10-02T01:48:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:09 crc kubenswrapper[4885]: E1002 01:48:09.806928 4885 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:48:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:48:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:48:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:48:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:48:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:48:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:48:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:48:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"92603ae5-ba38-4201-98e9-979fd56cb50d\\\",\\\"systemUUID\\\":\\\"50ddf2c2-ea95-4960-ae48-d17e9484ee2f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:09Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:09 crc kubenswrapper[4885]: I1002 01:48:09.811672 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:09 crc kubenswrapper[4885]: I1002 01:48:09.811734 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:09 crc kubenswrapper[4885]: I1002 01:48:09.811754 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:09 crc kubenswrapper[4885]: I1002 01:48:09.811783 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:09 crc kubenswrapper[4885]: I1002 01:48:09.811803 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:09Z","lastTransitionTime":"2025-10-02T01:48:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:09 crc kubenswrapper[4885]: E1002 01:48:09.831322 4885 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:48:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:48:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:48:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:48:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:48:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:48:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:48:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:48:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"92603ae5-ba38-4201-98e9-979fd56cb50d\\\",\\\"systemUUID\\\":\\\"50ddf2c2-ea95-4960-ae48-d17e9484ee2f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:09Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:09 crc kubenswrapper[4885]: I1002 01:48:09.835944 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:09 crc kubenswrapper[4885]: I1002 01:48:09.836013 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:09 crc kubenswrapper[4885]: I1002 01:48:09.836037 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:09 crc kubenswrapper[4885]: I1002 01:48:09.836063 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:09 crc kubenswrapper[4885]: I1002 01:48:09.836082 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:09Z","lastTransitionTime":"2025-10-02T01:48:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:09 crc kubenswrapper[4885]: E1002 01:48:09.856225 4885 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:48:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:48:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:48:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:48:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:48:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:48:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:48:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:48:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"92603ae5-ba38-4201-98e9-979fd56cb50d\\\",\\\"systemUUID\\\":\\\"50ddf2c2-ea95-4960-ae48-d17e9484ee2f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:09Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:09 crc kubenswrapper[4885]: I1002 01:48:09.861712 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:09 crc kubenswrapper[4885]: I1002 01:48:09.861768 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:09 crc kubenswrapper[4885]: I1002 01:48:09.861791 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:09 crc kubenswrapper[4885]: I1002 01:48:09.861816 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:09 crc kubenswrapper[4885]: I1002 01:48:09.861834 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:09Z","lastTransitionTime":"2025-10-02T01:48:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:09 crc kubenswrapper[4885]: E1002 01:48:09.881663 4885 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:48:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:48:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:48:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:48:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:48:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:48:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:48:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:48:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"92603ae5-ba38-4201-98e9-979fd56cb50d\\\",\\\"systemUUID\\\":\\\"50ddf2c2-ea95-4960-ae48-d17e9484ee2f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:09Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:09 crc kubenswrapper[4885]: I1002 01:48:09.886372 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:09 crc kubenswrapper[4885]: I1002 01:48:09.886423 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:09 crc kubenswrapper[4885]: I1002 01:48:09.886443 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:09 crc kubenswrapper[4885]: I1002 01:48:09.886465 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:09 crc kubenswrapper[4885]: I1002 01:48:09.886480 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:09Z","lastTransitionTime":"2025-10-02T01:48:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:09 crc kubenswrapper[4885]: E1002 01:48:09.906504 4885 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:48:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:48:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:48:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:48:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:48:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:48:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:48:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:48:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"92603ae5-ba38-4201-98e9-979fd56cb50d\\\",\\\"systemUUID\\\":\\\"50ddf2c2-ea95-4960-ae48-d17e9484ee2f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:09Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:09 crc kubenswrapper[4885]: E1002 01:48:09.906727 4885 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 02 01:48:09 crc kubenswrapper[4885]: I1002 01:48:09.908635 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:09 crc kubenswrapper[4885]: I1002 01:48:09.908691 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:09 crc kubenswrapper[4885]: I1002 01:48:09.908709 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:09 crc kubenswrapper[4885]: I1002 01:48:09.908755 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:09 crc kubenswrapper[4885]: I1002 01:48:09.908774 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:09Z","lastTransitionTime":"2025-10-02T01:48:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:10 crc kubenswrapper[4885]: I1002 01:48:10.011405 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:10 crc kubenswrapper[4885]: I1002 01:48:10.011472 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:10 crc kubenswrapper[4885]: I1002 01:48:10.011492 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:10 crc kubenswrapper[4885]: I1002 01:48:10.011521 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:10 crc kubenswrapper[4885]: I1002 01:48:10.011655 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:10Z","lastTransitionTime":"2025-10-02T01:48:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:10 crc kubenswrapper[4885]: I1002 01:48:10.046670 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:48:10 crc kubenswrapper[4885]: I1002 01:48:10.046721 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:48:10 crc kubenswrapper[4885]: I1002 01:48:10.046832 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:48:10 crc kubenswrapper[4885]: E1002 01:48:10.046904 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:48:10 crc kubenswrapper[4885]: I1002 01:48:10.046932 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sv4md" Oct 02 01:48:10 crc kubenswrapper[4885]: E1002 01:48:10.047047 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:48:10 crc kubenswrapper[4885]: E1002 01:48:10.047133 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sv4md" podUID="724a3e49-001c-45a4-a896-82c13e0ee0e7" Oct 02 01:48:10 crc kubenswrapper[4885]: I1002 01:48:10.047208 4885 scope.go:117] "RemoveContainer" containerID="188cb50b58926683ce8862d618258c1d1a6f33310608546c8651ece28f7c338f" Oct 02 01:48:10 crc kubenswrapper[4885]: E1002 01:48:10.047352 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:48:10 crc kubenswrapper[4885]: I1002 01:48:10.114549 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:10 crc kubenswrapper[4885]: I1002 01:48:10.114608 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:10 crc kubenswrapper[4885]: I1002 01:48:10.114625 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:10 crc kubenswrapper[4885]: I1002 01:48:10.114649 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:10 crc kubenswrapper[4885]: I1002 01:48:10.114665 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:10Z","lastTransitionTime":"2025-10-02T01:48:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:10 crc kubenswrapper[4885]: I1002 01:48:10.217936 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:10 crc kubenswrapper[4885]: I1002 01:48:10.218002 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:10 crc kubenswrapper[4885]: I1002 01:48:10.218019 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:10 crc kubenswrapper[4885]: I1002 01:48:10.218045 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:10 crc kubenswrapper[4885]: I1002 01:48:10.218063 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:10Z","lastTransitionTime":"2025-10-02T01:48:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:10 crc kubenswrapper[4885]: I1002 01:48:10.321338 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:10 crc kubenswrapper[4885]: I1002 01:48:10.321399 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:10 crc kubenswrapper[4885]: I1002 01:48:10.321417 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:10 crc kubenswrapper[4885]: I1002 01:48:10.321443 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:10 crc kubenswrapper[4885]: I1002 01:48:10.321461 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:10Z","lastTransitionTime":"2025-10-02T01:48:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:10 crc kubenswrapper[4885]: I1002 01:48:10.424767 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:10 crc kubenswrapper[4885]: I1002 01:48:10.424818 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:10 crc kubenswrapper[4885]: I1002 01:48:10.424835 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:10 crc kubenswrapper[4885]: I1002 01:48:10.424858 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:10 crc kubenswrapper[4885]: I1002 01:48:10.424877 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:10Z","lastTransitionTime":"2025-10-02T01:48:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:10 crc kubenswrapper[4885]: I1002 01:48:10.528368 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:10 crc kubenswrapper[4885]: I1002 01:48:10.528426 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:10 crc kubenswrapper[4885]: I1002 01:48:10.528443 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:10 crc kubenswrapper[4885]: I1002 01:48:10.528466 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:10 crc kubenswrapper[4885]: I1002 01:48:10.528483 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:10Z","lastTransitionTime":"2025-10-02T01:48:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:10 crc kubenswrapper[4885]: I1002 01:48:10.617121 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dngcm_9e4679dd-f870-41e9-a32b-360f5d2ee81b/ovnkube-controller/2.log" Oct 02 01:48:10 crc kubenswrapper[4885]: I1002 01:48:10.620089 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" event={"ID":"9e4679dd-f870-41e9-a32b-360f5d2ee81b","Type":"ContainerStarted","Data":"2ef80e838accc59ad7a579c66f4fcd6881d23fa86fc5922d4b0085d211d5d94b"} Oct 02 01:48:10 crc kubenswrapper[4885]: I1002 01:48:10.621279 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" Oct 02 01:48:10 crc kubenswrapper[4885]: I1002 01:48:10.630154 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:10 crc kubenswrapper[4885]: I1002 01:48:10.630194 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:10 crc kubenswrapper[4885]: I1002 01:48:10.630205 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:10 crc kubenswrapper[4885]: I1002 01:48:10.630219 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:10 crc kubenswrapper[4885]: I1002 01:48:10.630231 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:10Z","lastTransitionTime":"2025-10-02T01:48:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:10 crc kubenswrapper[4885]: I1002 01:48:10.636326 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"506a599f-fb1c-4d14-be88-0e4833f3d624\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31c2f0193ffef54965b79a3569b0f6f4547e540bac0774155159a1c996c196c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://533e73e5f90b7258acbcdf018ee01323c118f9dc45035b25b065ba01cabecccf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c00db833a593ec40eb63a6b383455e19e960c2e98b9b66f76877ee993057e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21e2f462a0deb0d740185663b2014f4a1e57b4fdbeb82571264b002f928035bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:10Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:10 crc kubenswrapper[4885]: I1002 01:48:10.660033 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e4679dd-f870-41e9-a32b-360f5d2ee81b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4310909ad9c7b609ed6814e04a525a6afae747775598e7275080413dec8b23a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7642033b0b5995d332ccf7992d6d7e6a5e9c9369ae2d1071ccb2f536938c8f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9f99f9e26a04fb42b01a0783a5c4e360514818ce9102edc7fa08626d50b9a34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6fa72b030d309656b73121e4ac2a3924a4e449aad66683dd9de5ee5f22d5a4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef403e519d197fb37b3e6b76a14a5128ba3f30a7a5e8bf0a420cf6a39b89575f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8482ba457683ffa03100ec1f07ea9d3f918be5e490b5ba05c2ba965d03033191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ef80e838accc59ad7a579c66f4fcd6881d23fa86fc5922d4b0085d211d5d94b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://188cb50b58926683ce8862d618258c1d1a6f33310608546c8651ece28f7c338f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T01:47:41Z\\\",\\\"message\\\":\\\"921 6578 obj_retry.go:434] periodicallyRetryResources: Retry channel got triggered: retrying failed objects of type *v1.Pod\\\\nI1002 01:47:40.991945 6578 obj_retry.go:409] Going to retry *v1.Pod resource setup for 16 objects: [openshift-multus/network-metrics-daemon-sv4md openshift-network-operator/iptables-alerter-4ln5h openshift-kube-apiserver/kube-apiserver-crc openshift-ovn-kubernetes/ovnkube-node-dngcm openshift-dns/node-resolver-58lhj openshift-machine-config-operator/machine-config-daemon-rttx8 openshift-multus/multus-vhgbd openshift-network-console/networking-console-plugin-85b44fc459-gdk6g openshift-network-diagnostics/network-check-source-55646444c4-trplf openshift-network-operator/network-operator-58b4c7f79c-55gtf openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rrn8b openshift-image-registry/node-ca-jgdvt openshift-kube-controller-manager/kube-controller-manager-crc openshift-multus/multus-additional-cni-plugins-2ll9b openshift-network-diagnostics/network-check-target-xd92c openshift-network-node-identity/network-node-identity-vrzqb]\\\\nF1002 01:47:40.991998 6578 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller ini\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:40Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:48:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fafdd1094bd02786d745c984989852b0c690db38655f92b419733a2004f417c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0125e907d061a97286026ee823c4eed303734ed03ab4864f742592b23a1aa2f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0125e907d061a97286026ee823c4eed303734ed03ab4864f742592b23a1aa2f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dngcm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:10Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:10 crc kubenswrapper[4885]: I1002 01:48:10.674703 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-sv4md" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"724a3e49-001c-45a4-a896-82c13e0ee0e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hpqgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hpqgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-sv4md\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:10Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:10 crc kubenswrapper[4885]: I1002 01:48:10.690798 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rrn8b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52967a77-525c-4537-9b1f-d7bb754a8494\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3dee451ebe9006bf1c519fed4c65267570b6139d76841858352e087ae78d26a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtr74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e9412788d6df7332592beac48281822b26e1aa19b39addce8e1be556af9d982\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtr74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:26Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rrn8b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:10Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:10 crc kubenswrapper[4885]: I1002 01:48:10.706041 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18466979-7c7c-4d73-86c6-b8b6eaf20484\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b742e4b288bb50569d2fd23f3f9857d649a46063998170eee582ecbbba64d190\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://298c2ace58c4a9d09cfdc864316317e370bb0ac5735d631af6d85c6298f686d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://620f54e02c463f6e603808e5e57e9e5770d125578cc258ba16117a2416e57d3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b4cd17b64e5657ae05e71f3ca2340cb0d7965b1b49d1f7b4bfe4494292293ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b4cd17b64e5657ae05e71f3ca2340cb0d7965b1b49d1f7b4bfe4494292293ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:10Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:10 crc kubenswrapper[4885]: I1002 01:48:10.733303 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:10 crc kubenswrapper[4885]: I1002 01:48:10.733350 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:10 crc kubenswrapper[4885]: I1002 01:48:10.733361 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:10 crc kubenswrapper[4885]: I1002 01:48:10.733377 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:10 crc kubenswrapper[4885]: I1002 01:48:10.733757 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:10Z","lastTransitionTime":"2025-10-02T01:48:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:10 crc kubenswrapper[4885]: I1002 01:48:10.738034 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8070bc1c-e3de-4e9a-985e-40d1a9858056\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3e634b3727c517cc48b17c2cfb38f26460f53d1bf25d631d08ec9acb8d780e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4adcc537c23697a324d82af94c31059dd444245efdf7c978db58ede256e112ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edcd1d99644a778dfb6d0a17d2a4abd906400fe8209e611ebf8d7a9abe7ccf28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://97cb59adc2ddfa365ebc9dec649600dc90466ffe59134a6d15a9d3134a7c9b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d2808bc658ea80d68469ed16422d56ef19447f4c90b4af7b2597c64406f8a44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://83acbc20919f7c1958b6d492fadb203855df02761662425df9f91a0757fe2bf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83acbc20919f7c1958b6d492fadb203855df02761662425df9f91a0757fe2bf7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e86ecd03835efc9a253b6804bc61e6ac690e75b8d1f3cc77f5647730e586574\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e86ecd03835efc9a253b6804bc61e6ac690e75b8d1f3cc77f5647730e586574\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://48f95af51feaf014a18496bdfadd59d7f1e11167c5c8b677fa7037e48368ff2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48f95af51feaf014a18496bdfadd59d7f1e11167c5c8b677fa7037e48368ff2a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:10Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:10 crc kubenswrapper[4885]: I1002 01:48:10.753108 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5c1f95f188569c9aa47e51f2916939fa0b65a5cdf4c16b798612d41f3c8411a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:10Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:10 crc kubenswrapper[4885]: I1002 01:48:10.769107 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vhgbd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df9792d80c807ed98c8442822d01f619327361c40425abfe26f11e5c27ba2584\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c63a7d20d39b65acfb5a9be4c28b21c72111b398bea76c238e15ddb4dd63e139\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T01:48:00Z\\\",\\\"message\\\":\\\"2025-10-02T01:47:15+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_9f42b0ef-8651-49db-8a31-d21c0c1c47e7\\\\n2025-10-02T01:47:15+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_9f42b0ef-8651-49db-8a31-d21c0c1c47e7 to /host/opt/cni/bin/\\\\n2025-10-02T01:47:15Z [verbose] multus-daemon started\\\\n2025-10-02T01:47:15Z [verbose] Readiness Indicator file check\\\\n2025-10-02T01:48:00Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7zwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vhgbd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:10Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:10 crc kubenswrapper[4885]: I1002 01:48:10.788373 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:10Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:10 crc kubenswrapper[4885]: I1002 01:48:10.808727 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74adcf912ccedb6b5d4f2803fd215e3e2ef62367597aa6da528354c4511ee536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7403ef968d040ac09d722b7557438bef23892d578992826f42b700b370d495d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:10Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:10 crc kubenswrapper[4885]: I1002 01:48:10.826406 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2ll9b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52b39712-5fa1-4fe0-814b-b170a6e3938b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9a96383d0fbfd116c419054aaf1a7adba1937e8b0bad294cefd68b4ab0d2298\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c979b91ba748aa4647826a8926337662ca929c54090fb62285572819ec2fbf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c979b91ba748aa4647826a8926337662ca929c54090fb62285572819ec2fbf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97d9d693e8e60ee9c183bb44461e698e8b3cfd20a516605aa8754c9205d2bdc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://97d9d693e8e60ee9c183bb44461e698e8b3cfd20a516605aa8754c9205d2bdc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c44252676359c671d65c9588b05572d76c99cf4c90666f0ac089b2f778450a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c44252676359c671d65c9588b05572d76c99cf4c90666f0ac089b2f778450a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80c982f07874df686fe7e29cc91043eb6d58f1be5fc93cacd46d33a21dd9b622\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80c982f07874df686fe7e29cc91043eb6d58f1be5fc93cacd46d33a21dd9b622\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a354c985652c7ad7fa8944fbfdfe223cfb0f97fb11eecd9bd429020d4892b78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a354c985652c7ad7fa8944fbfdfe223cfb0f97fb11eecd9bd429020d4892b78\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2efd7283df313d086c1b8b88f289ace28764de256205bee7610d6711002d8515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2efd7283df313d086c1b8b88f289ace28764de256205bee7610d6711002d8515\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2ll9b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:10Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:10 crc kubenswrapper[4885]: I1002 01:48:10.837060 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:10 crc kubenswrapper[4885]: I1002 01:48:10.837104 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:10 crc kubenswrapper[4885]: I1002 01:48:10.837117 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:10 crc kubenswrapper[4885]: I1002 01:48:10.837134 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:10 crc kubenswrapper[4885]: I1002 01:48:10.837146 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:10Z","lastTransitionTime":"2025-10-02T01:48:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:10 crc kubenswrapper[4885]: I1002 01:48:10.843725 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:10Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:10 crc kubenswrapper[4885]: I1002 01:48:10.858725 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-58lhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac6a19c4-f259-4dbd-a016-d6c2c849c6cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed76c0f968c063f7795d363af194a222487b647ceda234de0dabb39b25eafe0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hzc4c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-58lhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:10Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:10 crc kubenswrapper[4885]: I1002 01:48:10.873023 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d169c8ba-000a-4218-8e3b-4ae53035b110\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61e436b7876107ce5ea9e668c0bbc367be6e663fc804fa3a65b51adf8be8d51f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-77rzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09536e182a46e77318c10812d1777e551f834907c006cab5c726dc66f5bce88a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-77rzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rttx8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:10Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:10 crc kubenswrapper[4885]: I1002 01:48:10.884406 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jgdvt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81733e8d-67b0-46fa-8a41-08b40645bd3e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcef60c816ca1ba9a298da658e84c4d717bea3d352c4c2f807ebd8e56327b384\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cwbl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jgdvt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:10Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:10 crc kubenswrapper[4885]: I1002 01:48:10.908022 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a0bb885-2eb0-419c-8a7a-ef634cb3c144\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a89b49cbe08d927629564e607de13655ddf3d01ac4eb5d688b325210306c120c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06c5942805a498d9621a906088aa54538b4eed583d1b8d0e3ba2a6718378a36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b9f6f7fe44d5771a838d0b89c5bdea3ebb79023ae2675ee907ec2766b3d35a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://178b4340ee6904c60c814b99d83ee225161aeb23e438d8d02f856579a13ec4e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bb63bf30d6ecce9e71f0db0c16632bdfa6b33924b673378eb0c38ab229fa837\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T01:47:06Z\\\",\\\"message\\\":\\\"W1002 01:46:55.406027 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1002 01:46:55.406714 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759369615 cert, and key in /tmp/serving-cert-3795040473/serving-signer.crt, /tmp/serving-cert-3795040473/serving-signer.key\\\\nI1002 01:46:55.614670 1 observer_polling.go:159] Starting file observer\\\\nW1002 01:46:55.617662 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1002 01:46:55.618112 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 01:46:55.620347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3795040473/tls.crt::/tmp/serving-cert-3795040473/tls.key\\\\\\\"\\\\nF1002 01:47:06.224459 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90535f97ab2ed4626d47540ebd7baada837fd332140323bec0f49aa38d3c0009\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://592f89a326c9c1559193a71904b47ca27b1dbe1a7b30e03bc48924bb2f35f921\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://592f89a326c9c1559193a71904b47ca27b1dbe1a7b30e03bc48924bb2f35f921\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:10Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:10 crc kubenswrapper[4885]: I1002 01:48:10.926309 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b087880e05039816dcab154643ac72c7e4a0a1124817c51dc2c639b62bbee0fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:10Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:10 crc kubenswrapper[4885]: I1002 01:48:10.940185 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:10 crc kubenswrapper[4885]: I1002 01:48:10.940248 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:10 crc kubenswrapper[4885]: I1002 01:48:10.940348 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:10 crc kubenswrapper[4885]: I1002 01:48:10.940376 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:10 crc kubenswrapper[4885]: I1002 01:48:10.940394 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:10Z","lastTransitionTime":"2025-10-02T01:48:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:10 crc kubenswrapper[4885]: I1002 01:48:10.941433 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:10Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:11 crc kubenswrapper[4885]: I1002 01:48:11.043155 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:11 crc kubenswrapper[4885]: I1002 01:48:11.043212 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:11 crc kubenswrapper[4885]: I1002 01:48:11.043233 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:11 crc kubenswrapper[4885]: I1002 01:48:11.043257 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:11 crc kubenswrapper[4885]: I1002 01:48:11.043300 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:11Z","lastTransitionTime":"2025-10-02T01:48:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:11 crc kubenswrapper[4885]: I1002 01:48:11.146536 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:11 crc kubenswrapper[4885]: I1002 01:48:11.146601 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:11 crc kubenswrapper[4885]: I1002 01:48:11.146619 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:11 crc kubenswrapper[4885]: I1002 01:48:11.146651 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:11 crc kubenswrapper[4885]: I1002 01:48:11.146670 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:11Z","lastTransitionTime":"2025-10-02T01:48:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:11 crc kubenswrapper[4885]: I1002 01:48:11.249962 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:11 crc kubenswrapper[4885]: I1002 01:48:11.250027 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:11 crc kubenswrapper[4885]: I1002 01:48:11.250048 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:11 crc kubenswrapper[4885]: I1002 01:48:11.250072 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:11 crc kubenswrapper[4885]: I1002 01:48:11.250089 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:11Z","lastTransitionTime":"2025-10-02T01:48:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:11 crc kubenswrapper[4885]: I1002 01:48:11.353503 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:11 crc kubenswrapper[4885]: I1002 01:48:11.353544 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:11 crc kubenswrapper[4885]: I1002 01:48:11.353561 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:11 crc kubenswrapper[4885]: I1002 01:48:11.353580 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:11 crc kubenswrapper[4885]: I1002 01:48:11.353596 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:11Z","lastTransitionTime":"2025-10-02T01:48:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:11 crc kubenswrapper[4885]: I1002 01:48:11.457489 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:11 crc kubenswrapper[4885]: I1002 01:48:11.457543 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:11 crc kubenswrapper[4885]: I1002 01:48:11.457576 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:11 crc kubenswrapper[4885]: I1002 01:48:11.457599 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:11 crc kubenswrapper[4885]: I1002 01:48:11.457615 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:11Z","lastTransitionTime":"2025-10-02T01:48:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:11 crc kubenswrapper[4885]: I1002 01:48:11.561224 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:11 crc kubenswrapper[4885]: I1002 01:48:11.561301 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:11 crc kubenswrapper[4885]: I1002 01:48:11.561318 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:11 crc kubenswrapper[4885]: I1002 01:48:11.561340 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:11 crc kubenswrapper[4885]: I1002 01:48:11.561358 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:11Z","lastTransitionTime":"2025-10-02T01:48:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:11 crc kubenswrapper[4885]: I1002 01:48:11.626917 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dngcm_9e4679dd-f870-41e9-a32b-360f5d2ee81b/ovnkube-controller/3.log" Oct 02 01:48:11 crc kubenswrapper[4885]: I1002 01:48:11.627811 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dngcm_9e4679dd-f870-41e9-a32b-360f5d2ee81b/ovnkube-controller/2.log" Oct 02 01:48:11 crc kubenswrapper[4885]: I1002 01:48:11.631151 4885 generic.go:334] "Generic (PLEG): container finished" podID="9e4679dd-f870-41e9-a32b-360f5d2ee81b" containerID="2ef80e838accc59ad7a579c66f4fcd6881d23fa86fc5922d4b0085d211d5d94b" exitCode=1 Oct 02 01:48:11 crc kubenswrapper[4885]: I1002 01:48:11.631198 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" event={"ID":"9e4679dd-f870-41e9-a32b-360f5d2ee81b","Type":"ContainerDied","Data":"2ef80e838accc59ad7a579c66f4fcd6881d23fa86fc5922d4b0085d211d5d94b"} Oct 02 01:48:11 crc kubenswrapper[4885]: I1002 01:48:11.631243 4885 scope.go:117] "RemoveContainer" containerID="188cb50b58926683ce8862d618258c1d1a6f33310608546c8651ece28f7c338f" Oct 02 01:48:11 crc kubenswrapper[4885]: I1002 01:48:11.632329 4885 scope.go:117] "RemoveContainer" containerID="2ef80e838accc59ad7a579c66f4fcd6881d23fa86fc5922d4b0085d211d5d94b" Oct 02 01:48:11 crc kubenswrapper[4885]: E1002 01:48:11.632559 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-dngcm_openshift-ovn-kubernetes(9e4679dd-f870-41e9-a32b-360f5d2ee81b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" podUID="9e4679dd-f870-41e9-a32b-360f5d2ee81b" Oct 02 01:48:11 crc kubenswrapper[4885]: I1002 01:48:11.652933 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:11Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:11 crc kubenswrapper[4885]: I1002 01:48:11.663825 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:11 crc kubenswrapper[4885]: I1002 01:48:11.663914 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:11 crc kubenswrapper[4885]: I1002 01:48:11.663940 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:11 crc kubenswrapper[4885]: I1002 01:48:11.663972 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:11 crc kubenswrapper[4885]: I1002 01:48:11.664002 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:11Z","lastTransitionTime":"2025-10-02T01:48:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:11 crc kubenswrapper[4885]: I1002 01:48:11.673931 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74adcf912ccedb6b5d4f2803fd215e3e2ef62367597aa6da528354c4511ee536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7403ef968d040ac09d722b7557438bef23892d578992826f42b700b370d495d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:11Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:11 crc kubenswrapper[4885]: I1002 01:48:11.698177 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2ll9b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52b39712-5fa1-4fe0-814b-b170a6e3938b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9a96383d0fbfd116c419054aaf1a7adba1937e8b0bad294cefd68b4ab0d2298\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c979b91ba748aa4647826a8926337662ca929c54090fb62285572819ec2fbf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c979b91ba748aa4647826a8926337662ca929c54090fb62285572819ec2fbf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97d9d693e8e60ee9c183bb44461e698e8b3cfd20a516605aa8754c9205d2bdc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://97d9d693e8e60ee9c183bb44461e698e8b3cfd20a516605aa8754c9205d2bdc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c44252676359c671d65c9588b05572d76c99cf4c90666f0ac089b2f778450a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c44252676359c671d65c9588b05572d76c99cf4c90666f0ac089b2f778450a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80c982f07874df686fe7e29cc91043eb6d58f1be5fc93cacd46d33a21dd9b622\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80c982f07874df686fe7e29cc91043eb6d58f1be5fc93cacd46d33a21dd9b622\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a354c985652c7ad7fa8944fbfdfe223cfb0f97fb11eecd9bd429020d4892b78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a354c985652c7ad7fa8944fbfdfe223cfb0f97fb11eecd9bd429020d4892b78\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2efd7283df313d086c1b8b88f289ace28764de256205bee7610d6711002d8515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2efd7283df313d086c1b8b88f289ace28764de256205bee7610d6711002d8515\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2ll9b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:11Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:11 crc kubenswrapper[4885]: I1002 01:48:11.719663 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a0bb885-2eb0-419c-8a7a-ef634cb3c144\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a89b49cbe08d927629564e607de13655ddf3d01ac4eb5d688b325210306c120c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06c5942805a498d9621a906088aa54538b4eed583d1b8d0e3ba2a6718378a36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b9f6f7fe44d5771a838d0b89c5bdea3ebb79023ae2675ee907ec2766b3d35a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://178b4340ee6904c60c814b99d83ee225161aeb23e438d8d02f856579a13ec4e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bb63bf30d6ecce9e71f0db0c16632bdfa6b33924b673378eb0c38ab229fa837\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T01:47:06Z\\\",\\\"message\\\":\\\"W1002 01:46:55.406027 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1002 01:46:55.406714 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759369615 cert, and key in /tmp/serving-cert-3795040473/serving-signer.crt, /tmp/serving-cert-3795040473/serving-signer.key\\\\nI1002 01:46:55.614670 1 observer_polling.go:159] Starting file observer\\\\nW1002 01:46:55.617662 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1002 01:46:55.618112 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 01:46:55.620347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3795040473/tls.crt::/tmp/serving-cert-3795040473/tls.key\\\\\\\"\\\\nF1002 01:47:06.224459 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90535f97ab2ed4626d47540ebd7baada837fd332140323bec0f49aa38d3c0009\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://592f89a326c9c1559193a71904b47ca27b1dbe1a7b30e03bc48924bb2f35f921\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://592f89a326c9c1559193a71904b47ca27b1dbe1a7b30e03bc48924bb2f35f921\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:11Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:11 crc kubenswrapper[4885]: I1002 01:48:11.739408 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b087880e05039816dcab154643ac72c7e4a0a1124817c51dc2c639b62bbee0fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:11Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:11 crc kubenswrapper[4885]: I1002 01:48:11.759162 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:11Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:11 crc kubenswrapper[4885]: I1002 01:48:11.766841 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:11 crc kubenswrapper[4885]: I1002 01:48:11.766893 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:11 crc kubenswrapper[4885]: I1002 01:48:11.766912 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:11 crc kubenswrapper[4885]: I1002 01:48:11.766936 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:11 crc kubenswrapper[4885]: I1002 01:48:11.766954 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:11Z","lastTransitionTime":"2025-10-02T01:48:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:11 crc kubenswrapper[4885]: I1002 01:48:11.777321 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:11Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:11 crc kubenswrapper[4885]: I1002 01:48:11.792212 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-58lhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac6a19c4-f259-4dbd-a016-d6c2c849c6cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed76c0f968c063f7795d363af194a222487b647ceda234de0dabb39b25eafe0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hzc4c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-58lhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:11Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:11 crc kubenswrapper[4885]: I1002 01:48:11.809326 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d169c8ba-000a-4218-8e3b-4ae53035b110\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61e436b7876107ce5ea9e668c0bbc367be6e663fc804fa3a65b51adf8be8d51f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-77rzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09536e182a46e77318c10812d1777e551f834907c006cab5c726dc66f5bce88a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-77rzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rttx8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:11Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:11 crc kubenswrapper[4885]: I1002 01:48:11.825193 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jgdvt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81733e8d-67b0-46fa-8a41-08b40645bd3e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcef60c816ca1ba9a298da658e84c4d717bea3d352c4c2f807ebd8e56327b384\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cwbl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jgdvt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:11Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:11 crc kubenswrapper[4885]: I1002 01:48:11.845053 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"506a599f-fb1c-4d14-be88-0e4833f3d624\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31c2f0193ffef54965b79a3569b0f6f4547e540bac0774155159a1c996c196c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://533e73e5f90b7258acbcdf018ee01323c118f9dc45035b25b065ba01cabecccf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c00db833a593ec40eb63a6b383455e19e960c2e98b9b66f76877ee993057e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21e2f462a0deb0d740185663b2014f4a1e57b4fdbeb82571264b002f928035bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:11Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:11 crc kubenswrapper[4885]: I1002 01:48:11.870642 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:11 crc kubenswrapper[4885]: I1002 01:48:11.870707 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:11 crc kubenswrapper[4885]: I1002 01:48:11.870725 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:11 crc kubenswrapper[4885]: I1002 01:48:11.870749 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:11 crc kubenswrapper[4885]: I1002 01:48:11.870767 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:11Z","lastTransitionTime":"2025-10-02T01:48:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:11 crc kubenswrapper[4885]: I1002 01:48:11.881899 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e4679dd-f870-41e9-a32b-360f5d2ee81b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4310909ad9c7b609ed6814e04a525a6afae747775598e7275080413dec8b23a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7642033b0b5995d332ccf7992d6d7e6a5e9c9369ae2d1071ccb2f536938c8f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9f99f9e26a04fb42b01a0783a5c4e360514818ce9102edc7fa08626d50b9a34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6fa72b030d309656b73121e4ac2a3924a4e449aad66683dd9de5ee5f22d5a4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef403e519d197fb37b3e6b76a14a5128ba3f30a7a5e8bf0a420cf6a39b89575f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8482ba457683ffa03100ec1f07ea9d3f918be5e490b5ba05c2ba965d03033191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ef80e838accc59ad7a579c66f4fcd6881d23fa86fc5922d4b0085d211d5d94b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://188cb50b58926683ce8862d618258c1d1a6f33310608546c8651ece28f7c338f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T01:47:41Z\\\",\\\"message\\\":\\\"921 6578 obj_retry.go:434] periodicallyRetryResources: Retry channel got triggered: retrying failed objects of type *v1.Pod\\\\nI1002 01:47:40.991945 6578 obj_retry.go:409] Going to retry *v1.Pod resource setup for 16 objects: [openshift-multus/network-metrics-daemon-sv4md openshift-network-operator/iptables-alerter-4ln5h openshift-kube-apiserver/kube-apiserver-crc openshift-ovn-kubernetes/ovnkube-node-dngcm openshift-dns/node-resolver-58lhj openshift-machine-config-operator/machine-config-daemon-rttx8 openshift-multus/multus-vhgbd openshift-network-console/networking-console-plugin-85b44fc459-gdk6g openshift-network-diagnostics/network-check-source-55646444c4-trplf openshift-network-operator/network-operator-58b4c7f79c-55gtf openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rrn8b openshift-image-registry/node-ca-jgdvt openshift-kube-controller-manager/kube-controller-manager-crc openshift-multus/multus-additional-cni-plugins-2ll9b openshift-network-diagnostics/network-check-target-xd92c openshift-network-node-identity/network-node-identity-vrzqb]\\\\nF1002 01:47:40.991998 6578 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller ini\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:40Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ef80e838accc59ad7a579c66f4fcd6881d23fa86fc5922d4b0085d211d5d94b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T01:48:11Z\\\",\\\"message\\\":\\\" 6940 services_controller.go:451] Built service openshift-operator-lifecycle-manager/catalog-operator-metrics cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-operator-lifecycle-manager/catalog-operator-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-operator-lifecycle-manager/catalog-operator-metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.204\\\\\\\", Port:8443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1002 01:48:11.048667 6940 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:48:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fafdd1094bd02786d745c984989852b0c690db38655f92b419733a2004f417c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0125e907d061a97286026ee823c4eed303734ed03ab4864f742592b23a1aa2f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0125e907d061a97286026ee823c4eed303734ed03ab4864f742592b23a1aa2f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dngcm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:11Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:11 crc kubenswrapper[4885]: I1002 01:48:11.898388 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-sv4md" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"724a3e49-001c-45a4-a896-82c13e0ee0e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hpqgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hpqgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-sv4md\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:11Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:11 crc kubenswrapper[4885]: I1002 01:48:11.915214 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18466979-7c7c-4d73-86c6-b8b6eaf20484\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b742e4b288bb50569d2fd23f3f9857d649a46063998170eee582ecbbba64d190\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://298c2ace58c4a9d09cfdc864316317e370bb0ac5735d631af6d85c6298f686d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://620f54e02c463f6e603808e5e57e9e5770d125578cc258ba16117a2416e57d3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b4cd17b64e5657ae05e71f3ca2340cb0d7965b1b49d1f7b4bfe4494292293ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b4cd17b64e5657ae05e71f3ca2340cb0d7965b1b49d1f7b4bfe4494292293ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:11Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:11 crc kubenswrapper[4885]: I1002 01:48:11.945054 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8070bc1c-e3de-4e9a-985e-40d1a9858056\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3e634b3727c517cc48b17c2cfb38f26460f53d1bf25d631d08ec9acb8d780e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4adcc537c23697a324d82af94c31059dd444245efdf7c978db58ede256e112ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edcd1d99644a778dfb6d0a17d2a4abd906400fe8209e611ebf8d7a9abe7ccf28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://97cb59adc2ddfa365ebc9dec649600dc90466ffe59134a6d15a9d3134a7c9b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d2808bc658ea80d68469ed16422d56ef19447f4c90b4af7b2597c64406f8a44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://83acbc20919f7c1958b6d492fadb203855df02761662425df9f91a0757fe2bf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83acbc20919f7c1958b6d492fadb203855df02761662425df9f91a0757fe2bf7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e86ecd03835efc9a253b6804bc61e6ac690e75b8d1f3cc77f5647730e586574\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e86ecd03835efc9a253b6804bc61e6ac690e75b8d1f3cc77f5647730e586574\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://48f95af51feaf014a18496bdfadd59d7f1e11167c5c8b677fa7037e48368ff2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48f95af51feaf014a18496bdfadd59d7f1e11167c5c8b677fa7037e48368ff2a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:11Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:11 crc kubenswrapper[4885]: I1002 01:48:11.960449 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5c1f95f188569c9aa47e51f2916939fa0b65a5cdf4c16b798612d41f3c8411a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:11Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:11 crc kubenswrapper[4885]: I1002 01:48:11.974224 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:11 crc kubenswrapper[4885]: I1002 01:48:11.974299 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:11 crc kubenswrapper[4885]: I1002 01:48:11.974318 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:11 crc kubenswrapper[4885]: I1002 01:48:11.974341 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:11 crc kubenswrapper[4885]: I1002 01:48:11.974361 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:11Z","lastTransitionTime":"2025-10-02T01:48:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:11 crc kubenswrapper[4885]: I1002 01:48:11.978134 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vhgbd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df9792d80c807ed98c8442822d01f619327361c40425abfe26f11e5c27ba2584\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c63a7d20d39b65acfb5a9be4c28b21c72111b398bea76c238e15ddb4dd63e139\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T01:48:00Z\\\",\\\"message\\\":\\\"2025-10-02T01:47:15+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_9f42b0ef-8651-49db-8a31-d21c0c1c47e7\\\\n2025-10-02T01:47:15+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_9f42b0ef-8651-49db-8a31-d21c0c1c47e7 to /host/opt/cni/bin/\\\\n2025-10-02T01:47:15Z [verbose] multus-daemon started\\\\n2025-10-02T01:47:15Z [verbose] Readiness Indicator file check\\\\n2025-10-02T01:48:00Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7zwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vhgbd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:11Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:11 crc kubenswrapper[4885]: I1002 01:48:11.995512 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rrn8b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52967a77-525c-4537-9b1f-d7bb754a8494\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3dee451ebe9006bf1c519fed4c65267570b6139d76841858352e087ae78d26a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtr74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e9412788d6df7332592beac48281822b26e1aa19b39addce8e1be556af9d982\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtr74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:26Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rrn8b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:11Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:12 crc kubenswrapper[4885]: I1002 01:48:12.046481 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:48:12 crc kubenswrapper[4885]: I1002 01:48:12.046499 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sv4md" Oct 02 01:48:12 crc kubenswrapper[4885]: I1002 01:48:12.046619 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:48:12 crc kubenswrapper[4885]: I1002 01:48:12.046685 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:48:12 crc kubenswrapper[4885]: E1002 01:48:12.046878 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:48:12 crc kubenswrapper[4885]: E1002 01:48:12.047067 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sv4md" podUID="724a3e49-001c-45a4-a896-82c13e0ee0e7" Oct 02 01:48:12 crc kubenswrapper[4885]: E1002 01:48:12.047223 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:48:12 crc kubenswrapper[4885]: E1002 01:48:12.047509 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:48:12 crc kubenswrapper[4885]: I1002 01:48:12.076907 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e4679dd-f870-41e9-a32b-360f5d2ee81b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4310909ad9c7b609ed6814e04a525a6afae747775598e7275080413dec8b23a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7642033b0b5995d332ccf7992d6d7e6a5e9c9369ae2d1071ccb2f536938c8f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9f99f9e26a04fb42b01a0783a5c4e360514818ce9102edc7fa08626d50b9a34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6fa72b030d309656b73121e4ac2a3924a4e449aad66683dd9de5ee5f22d5a4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef403e519d197fb37b3e6b76a14a5128ba3f30a7a5e8bf0a420cf6a39b89575f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8482ba457683ffa03100ec1f07ea9d3f918be5e490b5ba05c2ba965d03033191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ef80e838accc59ad7a579c66f4fcd6881d23fa86fc5922d4b0085d211d5d94b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://188cb50b58926683ce8862d618258c1d1a6f33310608546c8651ece28f7c338f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T01:47:41Z\\\",\\\"message\\\":\\\"921 6578 obj_retry.go:434] periodicallyRetryResources: Retry channel got triggered: retrying failed objects of type *v1.Pod\\\\nI1002 01:47:40.991945 6578 obj_retry.go:409] Going to retry *v1.Pod resource setup for 16 objects: [openshift-multus/network-metrics-daemon-sv4md openshift-network-operator/iptables-alerter-4ln5h openshift-kube-apiserver/kube-apiserver-crc openshift-ovn-kubernetes/ovnkube-node-dngcm openshift-dns/node-resolver-58lhj openshift-machine-config-operator/machine-config-daemon-rttx8 openshift-multus/multus-vhgbd openshift-network-console/networking-console-plugin-85b44fc459-gdk6g openshift-network-diagnostics/network-check-source-55646444c4-trplf openshift-network-operator/network-operator-58b4c7f79c-55gtf openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rrn8b openshift-image-registry/node-ca-jgdvt openshift-kube-controller-manager/kube-controller-manager-crc openshift-multus/multus-additional-cni-plugins-2ll9b openshift-network-diagnostics/network-check-target-xd92c openshift-network-node-identity/network-node-identity-vrzqb]\\\\nF1002 01:47:40.991998 6578 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller ini\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:40Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ef80e838accc59ad7a579c66f4fcd6881d23fa86fc5922d4b0085d211d5d94b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T01:48:11Z\\\",\\\"message\\\":\\\" 6940 services_controller.go:451] Built service openshift-operator-lifecycle-manager/catalog-operator-metrics cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-operator-lifecycle-manager/catalog-operator-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-operator-lifecycle-manager/catalog-operator-metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.204\\\\\\\", Port:8443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1002 01:48:11.048667 6940 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:48:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fafdd1094bd02786d745c984989852b0c690db38655f92b419733a2004f417c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0125e907d061a97286026ee823c4eed303734ed03ab4864f742592b23a1aa2f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0125e907d061a97286026ee823c4eed303734ed03ab4864f742592b23a1aa2f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dngcm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:12Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:12 crc kubenswrapper[4885]: I1002 01:48:12.077525 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:12 crc kubenswrapper[4885]: I1002 01:48:12.077571 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:12 crc kubenswrapper[4885]: I1002 01:48:12.077588 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:12 crc kubenswrapper[4885]: I1002 01:48:12.077609 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:12 crc kubenswrapper[4885]: I1002 01:48:12.077626 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:12Z","lastTransitionTime":"2025-10-02T01:48:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:12 crc kubenswrapper[4885]: I1002 01:48:12.093957 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-sv4md" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"724a3e49-001c-45a4-a896-82c13e0ee0e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hpqgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hpqgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-sv4md\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:12Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:12 crc kubenswrapper[4885]: I1002 01:48:12.112585 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"506a599f-fb1c-4d14-be88-0e4833f3d624\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31c2f0193ffef54965b79a3569b0f6f4547e540bac0774155159a1c996c196c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://533e73e5f90b7258acbcdf018ee01323c118f9dc45035b25b065ba01cabecccf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c00db833a593ec40eb63a6b383455e19e960c2e98b9b66f76877ee993057e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21e2f462a0deb0d740185663b2014f4a1e57b4fdbeb82571264b002f928035bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:12Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:12 crc kubenswrapper[4885]: I1002 01:48:12.143082 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8070bc1c-e3de-4e9a-985e-40d1a9858056\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3e634b3727c517cc48b17c2cfb38f26460f53d1bf25d631d08ec9acb8d780e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4adcc537c23697a324d82af94c31059dd444245efdf7c978db58ede256e112ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edcd1d99644a778dfb6d0a17d2a4abd906400fe8209e611ebf8d7a9abe7ccf28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://97cb59adc2ddfa365ebc9dec649600dc90466ffe59134a6d15a9d3134a7c9b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d2808bc658ea80d68469ed16422d56ef19447f4c90b4af7b2597c64406f8a44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://83acbc20919f7c1958b6d492fadb203855df02761662425df9f91a0757fe2bf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83acbc20919f7c1958b6d492fadb203855df02761662425df9f91a0757fe2bf7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e86ecd03835efc9a253b6804bc61e6ac690e75b8d1f3cc77f5647730e586574\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e86ecd03835efc9a253b6804bc61e6ac690e75b8d1f3cc77f5647730e586574\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://48f95af51feaf014a18496bdfadd59d7f1e11167c5c8b677fa7037e48368ff2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48f95af51feaf014a18496bdfadd59d7f1e11167c5c8b677fa7037e48368ff2a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:12Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:12 crc kubenswrapper[4885]: I1002 01:48:12.162066 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5c1f95f188569c9aa47e51f2916939fa0b65a5cdf4c16b798612d41f3c8411a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:12Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:12 crc kubenswrapper[4885]: I1002 01:48:12.183215 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:12 crc kubenswrapper[4885]: I1002 01:48:12.183344 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:12 crc kubenswrapper[4885]: I1002 01:48:12.183422 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:12 crc kubenswrapper[4885]: I1002 01:48:12.183449 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:12 crc kubenswrapper[4885]: I1002 01:48:12.183503 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:12Z","lastTransitionTime":"2025-10-02T01:48:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:12 crc kubenswrapper[4885]: I1002 01:48:12.185747 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vhgbd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df9792d80c807ed98c8442822d01f619327361c40425abfe26f11e5c27ba2584\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c63a7d20d39b65acfb5a9be4c28b21c72111b398bea76c238e15ddb4dd63e139\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T01:48:00Z\\\",\\\"message\\\":\\\"2025-10-02T01:47:15+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_9f42b0ef-8651-49db-8a31-d21c0c1c47e7\\\\n2025-10-02T01:47:15+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_9f42b0ef-8651-49db-8a31-d21c0c1c47e7 to /host/opt/cni/bin/\\\\n2025-10-02T01:47:15Z [verbose] multus-daemon started\\\\n2025-10-02T01:47:15Z [verbose] Readiness Indicator file check\\\\n2025-10-02T01:48:00Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7zwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vhgbd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:12Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:12 crc kubenswrapper[4885]: I1002 01:48:12.204039 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rrn8b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52967a77-525c-4537-9b1f-d7bb754a8494\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3dee451ebe9006bf1c519fed4c65267570b6139d76841858352e087ae78d26a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtr74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e9412788d6df7332592beac48281822b26e1aa19b39addce8e1be556af9d982\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtr74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:26Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rrn8b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:12Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:12 crc kubenswrapper[4885]: I1002 01:48:12.222881 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18466979-7c7c-4d73-86c6-b8b6eaf20484\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b742e4b288bb50569d2fd23f3f9857d649a46063998170eee582ecbbba64d190\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://298c2ace58c4a9d09cfdc864316317e370bb0ac5735d631af6d85c6298f686d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://620f54e02c463f6e603808e5e57e9e5770d125578cc258ba16117a2416e57d3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b4cd17b64e5657ae05e71f3ca2340cb0d7965b1b49d1f7b4bfe4494292293ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b4cd17b64e5657ae05e71f3ca2340cb0d7965b1b49d1f7b4bfe4494292293ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:12Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:12 crc kubenswrapper[4885]: I1002 01:48:12.253322 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74adcf912ccedb6b5d4f2803fd215e3e2ef62367597aa6da528354c4511ee536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7403ef968d040ac09d722b7557438bef23892d578992826f42b700b370d495d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:12Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:12 crc kubenswrapper[4885]: I1002 01:48:12.278002 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2ll9b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52b39712-5fa1-4fe0-814b-b170a6e3938b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9a96383d0fbfd116c419054aaf1a7adba1937e8b0bad294cefd68b4ab0d2298\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c979b91ba748aa4647826a8926337662ca929c54090fb62285572819ec2fbf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c979b91ba748aa4647826a8926337662ca929c54090fb62285572819ec2fbf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97d9d693e8e60ee9c183bb44461e698e8b3cfd20a516605aa8754c9205d2bdc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://97d9d693e8e60ee9c183bb44461e698e8b3cfd20a516605aa8754c9205d2bdc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c44252676359c671d65c9588b05572d76c99cf4c90666f0ac089b2f778450a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c44252676359c671d65c9588b05572d76c99cf4c90666f0ac089b2f778450a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80c982f07874df686fe7e29cc91043eb6d58f1be5fc93cacd46d33a21dd9b622\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80c982f07874df686fe7e29cc91043eb6d58f1be5fc93cacd46d33a21dd9b622\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a354c985652c7ad7fa8944fbfdfe223cfb0f97fb11eecd9bd429020d4892b78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a354c985652c7ad7fa8944fbfdfe223cfb0f97fb11eecd9bd429020d4892b78\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2efd7283df313d086c1b8b88f289ace28764de256205bee7610d6711002d8515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2efd7283df313d086c1b8b88f289ace28764de256205bee7610d6711002d8515\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2ll9b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:12Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:12 crc kubenswrapper[4885]: I1002 01:48:12.286589 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:12 crc kubenswrapper[4885]: I1002 01:48:12.286626 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:12 crc kubenswrapper[4885]: I1002 01:48:12.286642 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:12 crc kubenswrapper[4885]: I1002 01:48:12.286667 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:12 crc kubenswrapper[4885]: I1002 01:48:12.286683 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:12Z","lastTransitionTime":"2025-10-02T01:48:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:12 crc kubenswrapper[4885]: I1002 01:48:12.297696 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:12Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:12 crc kubenswrapper[4885]: I1002 01:48:12.318508 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b087880e05039816dcab154643ac72c7e4a0a1124817c51dc2c639b62bbee0fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:12Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:12 crc kubenswrapper[4885]: I1002 01:48:12.337421 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:12Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:12 crc kubenswrapper[4885]: I1002 01:48:12.355337 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:12Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:12 crc kubenswrapper[4885]: I1002 01:48:12.370129 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-58lhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac6a19c4-f259-4dbd-a016-d6c2c849c6cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed76c0f968c063f7795d363af194a222487b647ceda234de0dabb39b25eafe0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hzc4c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-58lhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:12Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:12 crc kubenswrapper[4885]: I1002 01:48:12.388600 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d169c8ba-000a-4218-8e3b-4ae53035b110\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61e436b7876107ce5ea9e668c0bbc367be6e663fc804fa3a65b51adf8be8d51f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-77rzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09536e182a46e77318c10812d1777e551f834907c006cab5c726dc66f5bce88a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-77rzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rttx8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:12Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:12 crc kubenswrapper[4885]: I1002 01:48:12.390527 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:12 crc kubenswrapper[4885]: I1002 01:48:12.390573 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:12 crc kubenswrapper[4885]: I1002 01:48:12.390590 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:12 crc kubenswrapper[4885]: I1002 01:48:12.390612 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:12 crc kubenswrapper[4885]: I1002 01:48:12.390630 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:12Z","lastTransitionTime":"2025-10-02T01:48:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:12 crc kubenswrapper[4885]: I1002 01:48:12.402509 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jgdvt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81733e8d-67b0-46fa-8a41-08b40645bd3e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcef60c816ca1ba9a298da658e84c4d717bea3d352c4c2f807ebd8e56327b384\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cwbl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jgdvt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:12Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:12 crc kubenswrapper[4885]: I1002 01:48:12.423166 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a0bb885-2eb0-419c-8a7a-ef634cb3c144\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a89b49cbe08d927629564e607de13655ddf3d01ac4eb5d688b325210306c120c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06c5942805a498d9621a906088aa54538b4eed583d1b8d0e3ba2a6718378a36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b9f6f7fe44d5771a838d0b89c5bdea3ebb79023ae2675ee907ec2766b3d35a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://178b4340ee6904c60c814b99d83ee225161aeb23e438d8d02f856579a13ec4e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bb63bf30d6ecce9e71f0db0c16632bdfa6b33924b673378eb0c38ab229fa837\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T01:47:06Z\\\",\\\"message\\\":\\\"W1002 01:46:55.406027 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1002 01:46:55.406714 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759369615 cert, and key in /tmp/serving-cert-3795040473/serving-signer.crt, /tmp/serving-cert-3795040473/serving-signer.key\\\\nI1002 01:46:55.614670 1 observer_polling.go:159] Starting file observer\\\\nW1002 01:46:55.617662 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1002 01:46:55.618112 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 01:46:55.620347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3795040473/tls.crt::/tmp/serving-cert-3795040473/tls.key\\\\\\\"\\\\nF1002 01:47:06.224459 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90535f97ab2ed4626d47540ebd7baada837fd332140323bec0f49aa38d3c0009\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://592f89a326c9c1559193a71904b47ca27b1dbe1a7b30e03bc48924bb2f35f921\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://592f89a326c9c1559193a71904b47ca27b1dbe1a7b30e03bc48924bb2f35f921\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:12Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:12 crc kubenswrapper[4885]: I1002 01:48:12.493203 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:12 crc kubenswrapper[4885]: I1002 01:48:12.493285 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:12 crc kubenswrapper[4885]: I1002 01:48:12.493304 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:12 crc kubenswrapper[4885]: I1002 01:48:12.493325 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:12 crc kubenswrapper[4885]: I1002 01:48:12.493339 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:12Z","lastTransitionTime":"2025-10-02T01:48:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:12 crc kubenswrapper[4885]: I1002 01:48:12.596491 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:12 crc kubenswrapper[4885]: I1002 01:48:12.596540 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:12 crc kubenswrapper[4885]: I1002 01:48:12.596552 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:12 crc kubenswrapper[4885]: I1002 01:48:12.596568 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:12 crc kubenswrapper[4885]: I1002 01:48:12.596580 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:12Z","lastTransitionTime":"2025-10-02T01:48:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:12 crc kubenswrapper[4885]: I1002 01:48:12.637959 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dngcm_9e4679dd-f870-41e9-a32b-360f5d2ee81b/ovnkube-controller/3.log" Oct 02 01:48:12 crc kubenswrapper[4885]: I1002 01:48:12.643374 4885 scope.go:117] "RemoveContainer" containerID="2ef80e838accc59ad7a579c66f4fcd6881d23fa86fc5922d4b0085d211d5d94b" Oct 02 01:48:12 crc kubenswrapper[4885]: E1002 01:48:12.643625 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-dngcm_openshift-ovn-kubernetes(9e4679dd-f870-41e9-a32b-360f5d2ee81b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" podUID="9e4679dd-f870-41e9-a32b-360f5d2ee81b" Oct 02 01:48:12 crc kubenswrapper[4885]: I1002 01:48:12.661039 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-sv4md" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"724a3e49-001c-45a4-a896-82c13e0ee0e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hpqgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hpqgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-sv4md\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:12Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:12 crc kubenswrapper[4885]: I1002 01:48:12.680314 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"506a599f-fb1c-4d14-be88-0e4833f3d624\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31c2f0193ffef54965b79a3569b0f6f4547e540bac0774155159a1c996c196c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://533e73e5f90b7258acbcdf018ee01323c118f9dc45035b25b065ba01cabecccf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c00db833a593ec40eb63a6b383455e19e960c2e98b9b66f76877ee993057e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21e2f462a0deb0d740185663b2014f4a1e57b4fdbeb82571264b002f928035bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:12Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:12 crc kubenswrapper[4885]: I1002 01:48:12.699872 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:12 crc kubenswrapper[4885]: I1002 01:48:12.699925 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:12 crc kubenswrapper[4885]: I1002 01:48:12.699943 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:12 crc kubenswrapper[4885]: I1002 01:48:12.699968 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:12 crc kubenswrapper[4885]: I1002 01:48:12.699985 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:12Z","lastTransitionTime":"2025-10-02T01:48:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:12 crc kubenswrapper[4885]: I1002 01:48:12.710345 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e4679dd-f870-41e9-a32b-360f5d2ee81b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4310909ad9c7b609ed6814e04a525a6afae747775598e7275080413dec8b23a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7642033b0b5995d332ccf7992d6d7e6a5e9c9369ae2d1071ccb2f536938c8f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9f99f9e26a04fb42b01a0783a5c4e360514818ce9102edc7fa08626d50b9a34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6fa72b030d309656b73121e4ac2a3924a4e449aad66683dd9de5ee5f22d5a4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef403e519d197fb37b3e6b76a14a5128ba3f30a7a5e8bf0a420cf6a39b89575f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8482ba457683ffa03100ec1f07ea9d3f918be5e490b5ba05c2ba965d03033191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ef80e838accc59ad7a579c66f4fcd6881d23fa86fc5922d4b0085d211d5d94b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ef80e838accc59ad7a579c66f4fcd6881d23fa86fc5922d4b0085d211d5d94b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T01:48:11Z\\\",\\\"message\\\":\\\" 6940 services_controller.go:451] Built service openshift-operator-lifecycle-manager/catalog-operator-metrics cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-operator-lifecycle-manager/catalog-operator-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-operator-lifecycle-manager/catalog-operator-metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.204\\\\\\\", Port:8443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1002 01:48:11.048667 6940 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:48:10Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-dngcm_openshift-ovn-kubernetes(9e4679dd-f870-41e9-a32b-360f5d2ee81b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fafdd1094bd02786d745c984989852b0c690db38655f92b419733a2004f417c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0125e907d061a97286026ee823c4eed303734ed03ab4864f742592b23a1aa2f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0125e907d061a97286026ee823c4eed303734ed03ab4864f742592b23a1aa2f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dngcm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:12Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:12 crc kubenswrapper[4885]: I1002 01:48:12.729367 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5c1f95f188569c9aa47e51f2916939fa0b65a5cdf4c16b798612d41f3c8411a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:12Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:12 crc kubenswrapper[4885]: I1002 01:48:12.749320 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vhgbd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df9792d80c807ed98c8442822d01f619327361c40425abfe26f11e5c27ba2584\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c63a7d20d39b65acfb5a9be4c28b21c72111b398bea76c238e15ddb4dd63e139\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T01:48:00Z\\\",\\\"message\\\":\\\"2025-10-02T01:47:15+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_9f42b0ef-8651-49db-8a31-d21c0c1c47e7\\\\n2025-10-02T01:47:15+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_9f42b0ef-8651-49db-8a31-d21c0c1c47e7 to /host/opt/cni/bin/\\\\n2025-10-02T01:47:15Z [verbose] multus-daemon started\\\\n2025-10-02T01:47:15Z [verbose] Readiness Indicator file check\\\\n2025-10-02T01:48:00Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7zwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vhgbd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:12Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:12 crc kubenswrapper[4885]: I1002 01:48:12.766239 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rrn8b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52967a77-525c-4537-9b1f-d7bb754a8494\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3dee451ebe9006bf1c519fed4c65267570b6139d76841858352e087ae78d26a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtr74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e9412788d6df7332592beac48281822b26e1aa19b39addce8e1be556af9d982\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtr74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:26Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rrn8b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:12Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:12 crc kubenswrapper[4885]: I1002 01:48:12.783702 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18466979-7c7c-4d73-86c6-b8b6eaf20484\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b742e4b288bb50569d2fd23f3f9857d649a46063998170eee582ecbbba64d190\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://298c2ace58c4a9d09cfdc864316317e370bb0ac5735d631af6d85c6298f686d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://620f54e02c463f6e603808e5e57e9e5770d125578cc258ba16117a2416e57d3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b4cd17b64e5657ae05e71f3ca2340cb0d7965b1b49d1f7b4bfe4494292293ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b4cd17b64e5657ae05e71f3ca2340cb0d7965b1b49d1f7b4bfe4494292293ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:12Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:12 crc kubenswrapper[4885]: I1002 01:48:12.802788 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:12 crc kubenswrapper[4885]: I1002 01:48:12.802841 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:12 crc kubenswrapper[4885]: I1002 01:48:12.802860 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:12 crc kubenswrapper[4885]: I1002 01:48:12.802890 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:12 crc kubenswrapper[4885]: I1002 01:48:12.802908 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:12Z","lastTransitionTime":"2025-10-02T01:48:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:12 crc kubenswrapper[4885]: I1002 01:48:12.815251 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8070bc1c-e3de-4e9a-985e-40d1a9858056\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3e634b3727c517cc48b17c2cfb38f26460f53d1bf25d631d08ec9acb8d780e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4adcc537c23697a324d82af94c31059dd444245efdf7c978db58ede256e112ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edcd1d99644a778dfb6d0a17d2a4abd906400fe8209e611ebf8d7a9abe7ccf28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://97cb59adc2ddfa365ebc9dec649600dc90466ffe59134a6d15a9d3134a7c9b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d2808bc658ea80d68469ed16422d56ef19447f4c90b4af7b2597c64406f8a44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://83acbc20919f7c1958b6d492fadb203855df02761662425df9f91a0757fe2bf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83acbc20919f7c1958b6d492fadb203855df02761662425df9f91a0757fe2bf7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e86ecd03835efc9a253b6804bc61e6ac690e75b8d1f3cc77f5647730e586574\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e86ecd03835efc9a253b6804bc61e6ac690e75b8d1f3cc77f5647730e586574\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://48f95af51feaf014a18496bdfadd59d7f1e11167c5c8b677fa7037e48368ff2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48f95af51feaf014a18496bdfadd59d7f1e11167c5c8b677fa7037e48368ff2a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:12Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:12 crc kubenswrapper[4885]: I1002 01:48:12.837451 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2ll9b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52b39712-5fa1-4fe0-814b-b170a6e3938b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9a96383d0fbfd116c419054aaf1a7adba1937e8b0bad294cefd68b4ab0d2298\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c979b91ba748aa4647826a8926337662ca929c54090fb62285572819ec2fbf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c979b91ba748aa4647826a8926337662ca929c54090fb62285572819ec2fbf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97d9d693e8e60ee9c183bb44461e698e8b3cfd20a516605aa8754c9205d2bdc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://97d9d693e8e60ee9c183bb44461e698e8b3cfd20a516605aa8754c9205d2bdc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c44252676359c671d65c9588b05572d76c99cf4c90666f0ac089b2f778450a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c44252676359c671d65c9588b05572d76c99cf4c90666f0ac089b2f778450a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80c982f07874df686fe7e29cc91043eb6d58f1be5fc93cacd46d33a21dd9b622\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80c982f07874df686fe7e29cc91043eb6d58f1be5fc93cacd46d33a21dd9b622\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a354c985652c7ad7fa8944fbfdfe223cfb0f97fb11eecd9bd429020d4892b78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a354c985652c7ad7fa8944fbfdfe223cfb0f97fb11eecd9bd429020d4892b78\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2efd7283df313d086c1b8b88f289ace28764de256205bee7610d6711002d8515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2efd7283df313d086c1b8b88f289ace28764de256205bee7610d6711002d8515\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2ll9b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:12Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:12 crc kubenswrapper[4885]: I1002 01:48:12.856637 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:12Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:12 crc kubenswrapper[4885]: I1002 01:48:12.877364 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74adcf912ccedb6b5d4f2803fd215e3e2ef62367597aa6da528354c4511ee536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7403ef968d040ac09d722b7557438bef23892d578992826f42b700b370d495d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:12Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:12 crc kubenswrapper[4885]: I1002 01:48:12.897803 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b087880e05039816dcab154643ac72c7e4a0a1124817c51dc2c639b62bbee0fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:12Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:12 crc kubenswrapper[4885]: I1002 01:48:12.905647 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:12 crc kubenswrapper[4885]: I1002 01:48:12.905717 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:12 crc kubenswrapper[4885]: I1002 01:48:12.905737 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:12 crc kubenswrapper[4885]: I1002 01:48:12.905761 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:12 crc kubenswrapper[4885]: I1002 01:48:12.905779 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:12Z","lastTransitionTime":"2025-10-02T01:48:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:12 crc kubenswrapper[4885]: I1002 01:48:12.918929 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:12Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:12 crc kubenswrapper[4885]: I1002 01:48:12.938571 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:12Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:12 crc kubenswrapper[4885]: I1002 01:48:12.955568 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-58lhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac6a19c4-f259-4dbd-a016-d6c2c849c6cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed76c0f968c063f7795d363af194a222487b647ceda234de0dabb39b25eafe0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hzc4c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-58lhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:12Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:12 crc kubenswrapper[4885]: I1002 01:48:12.977726 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d169c8ba-000a-4218-8e3b-4ae53035b110\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61e436b7876107ce5ea9e668c0bbc367be6e663fc804fa3a65b51adf8be8d51f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-77rzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09536e182a46e77318c10812d1777e551f834907c006cab5c726dc66f5bce88a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-77rzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rttx8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:12Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:12 crc kubenswrapper[4885]: I1002 01:48:12.990493 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jgdvt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81733e8d-67b0-46fa-8a41-08b40645bd3e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcef60c816ca1ba9a298da658e84c4d717bea3d352c4c2f807ebd8e56327b384\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cwbl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jgdvt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:12Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:13 crc kubenswrapper[4885]: I1002 01:48:13.008908 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:13 crc kubenswrapper[4885]: I1002 01:48:13.008953 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:13 crc kubenswrapper[4885]: I1002 01:48:13.008966 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:13 crc kubenswrapper[4885]: I1002 01:48:13.008985 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:13 crc kubenswrapper[4885]: I1002 01:48:13.009002 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:13Z","lastTransitionTime":"2025-10-02T01:48:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:13 crc kubenswrapper[4885]: I1002 01:48:13.009211 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a0bb885-2eb0-419c-8a7a-ef634cb3c144\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a89b49cbe08d927629564e607de13655ddf3d01ac4eb5d688b325210306c120c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06c5942805a498d9621a906088aa54538b4eed583d1b8d0e3ba2a6718378a36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b9f6f7fe44d5771a838d0b89c5bdea3ebb79023ae2675ee907ec2766b3d35a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://178b4340ee6904c60c814b99d83ee225161aeb23e438d8d02f856579a13ec4e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bb63bf30d6ecce9e71f0db0c16632bdfa6b33924b673378eb0c38ab229fa837\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T01:47:06Z\\\",\\\"message\\\":\\\"W1002 01:46:55.406027 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1002 01:46:55.406714 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759369615 cert, and key in /tmp/serving-cert-3795040473/serving-signer.crt, /tmp/serving-cert-3795040473/serving-signer.key\\\\nI1002 01:46:55.614670 1 observer_polling.go:159] Starting file observer\\\\nW1002 01:46:55.617662 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1002 01:46:55.618112 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 01:46:55.620347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3795040473/tls.crt::/tmp/serving-cert-3795040473/tls.key\\\\\\\"\\\\nF1002 01:47:06.224459 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90535f97ab2ed4626d47540ebd7baada837fd332140323bec0f49aa38d3c0009\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://592f89a326c9c1559193a71904b47ca27b1dbe1a7b30e03bc48924bb2f35f921\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://592f89a326c9c1559193a71904b47ca27b1dbe1a7b30e03bc48924bb2f35f921\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:13Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:13 crc kubenswrapper[4885]: I1002 01:48:13.112141 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:13 crc kubenswrapper[4885]: I1002 01:48:13.112215 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:13 crc kubenswrapper[4885]: I1002 01:48:13.112234 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:13 crc kubenswrapper[4885]: I1002 01:48:13.112299 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:13 crc kubenswrapper[4885]: I1002 01:48:13.112317 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:13Z","lastTransitionTime":"2025-10-02T01:48:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:13 crc kubenswrapper[4885]: I1002 01:48:13.215089 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:13 crc kubenswrapper[4885]: I1002 01:48:13.215159 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:13 crc kubenswrapper[4885]: I1002 01:48:13.215180 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:13 crc kubenswrapper[4885]: I1002 01:48:13.215209 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:13 crc kubenswrapper[4885]: I1002 01:48:13.215231 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:13Z","lastTransitionTime":"2025-10-02T01:48:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:13 crc kubenswrapper[4885]: I1002 01:48:13.318781 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:13 crc kubenswrapper[4885]: I1002 01:48:13.318878 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:13 crc kubenswrapper[4885]: I1002 01:48:13.318903 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:13 crc kubenswrapper[4885]: I1002 01:48:13.318935 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:13 crc kubenswrapper[4885]: I1002 01:48:13.318959 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:13Z","lastTransitionTime":"2025-10-02T01:48:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:13 crc kubenswrapper[4885]: I1002 01:48:13.421346 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:13 crc kubenswrapper[4885]: I1002 01:48:13.421421 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:13 crc kubenswrapper[4885]: I1002 01:48:13.421441 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:13 crc kubenswrapper[4885]: I1002 01:48:13.421468 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:13 crc kubenswrapper[4885]: I1002 01:48:13.421488 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:13Z","lastTransitionTime":"2025-10-02T01:48:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:13 crc kubenswrapper[4885]: I1002 01:48:13.524509 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:13 crc kubenswrapper[4885]: I1002 01:48:13.524584 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:13 crc kubenswrapper[4885]: I1002 01:48:13.524601 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:13 crc kubenswrapper[4885]: I1002 01:48:13.524627 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:13 crc kubenswrapper[4885]: I1002 01:48:13.524647 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:13Z","lastTransitionTime":"2025-10-02T01:48:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:13 crc kubenswrapper[4885]: I1002 01:48:13.627200 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:13 crc kubenswrapper[4885]: I1002 01:48:13.627339 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:13 crc kubenswrapper[4885]: I1002 01:48:13.627362 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:13 crc kubenswrapper[4885]: I1002 01:48:13.627388 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:13 crc kubenswrapper[4885]: I1002 01:48:13.627405 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:13Z","lastTransitionTime":"2025-10-02T01:48:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:13 crc kubenswrapper[4885]: I1002 01:48:13.730079 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:13 crc kubenswrapper[4885]: I1002 01:48:13.730154 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:13 crc kubenswrapper[4885]: I1002 01:48:13.730172 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:13 crc kubenswrapper[4885]: I1002 01:48:13.730203 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:13 crc kubenswrapper[4885]: I1002 01:48:13.730223 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:13Z","lastTransitionTime":"2025-10-02T01:48:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:13 crc kubenswrapper[4885]: I1002 01:48:13.833190 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:13 crc kubenswrapper[4885]: I1002 01:48:13.833327 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:13 crc kubenswrapper[4885]: I1002 01:48:13.833352 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:13 crc kubenswrapper[4885]: I1002 01:48:13.833380 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:13 crc kubenswrapper[4885]: I1002 01:48:13.833398 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:13Z","lastTransitionTime":"2025-10-02T01:48:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:13 crc kubenswrapper[4885]: I1002 01:48:13.936628 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:13 crc kubenswrapper[4885]: I1002 01:48:13.936685 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:13 crc kubenswrapper[4885]: I1002 01:48:13.936702 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:13 crc kubenswrapper[4885]: I1002 01:48:13.936725 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:13 crc kubenswrapper[4885]: I1002 01:48:13.936743 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:13Z","lastTransitionTime":"2025-10-02T01:48:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:14 crc kubenswrapper[4885]: I1002 01:48:14.046111 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:48:14 crc kubenswrapper[4885]: I1002 01:48:14.046372 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sv4md" Oct 02 01:48:14 crc kubenswrapper[4885]: E1002 01:48:14.047212 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:48:14 crc kubenswrapper[4885]: I1002 01:48:14.047664 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:48:14 crc kubenswrapper[4885]: I1002 01:48:14.048164 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:14 crc kubenswrapper[4885]: I1002 01:48:14.048213 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:14 crc kubenswrapper[4885]: E1002 01:48:14.048240 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:48:14 crc kubenswrapper[4885]: I1002 01:48:14.048251 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:14 crc kubenswrapper[4885]: I1002 01:48:14.048377 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:14 crc kubenswrapper[4885]: E1002 01:48:14.047777 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sv4md" podUID="724a3e49-001c-45a4-a896-82c13e0ee0e7" Oct 02 01:48:14 crc kubenswrapper[4885]: I1002 01:48:14.048399 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:14Z","lastTransitionTime":"2025-10-02T01:48:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:14 crc kubenswrapper[4885]: I1002 01:48:14.048077 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:48:14 crc kubenswrapper[4885]: E1002 01:48:14.048810 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:48:14 crc kubenswrapper[4885]: I1002 01:48:14.151586 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:14 crc kubenswrapper[4885]: I1002 01:48:14.151642 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:14 crc kubenswrapper[4885]: I1002 01:48:14.151656 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:14 crc kubenswrapper[4885]: I1002 01:48:14.151680 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:14 crc kubenswrapper[4885]: I1002 01:48:14.151695 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:14Z","lastTransitionTime":"2025-10-02T01:48:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:14 crc kubenswrapper[4885]: I1002 01:48:14.254842 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:14 crc kubenswrapper[4885]: I1002 01:48:14.254916 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:14 crc kubenswrapper[4885]: I1002 01:48:14.254937 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:14 crc kubenswrapper[4885]: I1002 01:48:14.254965 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:14 crc kubenswrapper[4885]: I1002 01:48:14.254987 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:14Z","lastTransitionTime":"2025-10-02T01:48:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:14 crc kubenswrapper[4885]: I1002 01:48:14.359798 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:14 crc kubenswrapper[4885]: I1002 01:48:14.359868 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:14 crc kubenswrapper[4885]: I1002 01:48:14.359891 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:14 crc kubenswrapper[4885]: I1002 01:48:14.359924 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:14 crc kubenswrapper[4885]: I1002 01:48:14.359953 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:14Z","lastTransitionTime":"2025-10-02T01:48:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:14 crc kubenswrapper[4885]: I1002 01:48:14.463321 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:14 crc kubenswrapper[4885]: I1002 01:48:14.463573 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:14 crc kubenswrapper[4885]: I1002 01:48:14.463601 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:14 crc kubenswrapper[4885]: I1002 01:48:14.463633 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:14 crc kubenswrapper[4885]: I1002 01:48:14.463655 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:14Z","lastTransitionTime":"2025-10-02T01:48:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:14 crc kubenswrapper[4885]: I1002 01:48:14.567173 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:14 crc kubenswrapper[4885]: I1002 01:48:14.567310 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:14 crc kubenswrapper[4885]: I1002 01:48:14.567333 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:14 crc kubenswrapper[4885]: I1002 01:48:14.567361 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:14 crc kubenswrapper[4885]: I1002 01:48:14.567382 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:14Z","lastTransitionTime":"2025-10-02T01:48:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:14 crc kubenswrapper[4885]: I1002 01:48:14.669976 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:14 crc kubenswrapper[4885]: I1002 01:48:14.670035 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:14 crc kubenswrapper[4885]: I1002 01:48:14.670054 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:14 crc kubenswrapper[4885]: I1002 01:48:14.670078 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:14 crc kubenswrapper[4885]: I1002 01:48:14.670097 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:14Z","lastTransitionTime":"2025-10-02T01:48:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:14 crc kubenswrapper[4885]: I1002 01:48:14.772736 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:14 crc kubenswrapper[4885]: I1002 01:48:14.772806 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:14 crc kubenswrapper[4885]: I1002 01:48:14.772825 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:14 crc kubenswrapper[4885]: I1002 01:48:14.772857 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:14 crc kubenswrapper[4885]: I1002 01:48:14.772877 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:14Z","lastTransitionTime":"2025-10-02T01:48:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:14 crc kubenswrapper[4885]: I1002 01:48:14.876045 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:14 crc kubenswrapper[4885]: I1002 01:48:14.876108 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:14 crc kubenswrapper[4885]: I1002 01:48:14.876126 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:14 crc kubenswrapper[4885]: I1002 01:48:14.876150 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:14 crc kubenswrapper[4885]: I1002 01:48:14.876171 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:14Z","lastTransitionTime":"2025-10-02T01:48:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:14 crc kubenswrapper[4885]: I1002 01:48:14.979543 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:14 crc kubenswrapper[4885]: I1002 01:48:14.979604 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:14 crc kubenswrapper[4885]: I1002 01:48:14.979623 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:14 crc kubenswrapper[4885]: I1002 01:48:14.979649 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:14 crc kubenswrapper[4885]: I1002 01:48:14.979667 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:14Z","lastTransitionTime":"2025-10-02T01:48:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:15 crc kubenswrapper[4885]: I1002 01:48:15.081941 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:15 crc kubenswrapper[4885]: I1002 01:48:15.082026 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:15 crc kubenswrapper[4885]: I1002 01:48:15.082045 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:15 crc kubenswrapper[4885]: I1002 01:48:15.082070 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:15 crc kubenswrapper[4885]: I1002 01:48:15.082088 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:15Z","lastTransitionTime":"2025-10-02T01:48:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:15 crc kubenswrapper[4885]: I1002 01:48:15.185459 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:15 crc kubenswrapper[4885]: I1002 01:48:15.185649 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:15 crc kubenswrapper[4885]: I1002 01:48:15.185669 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:15 crc kubenswrapper[4885]: I1002 01:48:15.185691 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:15 crc kubenswrapper[4885]: I1002 01:48:15.185714 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:15Z","lastTransitionTime":"2025-10-02T01:48:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:15 crc kubenswrapper[4885]: I1002 01:48:15.288801 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:15 crc kubenswrapper[4885]: I1002 01:48:15.288842 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:15 crc kubenswrapper[4885]: I1002 01:48:15.288859 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:15 crc kubenswrapper[4885]: I1002 01:48:15.288916 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:15 crc kubenswrapper[4885]: I1002 01:48:15.288935 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:15Z","lastTransitionTime":"2025-10-02T01:48:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:15 crc kubenswrapper[4885]: I1002 01:48:15.391610 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:15 crc kubenswrapper[4885]: I1002 01:48:15.391710 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:15 crc kubenswrapper[4885]: I1002 01:48:15.391764 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:15 crc kubenswrapper[4885]: I1002 01:48:15.391792 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:15 crc kubenswrapper[4885]: I1002 01:48:15.391820 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:15Z","lastTransitionTime":"2025-10-02T01:48:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:15 crc kubenswrapper[4885]: I1002 01:48:15.494801 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:15 crc kubenswrapper[4885]: I1002 01:48:15.494904 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:15 crc kubenswrapper[4885]: I1002 01:48:15.494923 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:15 crc kubenswrapper[4885]: I1002 01:48:15.494990 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:15 crc kubenswrapper[4885]: I1002 01:48:15.495006 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:15Z","lastTransitionTime":"2025-10-02T01:48:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:15 crc kubenswrapper[4885]: I1002 01:48:15.597707 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:15 crc kubenswrapper[4885]: I1002 01:48:15.597764 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:15 crc kubenswrapper[4885]: I1002 01:48:15.597780 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:15 crc kubenswrapper[4885]: I1002 01:48:15.597808 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:15 crc kubenswrapper[4885]: I1002 01:48:15.597830 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:15Z","lastTransitionTime":"2025-10-02T01:48:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:15 crc kubenswrapper[4885]: I1002 01:48:15.701039 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:15 crc kubenswrapper[4885]: I1002 01:48:15.701117 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:15 crc kubenswrapper[4885]: I1002 01:48:15.701142 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:15 crc kubenswrapper[4885]: I1002 01:48:15.701175 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:15 crc kubenswrapper[4885]: I1002 01:48:15.701198 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:15Z","lastTransitionTime":"2025-10-02T01:48:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:15 crc kubenswrapper[4885]: I1002 01:48:15.803744 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:15 crc kubenswrapper[4885]: I1002 01:48:15.803826 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:15 crc kubenswrapper[4885]: I1002 01:48:15.803852 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:15 crc kubenswrapper[4885]: I1002 01:48:15.803887 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:15 crc kubenswrapper[4885]: I1002 01:48:15.803914 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:15Z","lastTransitionTime":"2025-10-02T01:48:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:15 crc kubenswrapper[4885]: I1002 01:48:15.905842 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:15 crc kubenswrapper[4885]: I1002 01:48:15.905911 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:15 crc kubenswrapper[4885]: I1002 01:48:15.905930 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:15 crc kubenswrapper[4885]: I1002 01:48:15.905956 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:15 crc kubenswrapper[4885]: I1002 01:48:15.905976 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:15Z","lastTransitionTime":"2025-10-02T01:48:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:15 crc kubenswrapper[4885]: I1002 01:48:15.955201 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:48:15 crc kubenswrapper[4885]: E1002 01:48:15.955408 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:49:19.955374162 +0000 UTC m=+148.767121601 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:48:15 crc kubenswrapper[4885]: I1002 01:48:15.955524 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:48:15 crc kubenswrapper[4885]: E1002 01:48:15.955751 4885 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 01:48:15 crc kubenswrapper[4885]: E1002 01:48:15.955855 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 01:49:19.955824906 +0000 UTC m=+148.767572355 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 01:48:16 crc kubenswrapper[4885]: I1002 01:48:16.008895 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:16 crc kubenswrapper[4885]: I1002 01:48:16.008957 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:16 crc kubenswrapper[4885]: I1002 01:48:16.008980 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:16 crc kubenswrapper[4885]: I1002 01:48:16.009007 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:16 crc kubenswrapper[4885]: I1002 01:48:16.009029 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:16Z","lastTransitionTime":"2025-10-02T01:48:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:16 crc kubenswrapper[4885]: I1002 01:48:16.045842 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sv4md" Oct 02 01:48:16 crc kubenswrapper[4885]: I1002 01:48:16.045943 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:48:16 crc kubenswrapper[4885]: E1002 01:48:16.046027 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sv4md" podUID="724a3e49-001c-45a4-a896-82c13e0ee0e7" Oct 02 01:48:16 crc kubenswrapper[4885]: I1002 01:48:16.046244 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:48:16 crc kubenswrapper[4885]: E1002 01:48:16.046367 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:48:16 crc kubenswrapper[4885]: I1002 01:48:16.046581 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:48:16 crc kubenswrapper[4885]: E1002 01:48:16.046587 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:48:16 crc kubenswrapper[4885]: E1002 01:48:16.046789 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:48:16 crc kubenswrapper[4885]: I1002 01:48:16.057251 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:48:16 crc kubenswrapper[4885]: E1002 01:48:16.057536 4885 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 01:48:16 crc kubenswrapper[4885]: E1002 01:48:16.057578 4885 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 01:48:16 crc kubenswrapper[4885]: E1002 01:48:16.057596 4885 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 01:48:16 crc kubenswrapper[4885]: E1002 01:48:16.057663 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-02 01:49:20.057639291 +0000 UTC m=+148.869386730 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 01:48:16 crc kubenswrapper[4885]: E1002 01:48:16.057688 4885 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 01:48:16 crc kubenswrapper[4885]: I1002 01:48:16.057534 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:48:16 crc kubenswrapper[4885]: E1002 01:48:16.057724 4885 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 01:48:16 crc kubenswrapper[4885]: E1002 01:48:16.057748 4885 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 01:48:16 crc kubenswrapper[4885]: E1002 01:48:16.057821 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-02 01:49:20.057795755 +0000 UTC m=+148.869543244 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 01:48:16 crc kubenswrapper[4885]: I1002 01:48:16.057864 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:48:16 crc kubenswrapper[4885]: E1002 01:48:16.057968 4885 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 01:48:16 crc kubenswrapper[4885]: E1002 01:48:16.058052 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 01:49:20.058029123 +0000 UTC m=+148.869776682 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 01:48:16 crc kubenswrapper[4885]: I1002 01:48:16.112710 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:16 crc kubenswrapper[4885]: I1002 01:48:16.112778 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:16 crc kubenswrapper[4885]: I1002 01:48:16.112795 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:16 crc kubenswrapper[4885]: I1002 01:48:16.112819 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:16 crc kubenswrapper[4885]: I1002 01:48:16.112836 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:16Z","lastTransitionTime":"2025-10-02T01:48:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:16 crc kubenswrapper[4885]: I1002 01:48:16.216406 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:16 crc kubenswrapper[4885]: I1002 01:48:16.216493 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:16 crc kubenswrapper[4885]: I1002 01:48:16.216520 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:16 crc kubenswrapper[4885]: I1002 01:48:16.216561 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:16 crc kubenswrapper[4885]: I1002 01:48:16.216587 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:16Z","lastTransitionTime":"2025-10-02T01:48:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:16 crc kubenswrapper[4885]: I1002 01:48:16.320626 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:16 crc kubenswrapper[4885]: I1002 01:48:16.320710 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:16 crc kubenswrapper[4885]: I1002 01:48:16.320731 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:16 crc kubenswrapper[4885]: I1002 01:48:16.320758 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:16 crc kubenswrapper[4885]: I1002 01:48:16.320779 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:16Z","lastTransitionTime":"2025-10-02T01:48:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:16 crc kubenswrapper[4885]: I1002 01:48:16.424340 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:16 crc kubenswrapper[4885]: I1002 01:48:16.424409 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:16 crc kubenswrapper[4885]: I1002 01:48:16.424432 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:16 crc kubenswrapper[4885]: I1002 01:48:16.424460 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:16 crc kubenswrapper[4885]: I1002 01:48:16.424480 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:16Z","lastTransitionTime":"2025-10-02T01:48:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:16 crc kubenswrapper[4885]: I1002 01:48:16.527632 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:16 crc kubenswrapper[4885]: I1002 01:48:16.527701 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:16 crc kubenswrapper[4885]: I1002 01:48:16.527724 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:16 crc kubenswrapper[4885]: I1002 01:48:16.527757 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:16 crc kubenswrapper[4885]: I1002 01:48:16.527783 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:16Z","lastTransitionTime":"2025-10-02T01:48:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:16 crc kubenswrapper[4885]: I1002 01:48:16.631322 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:16 crc kubenswrapper[4885]: I1002 01:48:16.631451 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:16 crc kubenswrapper[4885]: I1002 01:48:16.631477 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:16 crc kubenswrapper[4885]: I1002 01:48:16.631511 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:16 crc kubenswrapper[4885]: I1002 01:48:16.631532 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:16Z","lastTransitionTime":"2025-10-02T01:48:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:16 crc kubenswrapper[4885]: I1002 01:48:16.735755 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:16 crc kubenswrapper[4885]: I1002 01:48:16.735807 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:16 crc kubenswrapper[4885]: I1002 01:48:16.735827 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:16 crc kubenswrapper[4885]: I1002 01:48:16.735853 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:16 crc kubenswrapper[4885]: I1002 01:48:16.735875 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:16Z","lastTransitionTime":"2025-10-02T01:48:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:16 crc kubenswrapper[4885]: I1002 01:48:16.839988 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:16 crc kubenswrapper[4885]: I1002 01:48:16.840056 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:16 crc kubenswrapper[4885]: I1002 01:48:16.840079 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:16 crc kubenswrapper[4885]: I1002 01:48:16.840108 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:16 crc kubenswrapper[4885]: I1002 01:48:16.840126 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:16Z","lastTransitionTime":"2025-10-02T01:48:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:16 crc kubenswrapper[4885]: I1002 01:48:16.942645 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:16 crc kubenswrapper[4885]: I1002 01:48:16.943377 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:16 crc kubenswrapper[4885]: I1002 01:48:16.943405 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:16 crc kubenswrapper[4885]: I1002 01:48:16.943429 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:16 crc kubenswrapper[4885]: I1002 01:48:16.943482 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:16Z","lastTransitionTime":"2025-10-02T01:48:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:17 crc kubenswrapper[4885]: I1002 01:48:17.046787 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:17 crc kubenswrapper[4885]: I1002 01:48:17.046871 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:17 crc kubenswrapper[4885]: I1002 01:48:17.046891 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:17 crc kubenswrapper[4885]: I1002 01:48:17.047568 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:17 crc kubenswrapper[4885]: I1002 01:48:17.047648 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:17Z","lastTransitionTime":"2025-10-02T01:48:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:17 crc kubenswrapper[4885]: I1002 01:48:17.063726 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Oct 02 01:48:17 crc kubenswrapper[4885]: I1002 01:48:17.151471 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:17 crc kubenswrapper[4885]: I1002 01:48:17.151541 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:17 crc kubenswrapper[4885]: I1002 01:48:17.151559 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:17 crc kubenswrapper[4885]: I1002 01:48:17.151583 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:17 crc kubenswrapper[4885]: I1002 01:48:17.151602 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:17Z","lastTransitionTime":"2025-10-02T01:48:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:17 crc kubenswrapper[4885]: I1002 01:48:17.255419 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:17 crc kubenswrapper[4885]: I1002 01:48:17.255473 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:17 crc kubenswrapper[4885]: I1002 01:48:17.255489 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:17 crc kubenswrapper[4885]: I1002 01:48:17.255515 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:17 crc kubenswrapper[4885]: I1002 01:48:17.255535 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:17Z","lastTransitionTime":"2025-10-02T01:48:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:17 crc kubenswrapper[4885]: I1002 01:48:17.358569 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:17 crc kubenswrapper[4885]: I1002 01:48:17.358641 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:17 crc kubenswrapper[4885]: I1002 01:48:17.358658 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:17 crc kubenswrapper[4885]: I1002 01:48:17.358689 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:17 crc kubenswrapper[4885]: I1002 01:48:17.358710 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:17Z","lastTransitionTime":"2025-10-02T01:48:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:17 crc kubenswrapper[4885]: I1002 01:48:17.461662 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:17 crc kubenswrapper[4885]: I1002 01:48:17.461745 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:17 crc kubenswrapper[4885]: I1002 01:48:17.461770 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:17 crc kubenswrapper[4885]: I1002 01:48:17.461802 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:17 crc kubenswrapper[4885]: I1002 01:48:17.461829 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:17Z","lastTransitionTime":"2025-10-02T01:48:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:17 crc kubenswrapper[4885]: I1002 01:48:17.565393 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:17 crc kubenswrapper[4885]: I1002 01:48:17.565463 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:17 crc kubenswrapper[4885]: I1002 01:48:17.565486 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:17 crc kubenswrapper[4885]: I1002 01:48:17.565515 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:17 crc kubenswrapper[4885]: I1002 01:48:17.565533 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:17Z","lastTransitionTime":"2025-10-02T01:48:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:17 crc kubenswrapper[4885]: I1002 01:48:17.668887 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:17 crc kubenswrapper[4885]: I1002 01:48:17.668973 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:17 crc kubenswrapper[4885]: I1002 01:48:17.668993 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:17 crc kubenswrapper[4885]: I1002 01:48:17.669028 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:17 crc kubenswrapper[4885]: I1002 01:48:17.669052 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:17Z","lastTransitionTime":"2025-10-02T01:48:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:17 crc kubenswrapper[4885]: I1002 01:48:17.772958 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:17 crc kubenswrapper[4885]: I1002 01:48:17.773008 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:17 crc kubenswrapper[4885]: I1002 01:48:17.773025 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:17 crc kubenswrapper[4885]: I1002 01:48:17.773047 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:17 crc kubenswrapper[4885]: I1002 01:48:17.773063 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:17Z","lastTransitionTime":"2025-10-02T01:48:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:17 crc kubenswrapper[4885]: I1002 01:48:17.876914 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:17 crc kubenswrapper[4885]: I1002 01:48:17.877009 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:17 crc kubenswrapper[4885]: I1002 01:48:17.877029 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:17 crc kubenswrapper[4885]: I1002 01:48:17.877058 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:17 crc kubenswrapper[4885]: I1002 01:48:17.877079 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:17Z","lastTransitionTime":"2025-10-02T01:48:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:17 crc kubenswrapper[4885]: I1002 01:48:17.980062 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:17 crc kubenswrapper[4885]: I1002 01:48:17.980125 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:17 crc kubenswrapper[4885]: I1002 01:48:17.980148 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:17 crc kubenswrapper[4885]: I1002 01:48:17.980170 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:17 crc kubenswrapper[4885]: I1002 01:48:17.980188 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:17Z","lastTransitionTime":"2025-10-02T01:48:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:18 crc kubenswrapper[4885]: I1002 01:48:18.046172 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:48:18 crc kubenswrapper[4885]: I1002 01:48:18.046203 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:48:18 crc kubenswrapper[4885]: I1002 01:48:18.046229 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:48:18 crc kubenswrapper[4885]: I1002 01:48:18.046379 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sv4md" Oct 02 01:48:18 crc kubenswrapper[4885]: E1002 01:48:18.046378 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:48:18 crc kubenswrapper[4885]: E1002 01:48:18.046595 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sv4md" podUID="724a3e49-001c-45a4-a896-82c13e0ee0e7" Oct 02 01:48:18 crc kubenswrapper[4885]: E1002 01:48:18.046831 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:48:18 crc kubenswrapper[4885]: E1002 01:48:18.046914 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:48:18 crc kubenswrapper[4885]: I1002 01:48:18.082852 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:18 crc kubenswrapper[4885]: I1002 01:48:18.082904 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:18 crc kubenswrapper[4885]: I1002 01:48:18.082921 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:18 crc kubenswrapper[4885]: I1002 01:48:18.082946 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:18 crc kubenswrapper[4885]: I1002 01:48:18.082966 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:18Z","lastTransitionTime":"2025-10-02T01:48:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:18 crc kubenswrapper[4885]: I1002 01:48:18.187151 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:18 crc kubenswrapper[4885]: I1002 01:48:18.187292 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:18 crc kubenswrapper[4885]: I1002 01:48:18.187321 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:18 crc kubenswrapper[4885]: I1002 01:48:18.187366 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:18 crc kubenswrapper[4885]: I1002 01:48:18.187398 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:18Z","lastTransitionTime":"2025-10-02T01:48:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:18 crc kubenswrapper[4885]: I1002 01:48:18.291800 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:18 crc kubenswrapper[4885]: I1002 01:48:18.291863 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:18 crc kubenswrapper[4885]: I1002 01:48:18.291882 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:18 crc kubenswrapper[4885]: I1002 01:48:18.291911 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:18 crc kubenswrapper[4885]: I1002 01:48:18.291929 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:18Z","lastTransitionTime":"2025-10-02T01:48:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:18 crc kubenswrapper[4885]: I1002 01:48:18.394843 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:18 crc kubenswrapper[4885]: I1002 01:48:18.394943 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:18 crc kubenswrapper[4885]: I1002 01:48:18.394971 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:18 crc kubenswrapper[4885]: I1002 01:48:18.395013 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:18 crc kubenswrapper[4885]: I1002 01:48:18.395044 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:18Z","lastTransitionTime":"2025-10-02T01:48:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:18 crc kubenswrapper[4885]: I1002 01:48:18.497556 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:18 crc kubenswrapper[4885]: I1002 01:48:18.497625 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:18 crc kubenswrapper[4885]: I1002 01:48:18.497647 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:18 crc kubenswrapper[4885]: I1002 01:48:18.497674 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:18 crc kubenswrapper[4885]: I1002 01:48:18.497694 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:18Z","lastTransitionTime":"2025-10-02T01:48:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:18 crc kubenswrapper[4885]: I1002 01:48:18.600282 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:18 crc kubenswrapper[4885]: I1002 01:48:18.600333 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:18 crc kubenswrapper[4885]: I1002 01:48:18.600345 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:18 crc kubenswrapper[4885]: I1002 01:48:18.600364 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:18 crc kubenswrapper[4885]: I1002 01:48:18.600383 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:18Z","lastTransitionTime":"2025-10-02T01:48:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:18 crc kubenswrapper[4885]: I1002 01:48:18.703835 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:18 crc kubenswrapper[4885]: I1002 01:48:18.703927 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:18 crc kubenswrapper[4885]: I1002 01:48:18.703947 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:18 crc kubenswrapper[4885]: I1002 01:48:18.704002 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:18 crc kubenswrapper[4885]: I1002 01:48:18.704019 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:18Z","lastTransitionTime":"2025-10-02T01:48:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:18 crc kubenswrapper[4885]: I1002 01:48:18.807207 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:18 crc kubenswrapper[4885]: I1002 01:48:18.807730 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:18 crc kubenswrapper[4885]: I1002 01:48:18.807752 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:18 crc kubenswrapper[4885]: I1002 01:48:18.807779 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:18 crc kubenswrapper[4885]: I1002 01:48:18.807800 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:18Z","lastTransitionTime":"2025-10-02T01:48:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:18 crc kubenswrapper[4885]: I1002 01:48:18.911028 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:18 crc kubenswrapper[4885]: I1002 01:48:18.911154 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:18 crc kubenswrapper[4885]: I1002 01:48:18.911181 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:18 crc kubenswrapper[4885]: I1002 01:48:18.911209 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:18 crc kubenswrapper[4885]: I1002 01:48:18.911233 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:18Z","lastTransitionTime":"2025-10-02T01:48:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:19 crc kubenswrapper[4885]: I1002 01:48:19.014210 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:19 crc kubenswrapper[4885]: I1002 01:48:19.014313 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:19 crc kubenswrapper[4885]: I1002 01:48:19.014332 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:19 crc kubenswrapper[4885]: I1002 01:48:19.014358 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:19 crc kubenswrapper[4885]: I1002 01:48:19.014375 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:19Z","lastTransitionTime":"2025-10-02T01:48:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:19 crc kubenswrapper[4885]: I1002 01:48:19.117291 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:19 crc kubenswrapper[4885]: I1002 01:48:19.117330 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:19 crc kubenswrapper[4885]: I1002 01:48:19.117360 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:19 crc kubenswrapper[4885]: I1002 01:48:19.117378 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:19 crc kubenswrapper[4885]: I1002 01:48:19.117391 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:19Z","lastTransitionTime":"2025-10-02T01:48:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:19 crc kubenswrapper[4885]: I1002 01:48:19.220118 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:19 crc kubenswrapper[4885]: I1002 01:48:19.220188 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:19 crc kubenswrapper[4885]: I1002 01:48:19.220207 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:19 crc kubenswrapper[4885]: I1002 01:48:19.220231 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:19 crc kubenswrapper[4885]: I1002 01:48:19.220249 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:19Z","lastTransitionTime":"2025-10-02T01:48:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:19 crc kubenswrapper[4885]: I1002 01:48:19.323741 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:19 crc kubenswrapper[4885]: I1002 01:48:19.323812 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:19 crc kubenswrapper[4885]: I1002 01:48:19.323831 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:19 crc kubenswrapper[4885]: I1002 01:48:19.323859 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:19 crc kubenswrapper[4885]: I1002 01:48:19.323877 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:19Z","lastTransitionTime":"2025-10-02T01:48:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:19 crc kubenswrapper[4885]: I1002 01:48:19.427376 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:19 crc kubenswrapper[4885]: I1002 01:48:19.427444 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:19 crc kubenswrapper[4885]: I1002 01:48:19.427467 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:19 crc kubenswrapper[4885]: I1002 01:48:19.427500 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:19 crc kubenswrapper[4885]: I1002 01:48:19.427524 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:19Z","lastTransitionTime":"2025-10-02T01:48:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:19 crc kubenswrapper[4885]: I1002 01:48:19.530563 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:19 crc kubenswrapper[4885]: I1002 01:48:19.530631 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:19 crc kubenswrapper[4885]: I1002 01:48:19.530653 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:19 crc kubenswrapper[4885]: I1002 01:48:19.530685 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:19 crc kubenswrapper[4885]: I1002 01:48:19.530707 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:19Z","lastTransitionTime":"2025-10-02T01:48:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:19 crc kubenswrapper[4885]: I1002 01:48:19.634072 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:19 crc kubenswrapper[4885]: I1002 01:48:19.634172 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:19 crc kubenswrapper[4885]: I1002 01:48:19.634191 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:19 crc kubenswrapper[4885]: I1002 01:48:19.634218 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:19 crc kubenswrapper[4885]: I1002 01:48:19.634236 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:19Z","lastTransitionTime":"2025-10-02T01:48:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:19 crc kubenswrapper[4885]: I1002 01:48:19.738027 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:19 crc kubenswrapper[4885]: I1002 01:48:19.738111 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:19 crc kubenswrapper[4885]: I1002 01:48:19.738128 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:19 crc kubenswrapper[4885]: I1002 01:48:19.738152 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:19 crc kubenswrapper[4885]: I1002 01:48:19.738198 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:19Z","lastTransitionTime":"2025-10-02T01:48:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:19 crc kubenswrapper[4885]: I1002 01:48:19.841798 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:19 crc kubenswrapper[4885]: I1002 01:48:19.841874 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:19 crc kubenswrapper[4885]: I1002 01:48:19.841898 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:19 crc kubenswrapper[4885]: I1002 01:48:19.841929 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:19 crc kubenswrapper[4885]: I1002 01:48:19.841955 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:19Z","lastTransitionTime":"2025-10-02T01:48:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:19 crc kubenswrapper[4885]: I1002 01:48:19.945220 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:19 crc kubenswrapper[4885]: I1002 01:48:19.945311 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:19 crc kubenswrapper[4885]: I1002 01:48:19.945335 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:19 crc kubenswrapper[4885]: I1002 01:48:19.945363 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:19 crc kubenswrapper[4885]: I1002 01:48:19.945384 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:19Z","lastTransitionTime":"2025-10-02T01:48:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:20 crc kubenswrapper[4885]: I1002 01:48:20.046554 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:48:20 crc kubenswrapper[4885]: I1002 01:48:20.046616 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:48:20 crc kubenswrapper[4885]: E1002 01:48:20.046714 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:48:20 crc kubenswrapper[4885]: I1002 01:48:20.046744 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sv4md" Oct 02 01:48:20 crc kubenswrapper[4885]: I1002 01:48:20.046581 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:48:20 crc kubenswrapper[4885]: E1002 01:48:20.046841 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:48:20 crc kubenswrapper[4885]: E1002 01:48:20.047156 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:48:20 crc kubenswrapper[4885]: E1002 01:48:20.047406 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sv4md" podUID="724a3e49-001c-45a4-a896-82c13e0ee0e7" Oct 02 01:48:20 crc kubenswrapper[4885]: I1002 01:48:20.048018 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:20 crc kubenswrapper[4885]: I1002 01:48:20.048041 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:20 crc kubenswrapper[4885]: I1002 01:48:20.048049 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:20 crc kubenswrapper[4885]: I1002 01:48:20.048062 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:20 crc kubenswrapper[4885]: I1002 01:48:20.048071 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:20Z","lastTransitionTime":"2025-10-02T01:48:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:20 crc kubenswrapper[4885]: I1002 01:48:20.151064 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:20 crc kubenswrapper[4885]: I1002 01:48:20.151120 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:20 crc kubenswrapper[4885]: I1002 01:48:20.151137 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:20 crc kubenswrapper[4885]: I1002 01:48:20.151162 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:20 crc kubenswrapper[4885]: I1002 01:48:20.151179 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:20Z","lastTransitionTime":"2025-10-02T01:48:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:20 crc kubenswrapper[4885]: I1002 01:48:20.254569 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:20 crc kubenswrapper[4885]: I1002 01:48:20.254648 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:20 crc kubenswrapper[4885]: I1002 01:48:20.254669 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:20 crc kubenswrapper[4885]: I1002 01:48:20.254695 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:20 crc kubenswrapper[4885]: I1002 01:48:20.254713 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:20Z","lastTransitionTime":"2025-10-02T01:48:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:20 crc kubenswrapper[4885]: I1002 01:48:20.256396 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:20 crc kubenswrapper[4885]: I1002 01:48:20.256454 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:20 crc kubenswrapper[4885]: I1002 01:48:20.256473 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:20 crc kubenswrapper[4885]: I1002 01:48:20.256500 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:20 crc kubenswrapper[4885]: I1002 01:48:20.256524 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:20Z","lastTransitionTime":"2025-10-02T01:48:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:20 crc kubenswrapper[4885]: E1002 01:48:20.277990 4885 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:48:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:48:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:48:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:48:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:48:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:48:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:48:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:48:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"92603ae5-ba38-4201-98e9-979fd56cb50d\\\",\\\"systemUUID\\\":\\\"50ddf2c2-ea95-4960-ae48-d17e9484ee2f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:20Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:20 crc kubenswrapper[4885]: I1002 01:48:20.284995 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:20 crc kubenswrapper[4885]: I1002 01:48:20.285087 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:20 crc kubenswrapper[4885]: I1002 01:48:20.285112 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:20 crc kubenswrapper[4885]: I1002 01:48:20.285146 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:20 crc kubenswrapper[4885]: I1002 01:48:20.285168 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:20Z","lastTransitionTime":"2025-10-02T01:48:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:20 crc kubenswrapper[4885]: E1002 01:48:20.311167 4885 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:48:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:48:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:48:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:48:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:48:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:48:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:48:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:48:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"92603ae5-ba38-4201-98e9-979fd56cb50d\\\",\\\"systemUUID\\\":\\\"50ddf2c2-ea95-4960-ae48-d17e9484ee2f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:20Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:20 crc kubenswrapper[4885]: I1002 01:48:20.317811 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:20 crc kubenswrapper[4885]: I1002 01:48:20.317874 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:20 crc kubenswrapper[4885]: I1002 01:48:20.317890 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:20 crc kubenswrapper[4885]: I1002 01:48:20.317916 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:20 crc kubenswrapper[4885]: I1002 01:48:20.317934 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:20Z","lastTransitionTime":"2025-10-02T01:48:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:20 crc kubenswrapper[4885]: E1002 01:48:20.339121 4885 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:48:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:48:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:48:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:48:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:48:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:48:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:48:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:48:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"92603ae5-ba38-4201-98e9-979fd56cb50d\\\",\\\"systemUUID\\\":\\\"50ddf2c2-ea95-4960-ae48-d17e9484ee2f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:20Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:20 crc kubenswrapper[4885]: I1002 01:48:20.343933 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:20 crc kubenswrapper[4885]: I1002 01:48:20.343987 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:20 crc kubenswrapper[4885]: I1002 01:48:20.344006 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:20 crc kubenswrapper[4885]: I1002 01:48:20.344030 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:20 crc kubenswrapper[4885]: I1002 01:48:20.344047 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:20Z","lastTransitionTime":"2025-10-02T01:48:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:20 crc kubenswrapper[4885]: E1002 01:48:20.362359 4885 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:48:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:48:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:48:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:48:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:48:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:48:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:48:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:48:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"92603ae5-ba38-4201-98e9-979fd56cb50d\\\",\\\"systemUUID\\\":\\\"50ddf2c2-ea95-4960-ae48-d17e9484ee2f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:20Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:20 crc kubenswrapper[4885]: I1002 01:48:20.367498 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:20 crc kubenswrapper[4885]: I1002 01:48:20.367568 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:20 crc kubenswrapper[4885]: I1002 01:48:20.367621 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:20 crc kubenswrapper[4885]: I1002 01:48:20.367654 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:20 crc kubenswrapper[4885]: I1002 01:48:20.367682 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:20Z","lastTransitionTime":"2025-10-02T01:48:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:20 crc kubenswrapper[4885]: E1002 01:48:20.388148 4885 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:48:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:48:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:48:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:48:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:48:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:48:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:48:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:48:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"92603ae5-ba38-4201-98e9-979fd56cb50d\\\",\\\"systemUUID\\\":\\\"50ddf2c2-ea95-4960-ae48-d17e9484ee2f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:20Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:20 crc kubenswrapper[4885]: E1002 01:48:20.388396 4885 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 02 01:48:20 crc kubenswrapper[4885]: I1002 01:48:20.390254 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:20 crc kubenswrapper[4885]: I1002 01:48:20.390343 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:20 crc kubenswrapper[4885]: I1002 01:48:20.390362 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:20 crc kubenswrapper[4885]: I1002 01:48:20.390389 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:20 crc kubenswrapper[4885]: I1002 01:48:20.390410 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:20Z","lastTransitionTime":"2025-10-02T01:48:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:20 crc kubenswrapper[4885]: I1002 01:48:20.493855 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:20 crc kubenswrapper[4885]: I1002 01:48:20.493910 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:20 crc kubenswrapper[4885]: I1002 01:48:20.493926 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:20 crc kubenswrapper[4885]: I1002 01:48:20.493951 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:20 crc kubenswrapper[4885]: I1002 01:48:20.493973 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:20Z","lastTransitionTime":"2025-10-02T01:48:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:20 crc kubenswrapper[4885]: I1002 01:48:20.597118 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:20 crc kubenswrapper[4885]: I1002 01:48:20.597187 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:20 crc kubenswrapper[4885]: I1002 01:48:20.597204 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:20 crc kubenswrapper[4885]: I1002 01:48:20.597231 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:20 crc kubenswrapper[4885]: I1002 01:48:20.597249 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:20Z","lastTransitionTime":"2025-10-02T01:48:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:20 crc kubenswrapper[4885]: I1002 01:48:20.700013 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:20 crc kubenswrapper[4885]: I1002 01:48:20.700070 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:20 crc kubenswrapper[4885]: I1002 01:48:20.700089 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:20 crc kubenswrapper[4885]: I1002 01:48:20.700115 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:20 crc kubenswrapper[4885]: I1002 01:48:20.700133 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:20Z","lastTransitionTime":"2025-10-02T01:48:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:20 crc kubenswrapper[4885]: I1002 01:48:20.803509 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:20 crc kubenswrapper[4885]: I1002 01:48:20.803575 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:20 crc kubenswrapper[4885]: I1002 01:48:20.803593 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:20 crc kubenswrapper[4885]: I1002 01:48:20.803618 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:20 crc kubenswrapper[4885]: I1002 01:48:20.803635 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:20Z","lastTransitionTime":"2025-10-02T01:48:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:20 crc kubenswrapper[4885]: I1002 01:48:20.912874 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:20 crc kubenswrapper[4885]: I1002 01:48:20.912938 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:20 crc kubenswrapper[4885]: I1002 01:48:20.912957 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:20 crc kubenswrapper[4885]: I1002 01:48:20.912981 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:20 crc kubenswrapper[4885]: I1002 01:48:20.912998 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:20Z","lastTransitionTime":"2025-10-02T01:48:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:21 crc kubenswrapper[4885]: I1002 01:48:21.016019 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:21 crc kubenswrapper[4885]: I1002 01:48:21.016072 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:21 crc kubenswrapper[4885]: I1002 01:48:21.016089 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:21 crc kubenswrapper[4885]: I1002 01:48:21.016110 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:21 crc kubenswrapper[4885]: I1002 01:48:21.016127 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:21Z","lastTransitionTime":"2025-10-02T01:48:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:21 crc kubenswrapper[4885]: I1002 01:48:21.119133 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:21 crc kubenswrapper[4885]: I1002 01:48:21.119182 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:21 crc kubenswrapper[4885]: I1002 01:48:21.119200 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:21 crc kubenswrapper[4885]: I1002 01:48:21.119220 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:21 crc kubenswrapper[4885]: I1002 01:48:21.119236 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:21Z","lastTransitionTime":"2025-10-02T01:48:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:21 crc kubenswrapper[4885]: I1002 01:48:21.222105 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:21 crc kubenswrapper[4885]: I1002 01:48:21.222172 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:21 crc kubenswrapper[4885]: I1002 01:48:21.222190 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:21 crc kubenswrapper[4885]: I1002 01:48:21.222215 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:21 crc kubenswrapper[4885]: I1002 01:48:21.222233 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:21Z","lastTransitionTime":"2025-10-02T01:48:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:21 crc kubenswrapper[4885]: I1002 01:48:21.326052 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:21 crc kubenswrapper[4885]: I1002 01:48:21.326127 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:21 crc kubenswrapper[4885]: I1002 01:48:21.326152 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:21 crc kubenswrapper[4885]: I1002 01:48:21.326179 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:21 crc kubenswrapper[4885]: I1002 01:48:21.326197 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:21Z","lastTransitionTime":"2025-10-02T01:48:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:21 crc kubenswrapper[4885]: I1002 01:48:21.429417 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:21 crc kubenswrapper[4885]: I1002 01:48:21.429486 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:21 crc kubenswrapper[4885]: I1002 01:48:21.429502 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:21 crc kubenswrapper[4885]: I1002 01:48:21.429527 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:21 crc kubenswrapper[4885]: I1002 01:48:21.429545 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:21Z","lastTransitionTime":"2025-10-02T01:48:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:21 crc kubenswrapper[4885]: I1002 01:48:21.532973 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:21 crc kubenswrapper[4885]: I1002 01:48:21.533032 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:21 crc kubenswrapper[4885]: I1002 01:48:21.533048 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:21 crc kubenswrapper[4885]: I1002 01:48:21.533073 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:21 crc kubenswrapper[4885]: I1002 01:48:21.533089 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:21Z","lastTransitionTime":"2025-10-02T01:48:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:21 crc kubenswrapper[4885]: I1002 01:48:21.636529 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:21 crc kubenswrapper[4885]: I1002 01:48:21.636574 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:21 crc kubenswrapper[4885]: I1002 01:48:21.636592 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:21 crc kubenswrapper[4885]: I1002 01:48:21.636612 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:21 crc kubenswrapper[4885]: I1002 01:48:21.636668 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:21Z","lastTransitionTime":"2025-10-02T01:48:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:21 crc kubenswrapper[4885]: I1002 01:48:21.739927 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:21 crc kubenswrapper[4885]: I1002 01:48:21.740024 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:21 crc kubenswrapper[4885]: I1002 01:48:21.740056 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:21 crc kubenswrapper[4885]: I1002 01:48:21.740087 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:21 crc kubenswrapper[4885]: I1002 01:48:21.740238 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:21Z","lastTransitionTime":"2025-10-02T01:48:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:21 crc kubenswrapper[4885]: I1002 01:48:21.843663 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:21 crc kubenswrapper[4885]: I1002 01:48:21.843720 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:21 crc kubenswrapper[4885]: I1002 01:48:21.843737 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:21 crc kubenswrapper[4885]: I1002 01:48:21.843760 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:21 crc kubenswrapper[4885]: I1002 01:48:21.843777 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:21Z","lastTransitionTime":"2025-10-02T01:48:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:21 crc kubenswrapper[4885]: I1002 01:48:21.947086 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:21 crc kubenswrapper[4885]: I1002 01:48:21.947148 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:21 crc kubenswrapper[4885]: I1002 01:48:21.947165 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:21 crc kubenswrapper[4885]: I1002 01:48:21.947191 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:21 crc kubenswrapper[4885]: I1002 01:48:21.947209 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:21Z","lastTransitionTime":"2025-10-02T01:48:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:22 crc kubenswrapper[4885]: I1002 01:48:22.046549 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sv4md" Oct 02 01:48:22 crc kubenswrapper[4885]: E1002 01:48:22.046961 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sv4md" podUID="724a3e49-001c-45a4-a896-82c13e0ee0e7" Oct 02 01:48:22 crc kubenswrapper[4885]: I1002 01:48:22.047554 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:48:22 crc kubenswrapper[4885]: I1002 01:48:22.047688 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:48:22 crc kubenswrapper[4885]: I1002 01:48:22.047555 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:48:22 crc kubenswrapper[4885]: E1002 01:48:22.048472 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:48:22 crc kubenswrapper[4885]: E1002 01:48:22.048481 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:48:22 crc kubenswrapper[4885]: E1002 01:48:22.048681 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:48:22 crc kubenswrapper[4885]: I1002 01:48:22.049743 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:22 crc kubenswrapper[4885]: I1002 01:48:22.049783 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:22 crc kubenswrapper[4885]: I1002 01:48:22.049801 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:22 crc kubenswrapper[4885]: I1002 01:48:22.049825 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:22 crc kubenswrapper[4885]: I1002 01:48:22.049845 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:22Z","lastTransitionTime":"2025-10-02T01:48:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:22 crc kubenswrapper[4885]: I1002 01:48:22.070715 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a0bb885-2eb0-419c-8a7a-ef634cb3c144\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a89b49cbe08d927629564e607de13655ddf3d01ac4eb5d688b325210306c120c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06c5942805a498d9621a906088aa54538b4eed583d1b8d0e3ba2a6718378a36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06b9f6f7fe44d5771a838d0b89c5bdea3ebb79023ae2675ee907ec2766b3d35a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://178b4340ee6904c60c814b99d83ee225161aeb23e438d8d02f856579a13ec4e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bb63bf30d6ecce9e71f0db0c16632bdfa6b33924b673378eb0c38ab229fa837\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T01:47:06Z\\\",\\\"message\\\":\\\"W1002 01:46:55.406027 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1002 01:46:55.406714 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759369615 cert, and key in /tmp/serving-cert-3795040473/serving-signer.crt, /tmp/serving-cert-3795040473/serving-signer.key\\\\nI1002 01:46:55.614670 1 observer_polling.go:159] Starting file observer\\\\nW1002 01:46:55.617662 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1002 01:46:55.618112 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 01:46:55.620347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3795040473/tls.crt::/tmp/serving-cert-3795040473/tls.key\\\\\\\"\\\\nF1002 01:47:06.224459 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90535f97ab2ed4626d47540ebd7baada837fd332140323bec0f49aa38d3c0009\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://592f89a326c9c1559193a71904b47ca27b1dbe1a7b30e03bc48924bb2f35f921\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://592f89a326c9c1559193a71904b47ca27b1dbe1a7b30e03bc48924bb2f35f921\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:22Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:22 crc kubenswrapper[4885]: I1002 01:48:22.088537 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d86d336f-d1f2-46b0-a1e2-9cb04c62f94e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://614992bc938987be5169d990ded500acb5120eb0ddfc221168faa0d8c16e0972\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25c85ae265d3c8450cd878ca9e4452c40df15c44493535be916b7500f1c5bac0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://25c85ae265d3c8450cd878ca9e4452c40df15c44493535be916b7500f1c5bac0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:22Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:22 crc kubenswrapper[4885]: I1002 01:48:22.110568 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b087880e05039816dcab154643ac72c7e4a0a1124817c51dc2c639b62bbee0fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:22Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:22 crc kubenswrapper[4885]: I1002 01:48:22.131012 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:22Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:22 crc kubenswrapper[4885]: I1002 01:48:22.148792 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:22Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:22 crc kubenswrapper[4885]: I1002 01:48:22.153503 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:22 crc kubenswrapper[4885]: I1002 01:48:22.153629 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:22 crc kubenswrapper[4885]: I1002 01:48:22.153703 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:22 crc kubenswrapper[4885]: I1002 01:48:22.153746 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:22 crc kubenswrapper[4885]: I1002 01:48:22.153825 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:22Z","lastTransitionTime":"2025-10-02T01:48:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:22 crc kubenswrapper[4885]: I1002 01:48:22.165611 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-58lhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac6a19c4-f259-4dbd-a016-d6c2c849c6cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed76c0f968c063f7795d363af194a222487b647ceda234de0dabb39b25eafe0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hzc4c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-58lhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:22Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:22 crc kubenswrapper[4885]: I1002 01:48:22.182382 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d169c8ba-000a-4218-8e3b-4ae53035b110\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61e436b7876107ce5ea9e668c0bbc367be6e663fc804fa3a65b51adf8be8d51f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-77rzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09536e182a46e77318c10812d1777e551f834907c006cab5c726dc66f5bce88a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-77rzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rttx8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:22Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:22 crc kubenswrapper[4885]: I1002 01:48:22.198053 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jgdvt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81733e8d-67b0-46fa-8a41-08b40645bd3e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcef60c816ca1ba9a298da658e84c4d717bea3d352c4c2f807ebd8e56327b384\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cwbl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jgdvt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:22Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:22 crc kubenswrapper[4885]: I1002 01:48:22.218092 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"506a599f-fb1c-4d14-be88-0e4833f3d624\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31c2f0193ffef54965b79a3569b0f6f4547e540bac0774155159a1c996c196c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://533e73e5f90b7258acbcdf018ee01323c118f9dc45035b25b065ba01cabecccf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c00db833a593ec40eb63a6b383455e19e960c2e98b9b66f76877ee993057e7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21e2f462a0deb0d740185663b2014f4a1e57b4fdbeb82571264b002f928035bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:22Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:22 crc kubenswrapper[4885]: I1002 01:48:22.248550 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e4679dd-f870-41e9-a32b-360f5d2ee81b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4310909ad9c7b609ed6814e04a525a6afae747775598e7275080413dec8b23a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7642033b0b5995d332ccf7992d6d7e6a5e9c9369ae2d1071ccb2f536938c8f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9f99f9e26a04fb42b01a0783a5c4e360514818ce9102edc7fa08626d50b9a34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6fa72b030d309656b73121e4ac2a3924a4e449aad66683dd9de5ee5f22d5a4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef403e519d197fb37b3e6b76a14a5128ba3f30a7a5e8bf0a420cf6a39b89575f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8482ba457683ffa03100ec1f07ea9d3f918be5e490b5ba05c2ba965d03033191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ef80e838accc59ad7a579c66f4fcd6881d23fa86fc5922d4b0085d211d5d94b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ef80e838accc59ad7a579c66f4fcd6881d23fa86fc5922d4b0085d211d5d94b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T01:48:11Z\\\",\\\"message\\\":\\\" 6940 services_controller.go:451] Built service openshift-operator-lifecycle-manager/catalog-operator-metrics cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-operator-lifecycle-manager/catalog-operator-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-operator-lifecycle-manager/catalog-operator-metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.204\\\\\\\", Port:8443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1002 01:48:11.048667 6940 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:48:10Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-dngcm_openshift-ovn-kubernetes(9e4679dd-f870-41e9-a32b-360f5d2ee81b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fafdd1094bd02786d745c984989852b0c690db38655f92b419733a2004f417c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0125e907d061a97286026ee823c4eed303734ed03ab4864f742592b23a1aa2f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0125e907d061a97286026ee823c4eed303734ed03ab4864f742592b23a1aa2f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4b7jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dngcm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:22Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:22 crc kubenswrapper[4885]: I1002 01:48:22.256202 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:22 crc kubenswrapper[4885]: I1002 01:48:22.256255 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:22 crc kubenswrapper[4885]: I1002 01:48:22.256308 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:22 crc kubenswrapper[4885]: I1002 01:48:22.256334 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:22 crc kubenswrapper[4885]: I1002 01:48:22.256353 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:22Z","lastTransitionTime":"2025-10-02T01:48:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:22 crc kubenswrapper[4885]: I1002 01:48:22.265109 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-sv4md" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"724a3e49-001c-45a4-a896-82c13e0ee0e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hpqgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hpqgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-sv4md\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:22Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:22 crc kubenswrapper[4885]: I1002 01:48:22.304729 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18466979-7c7c-4d73-86c6-b8b6eaf20484\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b742e4b288bb50569d2fd23f3f9857d649a46063998170eee582ecbbba64d190\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://298c2ace58c4a9d09cfdc864316317e370bb0ac5735d631af6d85c6298f686d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://620f54e02c463f6e603808e5e57e9e5770d125578cc258ba16117a2416e57d3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b4cd17b64e5657ae05e71f3ca2340cb0d7965b1b49d1f7b4bfe4494292293ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b4cd17b64e5657ae05e71f3ca2340cb0d7965b1b49d1f7b4bfe4494292293ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:22Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:22 crc kubenswrapper[4885]: I1002 01:48:22.336957 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8070bc1c-e3de-4e9a-985e-40d1a9858056\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3e634b3727c517cc48b17c2cfb38f26460f53d1bf25d631d08ec9acb8d780e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4adcc537c23697a324d82af94c31059dd444245efdf7c978db58ede256e112ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edcd1d99644a778dfb6d0a17d2a4abd906400fe8209e611ebf8d7a9abe7ccf28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://97cb59adc2ddfa365ebc9dec649600dc90466ffe59134a6d15a9d3134a7c9b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d2808bc658ea80d68469ed16422d56ef19447f4c90b4af7b2597c64406f8a44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://83acbc20919f7c1958b6d492fadb203855df02761662425df9f91a0757fe2bf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83acbc20919f7c1958b6d492fadb203855df02761662425df9f91a0757fe2bf7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e86ecd03835efc9a253b6804bc61e6ac690e75b8d1f3cc77f5647730e586574\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e86ecd03835efc9a253b6804bc61e6ac690e75b8d1f3cc77f5647730e586574\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://48f95af51feaf014a18496bdfadd59d7f1e11167c5c8b677fa7037e48368ff2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48f95af51feaf014a18496bdfadd59d7f1e11167c5c8b677fa7037e48368ff2a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:46:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:46:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:46:52Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:22Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:22 crc kubenswrapper[4885]: I1002 01:48:22.354896 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5c1f95f188569c9aa47e51f2916939fa0b65a5cdf4c16b798612d41f3c8411a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:22Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:22 crc kubenswrapper[4885]: I1002 01:48:22.359719 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:22 crc kubenswrapper[4885]: I1002 01:48:22.359798 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:22 crc kubenswrapper[4885]: I1002 01:48:22.359819 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:22 crc kubenswrapper[4885]: I1002 01:48:22.359846 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:22 crc kubenswrapper[4885]: I1002 01:48:22.359867 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:22Z","lastTransitionTime":"2025-10-02T01:48:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:22 crc kubenswrapper[4885]: I1002 01:48:22.376386 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vhgbd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df9792d80c807ed98c8442822d01f619327361c40425abfe26f11e5c27ba2584\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c63a7d20d39b65acfb5a9be4c28b21c72111b398bea76c238e15ddb4dd63e139\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T01:48:00Z\\\",\\\"message\\\":\\\"2025-10-02T01:47:15+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_9f42b0ef-8651-49db-8a31-d21c0c1c47e7\\\\n2025-10-02T01:47:15+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_9f42b0ef-8651-49db-8a31-d21c0c1c47e7 to /host/opt/cni/bin/\\\\n2025-10-02T01:47:15Z [verbose] multus-daemon started\\\\n2025-10-02T01:47:15Z [verbose] Readiness Indicator file check\\\\n2025-10-02T01:48:00Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7zwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vhgbd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:22Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:22 crc kubenswrapper[4885]: I1002 01:48:22.394696 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rrn8b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52967a77-525c-4537-9b1f-d7bb754a8494\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3dee451ebe9006bf1c519fed4c65267570b6139d76841858352e087ae78d26a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtr74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e9412788d6df7332592beac48281822b26e1aa19b39addce8e1be556af9d982\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtr74\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:26Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rrn8b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:22Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:22 crc kubenswrapper[4885]: I1002 01:48:22.413850 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:22Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:22 crc kubenswrapper[4885]: I1002 01:48:22.434332 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74adcf912ccedb6b5d4f2803fd215e3e2ef62367597aa6da528354c4511ee536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7403ef968d040ac09d722b7557438bef23892d578992826f42b700b370d495d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:22Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:22 crc kubenswrapper[4885]: I1002 01:48:22.456885 4885 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2ll9b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52b39712-5fa1-4fe0-814b-b170a6e3938b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9a96383d0fbfd116c419054aaf1a7adba1937e8b0bad294cefd68b4ab0d2298\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c979b91ba748aa4647826a8926337662ca929c54090fb62285572819ec2fbf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c979b91ba748aa4647826a8926337662ca929c54090fb62285572819ec2fbf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97d9d693e8e60ee9c183bb44461e698e8b3cfd20a516605aa8754c9205d2bdc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://97d9d693e8e60ee9c183bb44461e698e8b3cfd20a516605aa8754c9205d2bdc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c44252676359c671d65c9588b05572d76c99cf4c90666f0ac089b2f778450a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c44252676359c671d65c9588b05572d76c99cf4c90666f0ac089b2f778450a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80c982f07874df686fe7e29cc91043eb6d58f1be5fc93cacd46d33a21dd9b622\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80c982f07874df686fe7e29cc91043eb6d58f1be5fc93cacd46d33a21dd9b622\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a354c985652c7ad7fa8944fbfdfe223cfb0f97fb11eecd9bd429020d4892b78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a354c985652c7ad7fa8944fbfdfe223cfb0f97fb11eecd9bd429020d4892b78\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2efd7283df313d086c1b8b88f289ace28764de256205bee7610d6711002d8515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2efd7283df313d086c1b8b88f289ace28764de256205bee7610d6711002d8515\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-928wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2ll9b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:48:22Z is after 2025-08-24T17:21:41Z" Oct 02 01:48:22 crc kubenswrapper[4885]: I1002 01:48:22.463953 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:22 crc kubenswrapper[4885]: I1002 01:48:22.464034 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:22 crc kubenswrapper[4885]: I1002 01:48:22.464054 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:22 crc kubenswrapper[4885]: I1002 01:48:22.464104 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:22 crc kubenswrapper[4885]: I1002 01:48:22.464120 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:22Z","lastTransitionTime":"2025-10-02T01:48:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:22 crc kubenswrapper[4885]: I1002 01:48:22.567589 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:22 crc kubenswrapper[4885]: I1002 01:48:22.567647 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:22 crc kubenswrapper[4885]: I1002 01:48:22.567666 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:22 crc kubenswrapper[4885]: I1002 01:48:22.567693 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:22 crc kubenswrapper[4885]: I1002 01:48:22.567712 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:22Z","lastTransitionTime":"2025-10-02T01:48:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:22 crc kubenswrapper[4885]: I1002 01:48:22.670762 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:22 crc kubenswrapper[4885]: I1002 01:48:22.670829 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:22 crc kubenswrapper[4885]: I1002 01:48:22.670853 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:22 crc kubenswrapper[4885]: I1002 01:48:22.670883 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:22 crc kubenswrapper[4885]: I1002 01:48:22.670909 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:22Z","lastTransitionTime":"2025-10-02T01:48:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:22 crc kubenswrapper[4885]: I1002 01:48:22.774224 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:22 crc kubenswrapper[4885]: I1002 01:48:22.774316 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:22 crc kubenswrapper[4885]: I1002 01:48:22.774334 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:22 crc kubenswrapper[4885]: I1002 01:48:22.774358 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:22 crc kubenswrapper[4885]: I1002 01:48:22.774376 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:22Z","lastTransitionTime":"2025-10-02T01:48:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:22 crc kubenswrapper[4885]: I1002 01:48:22.877251 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:22 crc kubenswrapper[4885]: I1002 01:48:22.877311 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:22 crc kubenswrapper[4885]: I1002 01:48:22.877320 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:22 crc kubenswrapper[4885]: I1002 01:48:22.877335 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:22 crc kubenswrapper[4885]: I1002 01:48:22.877346 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:22Z","lastTransitionTime":"2025-10-02T01:48:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:22 crc kubenswrapper[4885]: I1002 01:48:22.980710 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:22 crc kubenswrapper[4885]: I1002 01:48:22.980771 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:22 crc kubenswrapper[4885]: I1002 01:48:22.980789 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:22 crc kubenswrapper[4885]: I1002 01:48:22.980813 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:22 crc kubenswrapper[4885]: I1002 01:48:22.980830 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:22Z","lastTransitionTime":"2025-10-02T01:48:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:23 crc kubenswrapper[4885]: I1002 01:48:23.083970 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:23 crc kubenswrapper[4885]: I1002 01:48:23.084031 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:23 crc kubenswrapper[4885]: I1002 01:48:23.084051 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:23 crc kubenswrapper[4885]: I1002 01:48:23.084075 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:23 crc kubenswrapper[4885]: I1002 01:48:23.084095 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:23Z","lastTransitionTime":"2025-10-02T01:48:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:23 crc kubenswrapper[4885]: I1002 01:48:23.187219 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:23 crc kubenswrapper[4885]: I1002 01:48:23.187313 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:23 crc kubenswrapper[4885]: I1002 01:48:23.187335 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:23 crc kubenswrapper[4885]: I1002 01:48:23.187362 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:23 crc kubenswrapper[4885]: I1002 01:48:23.187378 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:23Z","lastTransitionTime":"2025-10-02T01:48:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:23 crc kubenswrapper[4885]: I1002 01:48:23.290482 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:23 crc kubenswrapper[4885]: I1002 01:48:23.290548 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:23 crc kubenswrapper[4885]: I1002 01:48:23.290568 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:23 crc kubenswrapper[4885]: I1002 01:48:23.290594 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:23 crc kubenswrapper[4885]: I1002 01:48:23.290612 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:23Z","lastTransitionTime":"2025-10-02T01:48:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:23 crc kubenswrapper[4885]: I1002 01:48:23.392858 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:23 crc kubenswrapper[4885]: I1002 01:48:23.392896 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:23 crc kubenswrapper[4885]: I1002 01:48:23.392905 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:23 crc kubenswrapper[4885]: I1002 01:48:23.392917 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:23 crc kubenswrapper[4885]: I1002 01:48:23.392925 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:23Z","lastTransitionTime":"2025-10-02T01:48:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:23 crc kubenswrapper[4885]: I1002 01:48:23.495884 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:23 crc kubenswrapper[4885]: I1002 01:48:23.495951 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:23 crc kubenswrapper[4885]: I1002 01:48:23.495970 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:23 crc kubenswrapper[4885]: I1002 01:48:23.495998 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:23 crc kubenswrapper[4885]: I1002 01:48:23.496017 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:23Z","lastTransitionTime":"2025-10-02T01:48:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:23 crc kubenswrapper[4885]: I1002 01:48:23.598904 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:23 crc kubenswrapper[4885]: I1002 01:48:23.598992 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:23 crc kubenswrapper[4885]: I1002 01:48:23.599015 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:23 crc kubenswrapper[4885]: I1002 01:48:23.599054 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:23 crc kubenswrapper[4885]: I1002 01:48:23.599078 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:23Z","lastTransitionTime":"2025-10-02T01:48:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:23 crc kubenswrapper[4885]: I1002 01:48:23.702311 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:23 crc kubenswrapper[4885]: I1002 01:48:23.702364 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:23 crc kubenswrapper[4885]: I1002 01:48:23.702376 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:23 crc kubenswrapper[4885]: I1002 01:48:23.702397 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:23 crc kubenswrapper[4885]: I1002 01:48:23.702410 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:23Z","lastTransitionTime":"2025-10-02T01:48:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:23 crc kubenswrapper[4885]: I1002 01:48:23.805238 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:23 crc kubenswrapper[4885]: I1002 01:48:23.805305 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:23 crc kubenswrapper[4885]: I1002 01:48:23.805314 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:23 crc kubenswrapper[4885]: I1002 01:48:23.805328 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:23 crc kubenswrapper[4885]: I1002 01:48:23.805337 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:23Z","lastTransitionTime":"2025-10-02T01:48:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:23 crc kubenswrapper[4885]: I1002 01:48:23.908919 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:23 crc kubenswrapper[4885]: I1002 01:48:23.908977 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:23 crc kubenswrapper[4885]: I1002 01:48:23.908994 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:23 crc kubenswrapper[4885]: I1002 01:48:23.909021 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:23 crc kubenswrapper[4885]: I1002 01:48:23.909039 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:23Z","lastTransitionTime":"2025-10-02T01:48:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:24 crc kubenswrapper[4885]: I1002 01:48:24.011577 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:24 crc kubenswrapper[4885]: I1002 01:48:24.011658 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:24 crc kubenswrapper[4885]: I1002 01:48:24.011682 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:24 crc kubenswrapper[4885]: I1002 01:48:24.011714 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:24 crc kubenswrapper[4885]: I1002 01:48:24.011739 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:24Z","lastTransitionTime":"2025-10-02T01:48:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:24 crc kubenswrapper[4885]: I1002 01:48:24.046557 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sv4md" Oct 02 01:48:24 crc kubenswrapper[4885]: I1002 01:48:24.046604 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:48:24 crc kubenswrapper[4885]: I1002 01:48:24.046638 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:48:24 crc kubenswrapper[4885]: E1002 01:48:24.046737 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sv4md" podUID="724a3e49-001c-45a4-a896-82c13e0ee0e7" Oct 02 01:48:24 crc kubenswrapper[4885]: I1002 01:48:24.046763 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:48:24 crc kubenswrapper[4885]: E1002 01:48:24.046849 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:48:24 crc kubenswrapper[4885]: E1002 01:48:24.047178 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:48:24 crc kubenswrapper[4885]: E1002 01:48:24.047349 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:48:24 crc kubenswrapper[4885]: I1002 01:48:24.114347 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:24 crc kubenswrapper[4885]: I1002 01:48:24.114505 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:24 crc kubenswrapper[4885]: I1002 01:48:24.114524 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:24 crc kubenswrapper[4885]: I1002 01:48:24.114549 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:24 crc kubenswrapper[4885]: I1002 01:48:24.114567 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:24Z","lastTransitionTime":"2025-10-02T01:48:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:24 crc kubenswrapper[4885]: I1002 01:48:24.218176 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:24 crc kubenswrapper[4885]: I1002 01:48:24.218321 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:24 crc kubenswrapper[4885]: I1002 01:48:24.218345 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:24 crc kubenswrapper[4885]: I1002 01:48:24.218399 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:24 crc kubenswrapper[4885]: I1002 01:48:24.218419 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:24Z","lastTransitionTime":"2025-10-02T01:48:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:24 crc kubenswrapper[4885]: I1002 01:48:24.321865 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:24 crc kubenswrapper[4885]: I1002 01:48:24.321922 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:24 crc kubenswrapper[4885]: I1002 01:48:24.321938 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:24 crc kubenswrapper[4885]: I1002 01:48:24.321961 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:24 crc kubenswrapper[4885]: I1002 01:48:24.321979 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:24Z","lastTransitionTime":"2025-10-02T01:48:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:24 crc kubenswrapper[4885]: I1002 01:48:24.425482 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:24 crc kubenswrapper[4885]: I1002 01:48:24.425544 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:24 crc kubenswrapper[4885]: I1002 01:48:24.425561 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:24 crc kubenswrapper[4885]: I1002 01:48:24.425583 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:24 crc kubenswrapper[4885]: I1002 01:48:24.425601 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:24Z","lastTransitionTime":"2025-10-02T01:48:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:24 crc kubenswrapper[4885]: I1002 01:48:24.528772 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:24 crc kubenswrapper[4885]: I1002 01:48:24.528824 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:24 crc kubenswrapper[4885]: I1002 01:48:24.528845 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:24 crc kubenswrapper[4885]: I1002 01:48:24.528867 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:24 crc kubenswrapper[4885]: I1002 01:48:24.528885 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:24Z","lastTransitionTime":"2025-10-02T01:48:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:24 crc kubenswrapper[4885]: I1002 01:48:24.639072 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:24 crc kubenswrapper[4885]: I1002 01:48:24.639147 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:24 crc kubenswrapper[4885]: I1002 01:48:24.639165 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:24 crc kubenswrapper[4885]: I1002 01:48:24.639192 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:24 crc kubenswrapper[4885]: I1002 01:48:24.639214 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:24Z","lastTransitionTime":"2025-10-02T01:48:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:24 crc kubenswrapper[4885]: I1002 01:48:24.742188 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:24 crc kubenswrapper[4885]: I1002 01:48:24.742241 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:24 crc kubenswrapper[4885]: I1002 01:48:24.742290 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:24 crc kubenswrapper[4885]: I1002 01:48:24.742314 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:24 crc kubenswrapper[4885]: I1002 01:48:24.742332 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:24Z","lastTransitionTime":"2025-10-02T01:48:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:24 crc kubenswrapper[4885]: I1002 01:48:24.846054 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:24 crc kubenswrapper[4885]: I1002 01:48:24.846150 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:24 crc kubenswrapper[4885]: I1002 01:48:24.846169 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:24 crc kubenswrapper[4885]: I1002 01:48:24.846197 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:24 crc kubenswrapper[4885]: I1002 01:48:24.846214 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:24Z","lastTransitionTime":"2025-10-02T01:48:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:24 crc kubenswrapper[4885]: I1002 01:48:24.948812 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:24 crc kubenswrapper[4885]: I1002 01:48:24.948884 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:24 crc kubenswrapper[4885]: I1002 01:48:24.948907 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:24 crc kubenswrapper[4885]: I1002 01:48:24.948936 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:24 crc kubenswrapper[4885]: I1002 01:48:24.948961 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:24Z","lastTransitionTime":"2025-10-02T01:48:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:25 crc kubenswrapper[4885]: I1002 01:48:25.047585 4885 scope.go:117] "RemoveContainer" containerID="2ef80e838accc59ad7a579c66f4fcd6881d23fa86fc5922d4b0085d211d5d94b" Oct 02 01:48:25 crc kubenswrapper[4885]: E1002 01:48:25.047848 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-dngcm_openshift-ovn-kubernetes(9e4679dd-f870-41e9-a32b-360f5d2ee81b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" podUID="9e4679dd-f870-41e9-a32b-360f5d2ee81b" Oct 02 01:48:25 crc kubenswrapper[4885]: I1002 01:48:25.051787 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:25 crc kubenswrapper[4885]: I1002 01:48:25.051838 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:25 crc kubenswrapper[4885]: I1002 01:48:25.051857 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:25 crc kubenswrapper[4885]: I1002 01:48:25.051887 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:25 crc kubenswrapper[4885]: I1002 01:48:25.051903 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:25Z","lastTransitionTime":"2025-10-02T01:48:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:25 crc kubenswrapper[4885]: I1002 01:48:25.155056 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:25 crc kubenswrapper[4885]: I1002 01:48:25.155134 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:25 crc kubenswrapper[4885]: I1002 01:48:25.155152 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:25 crc kubenswrapper[4885]: I1002 01:48:25.155177 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:25 crc kubenswrapper[4885]: I1002 01:48:25.155194 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:25Z","lastTransitionTime":"2025-10-02T01:48:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:25 crc kubenswrapper[4885]: I1002 01:48:25.257917 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:25 crc kubenswrapper[4885]: I1002 01:48:25.257972 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:25 crc kubenswrapper[4885]: I1002 01:48:25.257989 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:25 crc kubenswrapper[4885]: I1002 01:48:25.258011 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:25 crc kubenswrapper[4885]: I1002 01:48:25.258028 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:25Z","lastTransitionTime":"2025-10-02T01:48:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:25 crc kubenswrapper[4885]: I1002 01:48:25.361120 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:25 crc kubenswrapper[4885]: I1002 01:48:25.361213 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:25 crc kubenswrapper[4885]: I1002 01:48:25.361238 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:25 crc kubenswrapper[4885]: I1002 01:48:25.361298 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:25 crc kubenswrapper[4885]: I1002 01:48:25.361318 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:25Z","lastTransitionTime":"2025-10-02T01:48:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:25 crc kubenswrapper[4885]: I1002 01:48:25.463420 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:25 crc kubenswrapper[4885]: I1002 01:48:25.463484 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:25 crc kubenswrapper[4885]: I1002 01:48:25.463504 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:25 crc kubenswrapper[4885]: I1002 01:48:25.463529 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:25 crc kubenswrapper[4885]: I1002 01:48:25.463546 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:25Z","lastTransitionTime":"2025-10-02T01:48:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:25 crc kubenswrapper[4885]: I1002 01:48:25.567061 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:25 crc kubenswrapper[4885]: I1002 01:48:25.567136 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:25 crc kubenswrapper[4885]: I1002 01:48:25.567156 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:25 crc kubenswrapper[4885]: I1002 01:48:25.567184 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:25 crc kubenswrapper[4885]: I1002 01:48:25.567207 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:25Z","lastTransitionTime":"2025-10-02T01:48:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:25 crc kubenswrapper[4885]: I1002 01:48:25.669831 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:25 crc kubenswrapper[4885]: I1002 01:48:25.669894 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:25 crc kubenswrapper[4885]: I1002 01:48:25.669913 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:25 crc kubenswrapper[4885]: I1002 01:48:25.669940 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:25 crc kubenswrapper[4885]: I1002 01:48:25.669958 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:25Z","lastTransitionTime":"2025-10-02T01:48:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:25 crc kubenswrapper[4885]: I1002 01:48:25.772678 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:25 crc kubenswrapper[4885]: I1002 01:48:25.772732 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:25 crc kubenswrapper[4885]: I1002 01:48:25.772748 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:25 crc kubenswrapper[4885]: I1002 01:48:25.772771 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:25 crc kubenswrapper[4885]: I1002 01:48:25.772789 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:25Z","lastTransitionTime":"2025-10-02T01:48:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:25 crc kubenswrapper[4885]: I1002 01:48:25.875433 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:25 crc kubenswrapper[4885]: I1002 01:48:25.875500 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:25 crc kubenswrapper[4885]: I1002 01:48:25.875518 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:25 crc kubenswrapper[4885]: I1002 01:48:25.875543 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:25 crc kubenswrapper[4885]: I1002 01:48:25.875563 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:25Z","lastTransitionTime":"2025-10-02T01:48:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:25 crc kubenswrapper[4885]: I1002 01:48:25.978910 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:25 crc kubenswrapper[4885]: I1002 01:48:25.978977 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:25 crc kubenswrapper[4885]: I1002 01:48:25.978994 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:25 crc kubenswrapper[4885]: I1002 01:48:25.979020 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:25 crc kubenswrapper[4885]: I1002 01:48:25.979037 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:25Z","lastTransitionTime":"2025-10-02T01:48:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:26 crc kubenswrapper[4885]: I1002 01:48:26.046205 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:48:26 crc kubenswrapper[4885]: I1002 01:48:26.046402 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sv4md" Oct 02 01:48:26 crc kubenswrapper[4885]: I1002 01:48:26.046585 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:48:26 crc kubenswrapper[4885]: E1002 01:48:26.046598 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:48:26 crc kubenswrapper[4885]: I1002 01:48:26.046700 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:48:26 crc kubenswrapper[4885]: E1002 01:48:26.046827 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:48:26 crc kubenswrapper[4885]: E1002 01:48:26.046972 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sv4md" podUID="724a3e49-001c-45a4-a896-82c13e0ee0e7" Oct 02 01:48:26 crc kubenswrapper[4885]: E1002 01:48:26.047315 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:48:26 crc kubenswrapper[4885]: I1002 01:48:26.081814 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:26 crc kubenswrapper[4885]: I1002 01:48:26.081889 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:26 crc kubenswrapper[4885]: I1002 01:48:26.081916 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:26 crc kubenswrapper[4885]: I1002 01:48:26.081942 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:26 crc kubenswrapper[4885]: I1002 01:48:26.081963 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:26Z","lastTransitionTime":"2025-10-02T01:48:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:26 crc kubenswrapper[4885]: I1002 01:48:26.184825 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:26 crc kubenswrapper[4885]: I1002 01:48:26.184887 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:26 crc kubenswrapper[4885]: I1002 01:48:26.184905 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:26 crc kubenswrapper[4885]: I1002 01:48:26.184929 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:26 crc kubenswrapper[4885]: I1002 01:48:26.184946 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:26Z","lastTransitionTime":"2025-10-02T01:48:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:26 crc kubenswrapper[4885]: I1002 01:48:26.288022 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:26 crc kubenswrapper[4885]: I1002 01:48:26.288094 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:26 crc kubenswrapper[4885]: I1002 01:48:26.288120 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:26 crc kubenswrapper[4885]: I1002 01:48:26.288151 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:26 crc kubenswrapper[4885]: I1002 01:48:26.288177 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:26Z","lastTransitionTime":"2025-10-02T01:48:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:26 crc kubenswrapper[4885]: I1002 01:48:26.390701 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:26 crc kubenswrapper[4885]: I1002 01:48:26.390786 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:26 crc kubenswrapper[4885]: I1002 01:48:26.390814 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:26 crc kubenswrapper[4885]: I1002 01:48:26.390846 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:26 crc kubenswrapper[4885]: I1002 01:48:26.390874 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:26Z","lastTransitionTime":"2025-10-02T01:48:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:26 crc kubenswrapper[4885]: I1002 01:48:26.493726 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:26 crc kubenswrapper[4885]: I1002 01:48:26.493770 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:26 crc kubenswrapper[4885]: I1002 01:48:26.493782 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:26 crc kubenswrapper[4885]: I1002 01:48:26.493798 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:26 crc kubenswrapper[4885]: I1002 01:48:26.493811 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:26Z","lastTransitionTime":"2025-10-02T01:48:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:26 crc kubenswrapper[4885]: I1002 01:48:26.596906 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:26 crc kubenswrapper[4885]: I1002 01:48:26.597065 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:26 crc kubenswrapper[4885]: I1002 01:48:26.597088 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:26 crc kubenswrapper[4885]: I1002 01:48:26.597154 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:26 crc kubenswrapper[4885]: I1002 01:48:26.597174 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:26Z","lastTransitionTime":"2025-10-02T01:48:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:26 crc kubenswrapper[4885]: I1002 01:48:26.699755 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:26 crc kubenswrapper[4885]: I1002 01:48:26.699785 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:26 crc kubenswrapper[4885]: I1002 01:48:26.699797 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:26 crc kubenswrapper[4885]: I1002 01:48:26.699812 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:26 crc kubenswrapper[4885]: I1002 01:48:26.699824 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:26Z","lastTransitionTime":"2025-10-02T01:48:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:26 crc kubenswrapper[4885]: I1002 01:48:26.802602 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:26 crc kubenswrapper[4885]: I1002 01:48:26.802669 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:26 crc kubenswrapper[4885]: I1002 01:48:26.802691 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:26 crc kubenswrapper[4885]: I1002 01:48:26.802719 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:26 crc kubenswrapper[4885]: I1002 01:48:26.802755 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:26Z","lastTransitionTime":"2025-10-02T01:48:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:26 crc kubenswrapper[4885]: I1002 01:48:26.905855 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:26 crc kubenswrapper[4885]: I1002 01:48:26.905906 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:26 crc kubenswrapper[4885]: I1002 01:48:26.905926 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:26 crc kubenswrapper[4885]: I1002 01:48:26.905949 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:26 crc kubenswrapper[4885]: I1002 01:48:26.905968 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:26Z","lastTransitionTime":"2025-10-02T01:48:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:27 crc kubenswrapper[4885]: I1002 01:48:27.008506 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:27 crc kubenswrapper[4885]: I1002 01:48:27.008565 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:27 crc kubenswrapper[4885]: I1002 01:48:27.008581 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:27 crc kubenswrapper[4885]: I1002 01:48:27.008607 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:27 crc kubenswrapper[4885]: I1002 01:48:27.008625 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:27Z","lastTransitionTime":"2025-10-02T01:48:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:27 crc kubenswrapper[4885]: I1002 01:48:27.111554 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:27 crc kubenswrapper[4885]: I1002 01:48:27.111622 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:27 crc kubenswrapper[4885]: I1002 01:48:27.111651 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:27 crc kubenswrapper[4885]: I1002 01:48:27.111681 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:27 crc kubenswrapper[4885]: I1002 01:48:27.111707 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:27Z","lastTransitionTime":"2025-10-02T01:48:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:27 crc kubenswrapper[4885]: I1002 01:48:27.215438 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:27 crc kubenswrapper[4885]: I1002 01:48:27.215505 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:27 crc kubenswrapper[4885]: I1002 01:48:27.215527 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:27 crc kubenswrapper[4885]: I1002 01:48:27.215557 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:27 crc kubenswrapper[4885]: I1002 01:48:27.215581 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:27Z","lastTransitionTime":"2025-10-02T01:48:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:27 crc kubenswrapper[4885]: I1002 01:48:27.320347 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:27 crc kubenswrapper[4885]: I1002 01:48:27.320417 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:27 crc kubenswrapper[4885]: I1002 01:48:27.320440 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:27 crc kubenswrapper[4885]: I1002 01:48:27.320471 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:27 crc kubenswrapper[4885]: I1002 01:48:27.320494 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:27Z","lastTransitionTime":"2025-10-02T01:48:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:27 crc kubenswrapper[4885]: I1002 01:48:27.423699 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:27 crc kubenswrapper[4885]: I1002 01:48:27.423769 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:27 crc kubenswrapper[4885]: I1002 01:48:27.423786 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:27 crc kubenswrapper[4885]: I1002 01:48:27.423811 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:27 crc kubenswrapper[4885]: I1002 01:48:27.423829 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:27Z","lastTransitionTime":"2025-10-02T01:48:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:27 crc kubenswrapper[4885]: I1002 01:48:27.526969 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:27 crc kubenswrapper[4885]: I1002 01:48:27.527032 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:27 crc kubenswrapper[4885]: I1002 01:48:27.527049 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:27 crc kubenswrapper[4885]: I1002 01:48:27.527077 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:27 crc kubenswrapper[4885]: I1002 01:48:27.527094 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:27Z","lastTransitionTime":"2025-10-02T01:48:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:27 crc kubenswrapper[4885]: I1002 01:48:27.630469 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:27 crc kubenswrapper[4885]: I1002 01:48:27.630527 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:27 crc kubenswrapper[4885]: I1002 01:48:27.630544 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:27 crc kubenswrapper[4885]: I1002 01:48:27.630567 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:27 crc kubenswrapper[4885]: I1002 01:48:27.630586 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:27Z","lastTransitionTime":"2025-10-02T01:48:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:27 crc kubenswrapper[4885]: I1002 01:48:27.732771 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:27 crc kubenswrapper[4885]: I1002 01:48:27.732890 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:27 crc kubenswrapper[4885]: I1002 01:48:27.732909 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:27 crc kubenswrapper[4885]: I1002 01:48:27.732932 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:27 crc kubenswrapper[4885]: I1002 01:48:27.732953 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:27Z","lastTransitionTime":"2025-10-02T01:48:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:27 crc kubenswrapper[4885]: I1002 01:48:27.835707 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:27 crc kubenswrapper[4885]: I1002 01:48:27.835767 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:27 crc kubenswrapper[4885]: I1002 01:48:27.835784 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:27 crc kubenswrapper[4885]: I1002 01:48:27.835808 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:27 crc kubenswrapper[4885]: I1002 01:48:27.835826 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:27Z","lastTransitionTime":"2025-10-02T01:48:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:27 crc kubenswrapper[4885]: I1002 01:48:27.939341 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:27 crc kubenswrapper[4885]: I1002 01:48:27.939395 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:27 crc kubenswrapper[4885]: I1002 01:48:27.939413 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:27 crc kubenswrapper[4885]: I1002 01:48:27.939437 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:27 crc kubenswrapper[4885]: I1002 01:48:27.939454 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:27Z","lastTransitionTime":"2025-10-02T01:48:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:28 crc kubenswrapper[4885]: I1002 01:48:28.042195 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:28 crc kubenswrapper[4885]: I1002 01:48:28.042224 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:28 crc kubenswrapper[4885]: I1002 01:48:28.042234 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:28 crc kubenswrapper[4885]: I1002 01:48:28.042248 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:28 crc kubenswrapper[4885]: I1002 01:48:28.042277 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:28Z","lastTransitionTime":"2025-10-02T01:48:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:28 crc kubenswrapper[4885]: I1002 01:48:28.045892 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sv4md" Oct 02 01:48:28 crc kubenswrapper[4885]: I1002 01:48:28.045936 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:48:28 crc kubenswrapper[4885]: I1002 01:48:28.046166 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:48:28 crc kubenswrapper[4885]: I1002 01:48:28.046210 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:48:28 crc kubenswrapper[4885]: E1002 01:48:28.046400 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sv4md" podUID="724a3e49-001c-45a4-a896-82c13e0ee0e7" Oct 02 01:48:28 crc kubenswrapper[4885]: E1002 01:48:28.046555 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:48:28 crc kubenswrapper[4885]: E1002 01:48:28.046664 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:48:28 crc kubenswrapper[4885]: E1002 01:48:28.046736 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:48:28 crc kubenswrapper[4885]: I1002 01:48:28.145384 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:28 crc kubenswrapper[4885]: I1002 01:48:28.145442 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:28 crc kubenswrapper[4885]: I1002 01:48:28.145457 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:28 crc kubenswrapper[4885]: I1002 01:48:28.145482 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:28 crc kubenswrapper[4885]: I1002 01:48:28.145502 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:28Z","lastTransitionTime":"2025-10-02T01:48:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:28 crc kubenswrapper[4885]: I1002 01:48:28.247867 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:28 crc kubenswrapper[4885]: I1002 01:48:28.247907 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:28 crc kubenswrapper[4885]: I1002 01:48:28.247920 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:28 crc kubenswrapper[4885]: I1002 01:48:28.247938 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:28 crc kubenswrapper[4885]: I1002 01:48:28.247950 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:28Z","lastTransitionTime":"2025-10-02T01:48:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:28 crc kubenswrapper[4885]: I1002 01:48:28.350982 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:28 crc kubenswrapper[4885]: I1002 01:48:28.351056 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:28 crc kubenswrapper[4885]: I1002 01:48:28.351072 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:28 crc kubenswrapper[4885]: I1002 01:48:28.351099 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:28 crc kubenswrapper[4885]: I1002 01:48:28.351117 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:28Z","lastTransitionTime":"2025-10-02T01:48:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:28 crc kubenswrapper[4885]: I1002 01:48:28.454102 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:28 crc kubenswrapper[4885]: I1002 01:48:28.454164 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:28 crc kubenswrapper[4885]: I1002 01:48:28.454184 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:28 crc kubenswrapper[4885]: I1002 01:48:28.454208 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:28 crc kubenswrapper[4885]: I1002 01:48:28.454226 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:28Z","lastTransitionTime":"2025-10-02T01:48:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:28 crc kubenswrapper[4885]: I1002 01:48:28.557791 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:28 crc kubenswrapper[4885]: I1002 01:48:28.557843 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:28 crc kubenswrapper[4885]: I1002 01:48:28.557861 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:28 crc kubenswrapper[4885]: I1002 01:48:28.557883 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:28 crc kubenswrapper[4885]: I1002 01:48:28.557900 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:28Z","lastTransitionTime":"2025-10-02T01:48:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:28 crc kubenswrapper[4885]: I1002 01:48:28.660454 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:28 crc kubenswrapper[4885]: I1002 01:48:28.660515 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:28 crc kubenswrapper[4885]: I1002 01:48:28.660534 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:28 crc kubenswrapper[4885]: I1002 01:48:28.660557 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:28 crc kubenswrapper[4885]: I1002 01:48:28.660574 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:28Z","lastTransitionTime":"2025-10-02T01:48:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:28 crc kubenswrapper[4885]: I1002 01:48:28.763658 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:28 crc kubenswrapper[4885]: I1002 01:48:28.763721 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:28 crc kubenswrapper[4885]: I1002 01:48:28.763740 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:28 crc kubenswrapper[4885]: I1002 01:48:28.763768 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:28 crc kubenswrapper[4885]: I1002 01:48:28.763787 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:28Z","lastTransitionTime":"2025-10-02T01:48:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:28 crc kubenswrapper[4885]: I1002 01:48:28.867205 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:28 crc kubenswrapper[4885]: I1002 01:48:28.867330 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:28 crc kubenswrapper[4885]: I1002 01:48:28.867354 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:28 crc kubenswrapper[4885]: I1002 01:48:28.867397 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:28 crc kubenswrapper[4885]: I1002 01:48:28.867419 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:28Z","lastTransitionTime":"2025-10-02T01:48:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:28 crc kubenswrapper[4885]: I1002 01:48:28.970685 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:28 crc kubenswrapper[4885]: I1002 01:48:28.970796 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:28 crc kubenswrapper[4885]: I1002 01:48:28.970819 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:28 crc kubenswrapper[4885]: I1002 01:48:28.970845 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:28 crc kubenswrapper[4885]: I1002 01:48:28.970863 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:28Z","lastTransitionTime":"2025-10-02T01:48:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:29 crc kubenswrapper[4885]: I1002 01:48:29.074370 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:29 crc kubenswrapper[4885]: I1002 01:48:29.074502 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:29 crc kubenswrapper[4885]: I1002 01:48:29.074519 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:29 crc kubenswrapper[4885]: I1002 01:48:29.074543 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:29 crc kubenswrapper[4885]: I1002 01:48:29.074561 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:29Z","lastTransitionTime":"2025-10-02T01:48:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:29 crc kubenswrapper[4885]: I1002 01:48:29.177312 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:29 crc kubenswrapper[4885]: I1002 01:48:29.177384 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:29 crc kubenswrapper[4885]: I1002 01:48:29.177410 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:29 crc kubenswrapper[4885]: I1002 01:48:29.177440 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:29 crc kubenswrapper[4885]: I1002 01:48:29.177462 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:29Z","lastTransitionTime":"2025-10-02T01:48:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:29 crc kubenswrapper[4885]: I1002 01:48:29.280637 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:29 crc kubenswrapper[4885]: I1002 01:48:29.280677 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:29 crc kubenswrapper[4885]: I1002 01:48:29.280685 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:29 crc kubenswrapper[4885]: I1002 01:48:29.280697 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:29 crc kubenswrapper[4885]: I1002 01:48:29.280706 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:29Z","lastTransitionTime":"2025-10-02T01:48:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:29 crc kubenswrapper[4885]: I1002 01:48:29.383865 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:29 crc kubenswrapper[4885]: I1002 01:48:29.383938 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:29 crc kubenswrapper[4885]: I1002 01:48:29.383958 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:29 crc kubenswrapper[4885]: I1002 01:48:29.383988 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:29 crc kubenswrapper[4885]: I1002 01:48:29.384006 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:29Z","lastTransitionTime":"2025-10-02T01:48:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:29 crc kubenswrapper[4885]: I1002 01:48:29.486975 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:29 crc kubenswrapper[4885]: I1002 01:48:29.487034 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:29 crc kubenswrapper[4885]: I1002 01:48:29.487050 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:29 crc kubenswrapper[4885]: I1002 01:48:29.487074 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:29 crc kubenswrapper[4885]: I1002 01:48:29.487092 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:29Z","lastTransitionTime":"2025-10-02T01:48:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:29 crc kubenswrapper[4885]: I1002 01:48:29.590103 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:29 crc kubenswrapper[4885]: I1002 01:48:29.590180 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:29 crc kubenswrapper[4885]: I1002 01:48:29.590192 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:29 crc kubenswrapper[4885]: I1002 01:48:29.590205 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:29 crc kubenswrapper[4885]: I1002 01:48:29.590215 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:29Z","lastTransitionTime":"2025-10-02T01:48:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:29 crc kubenswrapper[4885]: I1002 01:48:29.693649 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:29 crc kubenswrapper[4885]: I1002 01:48:29.693705 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:29 crc kubenswrapper[4885]: I1002 01:48:29.693721 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:29 crc kubenswrapper[4885]: I1002 01:48:29.693744 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:29 crc kubenswrapper[4885]: I1002 01:48:29.693762 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:29Z","lastTransitionTime":"2025-10-02T01:48:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:29 crc kubenswrapper[4885]: I1002 01:48:29.797141 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:29 crc kubenswrapper[4885]: I1002 01:48:29.797211 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:29 crc kubenswrapper[4885]: I1002 01:48:29.797232 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:29 crc kubenswrapper[4885]: I1002 01:48:29.797290 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:29 crc kubenswrapper[4885]: I1002 01:48:29.797309 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:29Z","lastTransitionTime":"2025-10-02T01:48:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:29 crc kubenswrapper[4885]: I1002 01:48:29.907481 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:29 crc kubenswrapper[4885]: I1002 01:48:29.907546 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:29 crc kubenswrapper[4885]: I1002 01:48:29.907566 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:29 crc kubenswrapper[4885]: I1002 01:48:29.907592 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:29 crc kubenswrapper[4885]: I1002 01:48:29.907615 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:29Z","lastTransitionTime":"2025-10-02T01:48:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:30 crc kubenswrapper[4885]: I1002 01:48:30.010949 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:30 crc kubenswrapper[4885]: I1002 01:48:30.011017 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:30 crc kubenswrapper[4885]: I1002 01:48:30.011034 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:30 crc kubenswrapper[4885]: I1002 01:48:30.011059 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:30 crc kubenswrapper[4885]: I1002 01:48:30.011076 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:30Z","lastTransitionTime":"2025-10-02T01:48:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:30 crc kubenswrapper[4885]: I1002 01:48:30.046205 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sv4md" Oct 02 01:48:30 crc kubenswrapper[4885]: I1002 01:48:30.046225 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:48:30 crc kubenswrapper[4885]: I1002 01:48:30.046240 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:48:30 crc kubenswrapper[4885]: I1002 01:48:30.046205 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:48:30 crc kubenswrapper[4885]: E1002 01:48:30.046439 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:48:30 crc kubenswrapper[4885]: E1002 01:48:30.046833 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:48:30 crc kubenswrapper[4885]: E1002 01:48:30.047005 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sv4md" podUID="724a3e49-001c-45a4-a896-82c13e0ee0e7" Oct 02 01:48:30 crc kubenswrapper[4885]: E1002 01:48:30.047117 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:48:30 crc kubenswrapper[4885]: I1002 01:48:30.113331 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:30 crc kubenswrapper[4885]: I1002 01:48:30.113385 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:30 crc kubenswrapper[4885]: I1002 01:48:30.113403 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:30 crc kubenswrapper[4885]: I1002 01:48:30.113426 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:30 crc kubenswrapper[4885]: I1002 01:48:30.113446 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:30Z","lastTransitionTime":"2025-10-02T01:48:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:30 crc kubenswrapper[4885]: I1002 01:48:30.216631 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:30 crc kubenswrapper[4885]: I1002 01:48:30.216692 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:30 crc kubenswrapper[4885]: I1002 01:48:30.216710 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:30 crc kubenswrapper[4885]: I1002 01:48:30.216738 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:30 crc kubenswrapper[4885]: I1002 01:48:30.216757 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:30Z","lastTransitionTime":"2025-10-02T01:48:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:30 crc kubenswrapper[4885]: I1002 01:48:30.319855 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:30 crc kubenswrapper[4885]: I1002 01:48:30.319905 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:30 crc kubenswrapper[4885]: I1002 01:48:30.319922 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:30 crc kubenswrapper[4885]: I1002 01:48:30.319947 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:30 crc kubenswrapper[4885]: I1002 01:48:30.319965 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:30Z","lastTransitionTime":"2025-10-02T01:48:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:30 crc kubenswrapper[4885]: I1002 01:48:30.422349 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:30 crc kubenswrapper[4885]: I1002 01:48:30.422415 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:30 crc kubenswrapper[4885]: I1002 01:48:30.422436 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:30 crc kubenswrapper[4885]: I1002 01:48:30.422461 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:30 crc kubenswrapper[4885]: I1002 01:48:30.422479 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:30Z","lastTransitionTime":"2025-10-02T01:48:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:30 crc kubenswrapper[4885]: I1002 01:48:30.505148 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:30 crc kubenswrapper[4885]: I1002 01:48:30.505208 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:30 crc kubenswrapper[4885]: I1002 01:48:30.505224 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:30 crc kubenswrapper[4885]: I1002 01:48:30.505248 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:30 crc kubenswrapper[4885]: I1002 01:48:30.505298 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:30Z","lastTransitionTime":"2025-10-02T01:48:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:30 crc kubenswrapper[4885]: I1002 01:48:30.532549 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:48:30 crc kubenswrapper[4885]: I1002 01:48:30.532577 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:48:30 crc kubenswrapper[4885]: I1002 01:48:30.532589 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:48:30 crc kubenswrapper[4885]: I1002 01:48:30.532602 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:48:30 crc kubenswrapper[4885]: I1002 01:48:30.532612 4885 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:48:30Z","lastTransitionTime":"2025-10-02T01:48:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:48:30 crc kubenswrapper[4885]: I1002 01:48:30.571217 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-hpcjj"] Oct 02 01:48:30 crc kubenswrapper[4885]: I1002 01:48:30.571837 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hpcjj" Oct 02 01:48:30 crc kubenswrapper[4885]: I1002 01:48:30.575641 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Oct 02 01:48:30 crc kubenswrapper[4885]: I1002 01:48:30.576192 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Oct 02 01:48:30 crc kubenswrapper[4885]: I1002 01:48:30.576226 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Oct 02 01:48:30 crc kubenswrapper[4885]: I1002 01:48:30.578494 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Oct 02 01:48:30 crc kubenswrapper[4885]: I1002 01:48:30.596932 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podStartSLOduration=77.59690595000001 podStartE2EDuration="1m17.59690595s" podCreationTimestamp="2025-10-02 01:47:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:48:30.596709794 +0000 UTC m=+99.408457223" watchObservedRunningTime="2025-10-02 01:48:30.59690595 +0000 UTC m=+99.408653379" Oct 02 01:48:30 crc kubenswrapper[4885]: I1002 01:48:30.617314 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/86662652-3afb-4311-88ed-73546397284b-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-hpcjj\" (UID: \"86662652-3afb-4311-88ed-73546397284b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hpcjj" Oct 02 01:48:30 crc kubenswrapper[4885]: I1002 01:48:30.617405 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/86662652-3afb-4311-88ed-73546397284b-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-hpcjj\" (UID: \"86662652-3afb-4311-88ed-73546397284b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hpcjj" Oct 02 01:48:30 crc kubenswrapper[4885]: I1002 01:48:30.617502 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/86662652-3afb-4311-88ed-73546397284b-service-ca\") pod \"cluster-version-operator-5c965bbfc6-hpcjj\" (UID: \"86662652-3afb-4311-88ed-73546397284b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hpcjj" Oct 02 01:48:30 crc kubenswrapper[4885]: I1002 01:48:30.617563 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/86662652-3afb-4311-88ed-73546397284b-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-hpcjj\" (UID: \"86662652-3afb-4311-88ed-73546397284b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hpcjj" Oct 02 01:48:30 crc kubenswrapper[4885]: I1002 01:48:30.617596 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/86662652-3afb-4311-88ed-73546397284b-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-hpcjj\" (UID: \"86662652-3afb-4311-88ed-73546397284b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hpcjj" Oct 02 01:48:30 crc kubenswrapper[4885]: I1002 01:48:30.640354 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-jgdvt" podStartSLOduration=77.640333153 podStartE2EDuration="1m17.640333153s" podCreationTimestamp="2025-10-02 01:47:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:48:30.614868009 +0000 UTC m=+99.426615438" watchObservedRunningTime="2025-10-02 01:48:30.640333153 +0000 UTC m=+99.452080592" Oct 02 01:48:30 crc kubenswrapper[4885]: I1002 01:48:30.640711 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=79.640656753 podStartE2EDuration="1m19.640656753s" podCreationTimestamp="2025-10-02 01:47:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:48:30.639528919 +0000 UTC m=+99.451276358" watchObservedRunningTime="2025-10-02 01:48:30.640656753 +0000 UTC m=+99.452404192" Oct 02 01:48:30 crc kubenswrapper[4885]: I1002 01:48:30.677914 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=13.67788856 podStartE2EDuration="13.67788856s" podCreationTimestamp="2025-10-02 01:48:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:48:30.656169189 +0000 UTC m=+99.467916628" watchObservedRunningTime="2025-10-02 01:48:30.67788856 +0000 UTC m=+99.489635989" Oct 02 01:48:30 crc kubenswrapper[4885]: I1002 01:48:30.718245 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/86662652-3afb-4311-88ed-73546397284b-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-hpcjj\" (UID: \"86662652-3afb-4311-88ed-73546397284b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hpcjj" Oct 02 01:48:30 crc kubenswrapper[4885]: I1002 01:48:30.718418 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/86662652-3afb-4311-88ed-73546397284b-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-hpcjj\" (UID: \"86662652-3afb-4311-88ed-73546397284b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hpcjj" Oct 02 01:48:30 crc kubenswrapper[4885]: I1002 01:48:30.718429 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/86662652-3afb-4311-88ed-73546397284b-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-hpcjj\" (UID: \"86662652-3afb-4311-88ed-73546397284b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hpcjj" Oct 02 01:48:30 crc kubenswrapper[4885]: I1002 01:48:30.718540 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/86662652-3afb-4311-88ed-73546397284b-service-ca\") pod \"cluster-version-operator-5c965bbfc6-hpcjj\" (UID: \"86662652-3afb-4311-88ed-73546397284b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hpcjj" Oct 02 01:48:30 crc kubenswrapper[4885]: I1002 01:48:30.718600 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/86662652-3afb-4311-88ed-73546397284b-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-hpcjj\" (UID: \"86662652-3afb-4311-88ed-73546397284b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hpcjj" Oct 02 01:48:30 crc kubenswrapper[4885]: I1002 01:48:30.718665 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/86662652-3afb-4311-88ed-73546397284b-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-hpcjj\" (UID: \"86662652-3afb-4311-88ed-73546397284b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hpcjj" Oct 02 01:48:30 crc kubenswrapper[4885]: I1002 01:48:30.718770 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/86662652-3afb-4311-88ed-73546397284b-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-hpcjj\" (UID: \"86662652-3afb-4311-88ed-73546397284b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hpcjj" Oct 02 01:48:30 crc kubenswrapper[4885]: I1002 01:48:30.719685 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/86662652-3afb-4311-88ed-73546397284b-service-ca\") pod \"cluster-version-operator-5c965bbfc6-hpcjj\" (UID: \"86662652-3afb-4311-88ed-73546397284b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hpcjj" Oct 02 01:48:30 crc kubenswrapper[4885]: I1002 01:48:30.735567 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/86662652-3afb-4311-88ed-73546397284b-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-hpcjj\" (UID: \"86662652-3afb-4311-88ed-73546397284b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hpcjj" Oct 02 01:48:30 crc kubenswrapper[4885]: I1002 01:48:30.747897 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-58lhj" podStartSLOduration=77.74787969 podStartE2EDuration="1m17.74787969s" podCreationTimestamp="2025-10-02 01:47:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:48:30.747332624 +0000 UTC m=+99.559080063" watchObservedRunningTime="2025-10-02 01:48:30.74787969 +0000 UTC m=+99.559627129" Oct 02 01:48:30 crc kubenswrapper[4885]: I1002 01:48:30.755385 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/86662652-3afb-4311-88ed-73546397284b-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-hpcjj\" (UID: \"86662652-3afb-4311-88ed-73546397284b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hpcjj" Oct 02 01:48:30 crc kubenswrapper[4885]: I1002 01:48:30.769414 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=78.769389726 podStartE2EDuration="1m18.769389726s" podCreationTimestamp="2025-10-02 01:47:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:48:30.769105218 +0000 UTC m=+99.580852657" watchObservedRunningTime="2025-10-02 01:48:30.769389726 +0000 UTC m=+99.581137165" Oct 02 01:48:30 crc kubenswrapper[4885]: I1002 01:48:30.894330 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hpcjj" Oct 02 01:48:30 crc kubenswrapper[4885]: I1002 01:48:30.917103 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=76.917084229 podStartE2EDuration="1m16.917084229s" podCreationTimestamp="2025-10-02 01:47:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:48:30.911328205 +0000 UTC m=+99.723075634" watchObservedRunningTime="2025-10-02 01:48:30.917084229 +0000 UTC m=+99.728831638" Oct 02 01:48:30 crc kubenswrapper[4885]: I1002 01:48:30.917719 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=47.917710817 podStartE2EDuration="47.917710817s" podCreationTimestamp="2025-10-02 01:47:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:48:30.883203371 +0000 UTC m=+99.694950800" watchObservedRunningTime="2025-10-02 01:48:30.917710817 +0000 UTC m=+99.729458226" Oct 02 01:48:30 crc kubenswrapper[4885]: I1002 01:48:30.953987 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-vhgbd" podStartSLOduration=77.953969645 podStartE2EDuration="1m17.953969645s" podCreationTimestamp="2025-10-02 01:47:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:48:30.953947374 +0000 UTC m=+99.765694813" watchObservedRunningTime="2025-10-02 01:48:30.953969645 +0000 UTC m=+99.765717054" Oct 02 01:48:30 crc kubenswrapper[4885]: I1002 01:48:30.988478 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rrn8b" podStartSLOduration=76.98844887999999 podStartE2EDuration="1m16.98844888s" podCreationTimestamp="2025-10-02 01:47:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:48:30.96913017 +0000 UTC m=+99.780877579" watchObservedRunningTime="2025-10-02 01:48:30.98844888 +0000 UTC m=+99.800196309" Oct 02 01:48:31 crc kubenswrapper[4885]: I1002 01:48:31.033327 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-2ll9b" podStartSLOduration=78.033311896 podStartE2EDuration="1m18.033311896s" podCreationTimestamp="2025-10-02 01:47:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:48:31.029702728 +0000 UTC m=+99.841450127" watchObservedRunningTime="2025-10-02 01:48:31.033311896 +0000 UTC m=+99.845059295" Oct 02 01:48:31 crc kubenswrapper[4885]: I1002 01:48:31.719303 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hpcjj" event={"ID":"86662652-3afb-4311-88ed-73546397284b","Type":"ContainerStarted","Data":"723f038d46838b23eb879c825df6c0538edc8429efeb5ac6f9757c2dfb0d40f5"} Oct 02 01:48:31 crc kubenswrapper[4885]: I1002 01:48:31.719598 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hpcjj" event={"ID":"86662652-3afb-4311-88ed-73546397284b","Type":"ContainerStarted","Data":"b3853244c7c05e4eef0f9f36916c23bcf9749cb18684309348e8c2d7c6199c87"} Oct 02 01:48:31 crc kubenswrapper[4885]: I1002 01:48:31.738466 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hpcjj" podStartSLOduration=78.738454347 podStartE2EDuration="1m18.738454347s" podCreationTimestamp="2025-10-02 01:47:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:48:31.738177679 +0000 UTC m=+100.549925118" watchObservedRunningTime="2025-10-02 01:48:31.738454347 +0000 UTC m=+100.550201756" Oct 02 01:48:32 crc kubenswrapper[4885]: I1002 01:48:32.045837 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:48:32 crc kubenswrapper[4885]: I1002 01:48:32.045926 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sv4md" Oct 02 01:48:32 crc kubenswrapper[4885]: I1002 01:48:32.046001 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:48:32 crc kubenswrapper[4885]: E1002 01:48:32.047598 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:48:32 crc kubenswrapper[4885]: I1002 01:48:32.047657 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:48:32 crc kubenswrapper[4885]: E1002 01:48:32.048001 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:48:32 crc kubenswrapper[4885]: E1002 01:48:32.048183 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sv4md" podUID="724a3e49-001c-45a4-a896-82c13e0ee0e7" Oct 02 01:48:32 crc kubenswrapper[4885]: E1002 01:48:32.048466 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:48:32 crc kubenswrapper[4885]: I1002 01:48:32.237193 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/724a3e49-001c-45a4-a896-82c13e0ee0e7-metrics-certs\") pod \"network-metrics-daemon-sv4md\" (UID: \"724a3e49-001c-45a4-a896-82c13e0ee0e7\") " pod="openshift-multus/network-metrics-daemon-sv4md" Oct 02 01:48:32 crc kubenswrapper[4885]: E1002 01:48:32.237468 4885 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 01:48:32 crc kubenswrapper[4885]: E1002 01:48:32.237592 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/724a3e49-001c-45a4-a896-82c13e0ee0e7-metrics-certs podName:724a3e49-001c-45a4-a896-82c13e0ee0e7 nodeName:}" failed. No retries permitted until 2025-10-02 01:49:36.237558824 +0000 UTC m=+165.049306263 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/724a3e49-001c-45a4-a896-82c13e0ee0e7-metrics-certs") pod "network-metrics-daemon-sv4md" (UID: "724a3e49-001c-45a4-a896-82c13e0ee0e7") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 01:48:34 crc kubenswrapper[4885]: I1002 01:48:34.046565 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:48:34 crc kubenswrapper[4885]: I1002 01:48:34.046602 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:48:34 crc kubenswrapper[4885]: I1002 01:48:34.047495 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sv4md" Oct 02 01:48:34 crc kubenswrapper[4885]: E1002 01:48:34.047677 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:48:34 crc kubenswrapper[4885]: I1002 01:48:34.047782 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:48:34 crc kubenswrapper[4885]: E1002 01:48:34.047813 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sv4md" podUID="724a3e49-001c-45a4-a896-82c13e0ee0e7" Oct 02 01:48:34 crc kubenswrapper[4885]: E1002 01:48:34.047975 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:48:34 crc kubenswrapper[4885]: E1002 01:48:34.048251 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:48:36 crc kubenswrapper[4885]: I1002 01:48:36.046562 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:48:36 crc kubenswrapper[4885]: I1002 01:48:36.046562 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sv4md" Oct 02 01:48:36 crc kubenswrapper[4885]: I1002 01:48:36.046608 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:48:36 crc kubenswrapper[4885]: I1002 01:48:36.046789 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:48:36 crc kubenswrapper[4885]: E1002 01:48:36.046958 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:48:36 crc kubenswrapper[4885]: E1002 01:48:36.047373 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sv4md" podUID="724a3e49-001c-45a4-a896-82c13e0ee0e7" Oct 02 01:48:36 crc kubenswrapper[4885]: I1002 01:48:36.048529 4885 scope.go:117] "RemoveContainer" containerID="2ef80e838accc59ad7a579c66f4fcd6881d23fa86fc5922d4b0085d211d5d94b" Oct 02 01:48:36 crc kubenswrapper[4885]: E1002 01:48:36.048758 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-dngcm_openshift-ovn-kubernetes(9e4679dd-f870-41e9-a32b-360f5d2ee81b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" podUID="9e4679dd-f870-41e9-a32b-360f5d2ee81b" Oct 02 01:48:36 crc kubenswrapper[4885]: E1002 01:48:36.049034 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:48:36 crc kubenswrapper[4885]: E1002 01:48:36.049287 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:48:38 crc kubenswrapper[4885]: I1002 01:48:38.046552 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:48:38 crc kubenswrapper[4885]: I1002 01:48:38.046595 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sv4md" Oct 02 01:48:38 crc kubenswrapper[4885]: E1002 01:48:38.046793 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sv4md" podUID="724a3e49-001c-45a4-a896-82c13e0ee0e7" Oct 02 01:48:38 crc kubenswrapper[4885]: I1002 01:48:38.046881 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:48:38 crc kubenswrapper[4885]: E1002 01:48:38.047088 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:48:38 crc kubenswrapper[4885]: I1002 01:48:38.047462 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:48:38 crc kubenswrapper[4885]: E1002 01:48:38.047598 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:48:38 crc kubenswrapper[4885]: E1002 01:48:38.047957 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:48:40 crc kubenswrapper[4885]: I1002 01:48:40.045728 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:48:40 crc kubenswrapper[4885]: I1002 01:48:40.045926 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sv4md" Oct 02 01:48:40 crc kubenswrapper[4885]: I1002 01:48:40.045958 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:48:40 crc kubenswrapper[4885]: I1002 01:48:40.046072 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:48:40 crc kubenswrapper[4885]: E1002 01:48:40.046213 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:48:40 crc kubenswrapper[4885]: E1002 01:48:40.046442 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:48:40 crc kubenswrapper[4885]: E1002 01:48:40.046629 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sv4md" podUID="724a3e49-001c-45a4-a896-82c13e0ee0e7" Oct 02 01:48:40 crc kubenswrapper[4885]: E1002 01:48:40.046716 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:48:42 crc kubenswrapper[4885]: I1002 01:48:42.045592 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:48:42 crc kubenswrapper[4885]: E1002 01:48:42.047348 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:48:42 crc kubenswrapper[4885]: I1002 01:48:42.047417 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:48:42 crc kubenswrapper[4885]: I1002 01:48:42.047462 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sv4md" Oct 02 01:48:42 crc kubenswrapper[4885]: I1002 01:48:42.047536 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:48:42 crc kubenswrapper[4885]: E1002 01:48:42.047726 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:48:42 crc kubenswrapper[4885]: E1002 01:48:42.047857 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sv4md" podUID="724a3e49-001c-45a4-a896-82c13e0ee0e7" Oct 02 01:48:42 crc kubenswrapper[4885]: E1002 01:48:42.048450 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:48:44 crc kubenswrapper[4885]: I1002 01:48:44.045749 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:48:44 crc kubenswrapper[4885]: I1002 01:48:44.045869 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:48:44 crc kubenswrapper[4885]: E1002 01:48:44.045955 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:48:44 crc kubenswrapper[4885]: E1002 01:48:44.046145 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:48:44 crc kubenswrapper[4885]: I1002 01:48:44.046344 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sv4md" Oct 02 01:48:44 crc kubenswrapper[4885]: E1002 01:48:44.046530 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sv4md" podUID="724a3e49-001c-45a4-a896-82c13e0ee0e7" Oct 02 01:48:44 crc kubenswrapper[4885]: I1002 01:48:44.046592 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:48:44 crc kubenswrapper[4885]: E1002 01:48:44.046780 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:48:46 crc kubenswrapper[4885]: I1002 01:48:46.046583 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:48:46 crc kubenswrapper[4885]: E1002 01:48:46.046757 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:48:46 crc kubenswrapper[4885]: I1002 01:48:46.047072 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:48:46 crc kubenswrapper[4885]: E1002 01:48:46.047158 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:48:46 crc kubenswrapper[4885]: I1002 01:48:46.047404 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:48:46 crc kubenswrapper[4885]: E1002 01:48:46.047487 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:48:46 crc kubenswrapper[4885]: I1002 01:48:46.047693 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sv4md" Oct 02 01:48:46 crc kubenswrapper[4885]: E1002 01:48:46.048078 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sv4md" podUID="724a3e49-001c-45a4-a896-82c13e0ee0e7" Oct 02 01:48:47 crc kubenswrapper[4885]: I1002 01:48:47.774549 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-vhgbd_ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f/kube-multus/1.log" Oct 02 01:48:47 crc kubenswrapper[4885]: I1002 01:48:47.775174 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-vhgbd_ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f/kube-multus/0.log" Oct 02 01:48:47 crc kubenswrapper[4885]: I1002 01:48:47.775221 4885 generic.go:334] "Generic (PLEG): container finished" podID="ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f" containerID="df9792d80c807ed98c8442822d01f619327361c40425abfe26f11e5c27ba2584" exitCode=1 Oct 02 01:48:47 crc kubenswrapper[4885]: I1002 01:48:47.775295 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-vhgbd" event={"ID":"ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f","Type":"ContainerDied","Data":"df9792d80c807ed98c8442822d01f619327361c40425abfe26f11e5c27ba2584"} Oct 02 01:48:47 crc kubenswrapper[4885]: I1002 01:48:47.775340 4885 scope.go:117] "RemoveContainer" containerID="c63a7d20d39b65acfb5a9be4c28b21c72111b398bea76c238e15ddb4dd63e139" Oct 02 01:48:47 crc kubenswrapper[4885]: I1002 01:48:47.775935 4885 scope.go:117] "RemoveContainer" containerID="df9792d80c807ed98c8442822d01f619327361c40425abfe26f11e5c27ba2584" Oct 02 01:48:47 crc kubenswrapper[4885]: E1002 01:48:47.776496 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-vhgbd_openshift-multus(ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f)\"" pod="openshift-multus/multus-vhgbd" podUID="ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f" Oct 02 01:48:48 crc kubenswrapper[4885]: I1002 01:48:48.046535 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:48:48 crc kubenswrapper[4885]: E1002 01:48:48.046732 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:48:48 crc kubenswrapper[4885]: I1002 01:48:48.047092 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:48:48 crc kubenswrapper[4885]: E1002 01:48:48.047241 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:48:48 crc kubenswrapper[4885]: I1002 01:48:48.047834 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sv4md" Oct 02 01:48:48 crc kubenswrapper[4885]: E1002 01:48:48.047991 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sv4md" podUID="724a3e49-001c-45a4-a896-82c13e0ee0e7" Oct 02 01:48:48 crc kubenswrapper[4885]: I1002 01:48:48.048329 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:48:48 crc kubenswrapper[4885]: E1002 01:48:48.048467 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:48:48 crc kubenswrapper[4885]: I1002 01:48:48.049988 4885 scope.go:117] "RemoveContainer" containerID="2ef80e838accc59ad7a579c66f4fcd6881d23fa86fc5922d4b0085d211d5d94b" Oct 02 01:48:48 crc kubenswrapper[4885]: E1002 01:48:48.050255 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-dngcm_openshift-ovn-kubernetes(9e4679dd-f870-41e9-a32b-360f5d2ee81b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" podUID="9e4679dd-f870-41e9-a32b-360f5d2ee81b" Oct 02 01:48:48 crc kubenswrapper[4885]: I1002 01:48:48.781815 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-vhgbd_ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f/kube-multus/1.log" Oct 02 01:48:50 crc kubenswrapper[4885]: I1002 01:48:50.045994 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sv4md" Oct 02 01:48:50 crc kubenswrapper[4885]: I1002 01:48:50.046069 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:48:50 crc kubenswrapper[4885]: E1002 01:48:50.046202 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sv4md" podUID="724a3e49-001c-45a4-a896-82c13e0ee0e7" Oct 02 01:48:50 crc kubenswrapper[4885]: E1002 01:48:50.046439 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:48:50 crc kubenswrapper[4885]: I1002 01:48:50.047747 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:48:50 crc kubenswrapper[4885]: I1002 01:48:50.047871 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:48:50 crc kubenswrapper[4885]: E1002 01:48:50.048253 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:48:50 crc kubenswrapper[4885]: E1002 01:48:50.048087 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:48:52 crc kubenswrapper[4885]: E1002 01:48:52.004452 4885 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Oct 02 01:48:52 crc kubenswrapper[4885]: I1002 01:48:52.046522 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sv4md" Oct 02 01:48:52 crc kubenswrapper[4885]: E1002 01:48:52.048539 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sv4md" podUID="724a3e49-001c-45a4-a896-82c13e0ee0e7" Oct 02 01:48:52 crc kubenswrapper[4885]: I1002 01:48:52.048593 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:48:52 crc kubenswrapper[4885]: I1002 01:48:52.048580 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:48:52 crc kubenswrapper[4885]: E1002 01:48:52.048744 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:48:52 crc kubenswrapper[4885]: I1002 01:48:52.049455 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:48:52 crc kubenswrapper[4885]: E1002 01:48:52.049672 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:48:52 crc kubenswrapper[4885]: E1002 01:48:52.049916 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:48:52 crc kubenswrapper[4885]: E1002 01:48:52.167988 4885 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 02 01:48:54 crc kubenswrapper[4885]: I1002 01:48:54.045914 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:48:54 crc kubenswrapper[4885]: I1002 01:48:54.045949 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sv4md" Oct 02 01:48:54 crc kubenswrapper[4885]: E1002 01:48:54.046571 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:48:54 crc kubenswrapper[4885]: I1002 01:48:54.045984 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:48:54 crc kubenswrapper[4885]: E1002 01:48:54.046673 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sv4md" podUID="724a3e49-001c-45a4-a896-82c13e0ee0e7" Oct 02 01:48:54 crc kubenswrapper[4885]: I1002 01:48:54.045914 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:48:54 crc kubenswrapper[4885]: E1002 01:48:54.046836 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:48:54 crc kubenswrapper[4885]: E1002 01:48:54.046971 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:48:56 crc kubenswrapper[4885]: I1002 01:48:56.046037 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sv4md" Oct 02 01:48:56 crc kubenswrapper[4885]: I1002 01:48:56.046308 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:48:56 crc kubenswrapper[4885]: E1002 01:48:56.046483 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sv4md" podUID="724a3e49-001c-45a4-a896-82c13e0ee0e7" Oct 02 01:48:56 crc kubenswrapper[4885]: I1002 01:48:56.046581 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:48:56 crc kubenswrapper[4885]: I1002 01:48:56.046605 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:48:56 crc kubenswrapper[4885]: E1002 01:48:56.046694 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:48:56 crc kubenswrapper[4885]: E1002 01:48:56.046817 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:48:56 crc kubenswrapper[4885]: E1002 01:48:56.046938 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:48:57 crc kubenswrapper[4885]: E1002 01:48:57.168990 4885 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 02 01:48:58 crc kubenswrapper[4885]: I1002 01:48:58.046428 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sv4md" Oct 02 01:48:58 crc kubenswrapper[4885]: I1002 01:48:58.046488 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:48:58 crc kubenswrapper[4885]: I1002 01:48:58.046545 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:48:58 crc kubenswrapper[4885]: I1002 01:48:58.046570 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:48:58 crc kubenswrapper[4885]: E1002 01:48:58.046681 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sv4md" podUID="724a3e49-001c-45a4-a896-82c13e0ee0e7" Oct 02 01:48:58 crc kubenswrapper[4885]: E1002 01:48:58.046836 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:48:58 crc kubenswrapper[4885]: E1002 01:48:58.046972 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:48:58 crc kubenswrapper[4885]: E1002 01:48:58.047185 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:48:59 crc kubenswrapper[4885]: I1002 01:48:59.046666 4885 scope.go:117] "RemoveContainer" containerID="df9792d80c807ed98c8442822d01f619327361c40425abfe26f11e5c27ba2584" Oct 02 01:48:59 crc kubenswrapper[4885]: I1002 01:48:59.826860 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-vhgbd_ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f/kube-multus/1.log" Oct 02 01:48:59 crc kubenswrapper[4885]: I1002 01:48:59.826949 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-vhgbd" event={"ID":"ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f","Type":"ContainerStarted","Data":"7476bbb9bcb8915493b4060254e1303bd96e825ec40425de522dce0a357fb6e6"} Oct 02 01:49:00 crc kubenswrapper[4885]: I1002 01:49:00.046312 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:49:00 crc kubenswrapper[4885]: I1002 01:49:00.046355 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:49:00 crc kubenswrapper[4885]: I1002 01:49:00.046410 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:49:00 crc kubenswrapper[4885]: I1002 01:49:00.046384 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sv4md" Oct 02 01:49:00 crc kubenswrapper[4885]: E1002 01:49:00.046565 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:49:00 crc kubenswrapper[4885]: E1002 01:49:00.046650 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:49:00 crc kubenswrapper[4885]: E1002 01:49:00.046743 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:49:00 crc kubenswrapper[4885]: E1002 01:49:00.046839 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sv4md" podUID="724a3e49-001c-45a4-a896-82c13e0ee0e7" Oct 02 01:49:01 crc kubenswrapper[4885]: I1002 01:49:01.047311 4885 scope.go:117] "RemoveContainer" containerID="2ef80e838accc59ad7a579c66f4fcd6881d23fa86fc5922d4b0085d211d5d94b" Oct 02 01:49:01 crc kubenswrapper[4885]: I1002 01:49:01.834466 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dngcm_9e4679dd-f870-41e9-a32b-360f5d2ee81b/ovnkube-controller/3.log" Oct 02 01:49:01 crc kubenswrapper[4885]: I1002 01:49:01.837685 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" event={"ID":"9e4679dd-f870-41e9-a32b-360f5d2ee81b","Type":"ContainerStarted","Data":"41121c480bb0f68a539497e7d1018e6582a1cf812f819b4c7327ad8827b316b7"} Oct 02 01:49:01 crc kubenswrapper[4885]: I1002 01:49:01.838146 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" Oct 02 01:49:01 crc kubenswrapper[4885]: I1002 01:49:01.880622 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" podStartSLOduration=108.880588242 podStartE2EDuration="1m48.880588242s" podCreationTimestamp="2025-10-02 01:47:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:49:01.87950565 +0000 UTC m=+130.691253059" watchObservedRunningTime="2025-10-02 01:49:01.880588242 +0000 UTC m=+130.692335681" Oct 02 01:49:01 crc kubenswrapper[4885]: I1002 01:49:01.977198 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-sv4md"] Oct 02 01:49:01 crc kubenswrapper[4885]: I1002 01:49:01.977373 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sv4md" Oct 02 01:49:01 crc kubenswrapper[4885]: E1002 01:49:01.977510 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sv4md" podUID="724a3e49-001c-45a4-a896-82c13e0ee0e7" Oct 02 01:49:02 crc kubenswrapper[4885]: I1002 01:49:02.045797 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:49:02 crc kubenswrapper[4885]: I1002 01:49:02.045895 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:49:02 crc kubenswrapper[4885]: I1002 01:49:02.045903 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:49:02 crc kubenswrapper[4885]: E1002 01:49:02.047518 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:49:02 crc kubenswrapper[4885]: E1002 01:49:02.047703 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:49:02 crc kubenswrapper[4885]: E1002 01:49:02.047804 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:49:02 crc kubenswrapper[4885]: E1002 01:49:02.169763 4885 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 02 01:49:03 crc kubenswrapper[4885]: I1002 01:49:03.046666 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sv4md" Oct 02 01:49:03 crc kubenswrapper[4885]: E1002 01:49:03.046906 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sv4md" podUID="724a3e49-001c-45a4-a896-82c13e0ee0e7" Oct 02 01:49:04 crc kubenswrapper[4885]: I1002 01:49:04.046443 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:49:04 crc kubenswrapper[4885]: I1002 01:49:04.046531 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:49:04 crc kubenswrapper[4885]: I1002 01:49:04.046573 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:49:04 crc kubenswrapper[4885]: E1002 01:49:04.046627 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:49:04 crc kubenswrapper[4885]: E1002 01:49:04.046765 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:49:04 crc kubenswrapper[4885]: E1002 01:49:04.046872 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:49:05 crc kubenswrapper[4885]: I1002 01:49:05.046416 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sv4md" Oct 02 01:49:05 crc kubenswrapper[4885]: E1002 01:49:05.046617 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sv4md" podUID="724a3e49-001c-45a4-a896-82c13e0ee0e7" Oct 02 01:49:06 crc kubenswrapper[4885]: I1002 01:49:06.046888 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:49:06 crc kubenswrapper[4885]: I1002 01:49:06.046907 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:49:06 crc kubenswrapper[4885]: E1002 01:49:06.047080 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:49:06 crc kubenswrapper[4885]: I1002 01:49:06.047199 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:49:06 crc kubenswrapper[4885]: E1002 01:49:06.047361 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:49:06 crc kubenswrapper[4885]: E1002 01:49:06.047528 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:49:07 crc kubenswrapper[4885]: I1002 01:49:07.046326 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sv4md" Oct 02 01:49:07 crc kubenswrapper[4885]: E1002 01:49:07.046527 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sv4md" podUID="724a3e49-001c-45a4-a896-82c13e0ee0e7" Oct 02 01:49:08 crc kubenswrapper[4885]: I1002 01:49:08.046432 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:49:08 crc kubenswrapper[4885]: I1002 01:49:08.046537 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:49:08 crc kubenswrapper[4885]: I1002 01:49:08.046612 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:49:08 crc kubenswrapper[4885]: I1002 01:49:08.050810 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Oct 02 01:49:08 crc kubenswrapper[4885]: I1002 01:49:08.050882 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Oct 02 01:49:08 crc kubenswrapper[4885]: I1002 01:49:08.052865 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Oct 02 01:49:08 crc kubenswrapper[4885]: I1002 01:49:08.054374 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Oct 02 01:49:09 crc kubenswrapper[4885]: I1002 01:49:09.045744 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sv4md" Oct 02 01:49:09 crc kubenswrapper[4885]: I1002 01:49:09.049106 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Oct 02 01:49:09 crc kubenswrapper[4885]: I1002 01:49:09.049227 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.357539 4885 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.408655 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-vs6d8"] Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.409083 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-vs6d8" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.410587 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-smtqm"] Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.410972 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.411442 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.412912 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-mc59z"] Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.413171 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-gg9lz"] Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.413314 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-smtqm" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.413886 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-g4vj4"] Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.414056 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-mc59z" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.414181 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-gg9lz" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.415838 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-g4vj4" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.435412 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.435527 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.435770 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.437157 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.437227 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.439070 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.440766 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8d6k4"] Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.441156 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.441204 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8d6k4" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.441438 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.441740 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.441884 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.442042 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.442203 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.442629 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.442856 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.443000 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.443135 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.443286 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.443438 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.443702 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.443869 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.444239 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.444491 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.444691 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.445082 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.445290 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.445922 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.446112 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.446253 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.446427 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.447634 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.448366 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-hg5rx"] Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.448496 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.449070 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-hg5rx" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.454425 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-5kn8g"] Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.464406 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.465455 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-5kn8g" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.481485 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-fxhqt"] Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.481969 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.482111 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-fxhqt" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.482637 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5vkgw"] Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.483149 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5vkgw" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.483340 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.483425 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.483757 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.483860 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-5wldz"] Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.484023 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.484171 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.484305 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.484564 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-5wldz" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.484657 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-hkwl9"] Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.485495 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.485670 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.485800 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.486026 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6w2pm"] Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.487085 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6w2pm" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.487397 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hkwl9" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.487869 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-kckm8"] Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.488276 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-kckm8" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.488649 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-98cfh"] Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.489062 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-98cfh" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.490832 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.496106 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.496314 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.499387 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.500603 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.500985 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-swrb9"] Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.501799 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.503228 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.507022 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-8j5cm"] Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.507336 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-gdxhh"] Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.507573 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-7nhrb"] Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.511252 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-gdxhh" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.512571 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.515248 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.515498 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-swrb9" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.516189 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.516473 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.516822 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.517904 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.517916 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.518133 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.518191 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-8j5cm" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.518394 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.541162 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.541348 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.541416 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.541476 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.541540 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.542561 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.542686 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.542826 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.543553 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.543674 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.543954 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.543734 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.543832 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.544092 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.544128 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.544156 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.543959 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.544191 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.544302 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.544476 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.544539 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.544564 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.544618 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.544824 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.544887 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.545351 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.545432 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.545537 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-66jkw"] Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.545922 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-zqxvm"] Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.546169 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-f6gkd"] Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.546225 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.546487 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.546536 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dpqkj"] Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.546669 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.546830 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-66jkw" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.546854 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-7nhrb" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.546942 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-zqxvm" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.546951 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.547008 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.547073 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.547146 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.546980 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.547241 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.547326 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-f6gkd" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.547350 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dpqkj" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.547467 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/45e90d9b-8520-4b1f-8f16-170787f4c8b6-serving-cert\") pod \"route-controller-manager-6576b87f9c-g4vj4\" (UID: \"45e90d9b-8520-4b1f-8f16-170787f4c8b6\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-g4vj4" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.547495 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f0062618-a397-45c0-a4f5-df3ebe244228-etcd-client\") pod \"apiserver-76f77b778f-hg5rx\" (UID: \"f0062618-a397-45c0-a4f5-df3ebe244228\") " pod="openshift-apiserver/apiserver-76f77b778f-hg5rx" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.547513 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/f0062618-a397-45c0-a4f5-df3ebe244228-image-import-ca\") pod \"apiserver-76f77b778f-hg5rx\" (UID: \"f0062618-a397-45c0-a4f5-df3ebe244228\") " pod="openshift-apiserver/apiserver-76f77b778f-hg5rx" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.547535 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c57e08c1-826c-4332-8a9e-50f419231d73-config\") pod \"console-operator-58897d9998-fxhqt\" (UID: \"c57e08c1-826c-4332-8a9e-50f419231d73\") " pod="openshift-console-operator/console-operator-58897d9998-fxhqt" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.547551 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wfjz8\" (UniqueName: \"kubernetes.io/projected/f0062618-a397-45c0-a4f5-df3ebe244228-kube-api-access-wfjz8\") pod \"apiserver-76f77b778f-hg5rx\" (UID: \"f0062618-a397-45c0-a4f5-df3ebe244228\") " pod="openshift-apiserver/apiserver-76f77b778f-hg5rx" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.547568 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5feb1fb5-73be-4ce5-a6d5-d426e2eada2e-config\") pod \"machine-approver-56656f9798-smtqm\" (UID: \"5feb1fb5-73be-4ce5-a6d5-d426e2eada2e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-smtqm" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.547593 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3a80ea92-4f96-4916-a00d-3f010959194d-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-8d6k4\" (UID: \"3a80ea92-4f96-4916-a00d-3f010959194d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8d6k4" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.547607 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f0062618-a397-45c0-a4f5-df3ebe244228-config\") pod \"apiserver-76f77b778f-hg5rx\" (UID: \"f0062618-a397-45c0-a4f5-df3ebe244228\") " pod="openshift-apiserver/apiserver-76f77b778f-hg5rx" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.547623 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/b6e3c195-a333-4c4f-8b75-71edbfcc26c3-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-mc59z\" (UID: \"b6e3c195-a333-4c4f-8b75-71edbfcc26c3\") " pod="openshift-authentication/oauth-openshift-558db77b4-mc59z" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.547643 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/b6e3c195-a333-4c4f-8b75-71edbfcc26c3-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-mc59z\" (UID: \"b6e3c195-a333-4c4f-8b75-71edbfcc26c3\") " pod="openshift-authentication/oauth-openshift-558db77b4-mc59z" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.547662 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hfdzk\" (UniqueName: \"kubernetes.io/projected/243fb6e4-1bf5-4c4a-b471-87f50d5e147d-kube-api-access-hfdzk\") pod \"openshift-config-operator-7777fb866f-gg9lz\" (UID: \"243fb6e4-1bf5-4c4a-b471-87f50d5e147d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-gg9lz" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.547677 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/f0062618-a397-45c0-a4f5-df3ebe244228-audit\") pod \"apiserver-76f77b778f-hg5rx\" (UID: \"f0062618-a397-45c0-a4f5-df3ebe244228\") " pod="openshift-apiserver/apiserver-76f77b778f-hg5rx" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.547683 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.547704 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/45e90d9b-8520-4b1f-8f16-170787f4c8b6-config\") pod \"route-controller-manager-6576b87f9c-g4vj4\" (UID: \"45e90d9b-8520-4b1f-8f16-170787f4c8b6\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-g4vj4" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.547720 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/b6e3c195-a333-4c4f-8b75-71edbfcc26c3-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-mc59z\" (UID: \"b6e3c195-a333-4c4f-8b75-71edbfcc26c3\") " pod="openshift-authentication/oauth-openshift-558db77b4-mc59z" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.547831 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.547863 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/075528cd-eab2-47bb-8aa0-3d39002fb3d1-client-ca\") pod \"controller-manager-879f6c89f-vs6d8\" (UID: \"075528cd-eab2-47bb-8aa0-3d39002fb3d1\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vs6d8" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.547869 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.547953 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b6e3c195-a333-4c4f-8b75-71edbfcc26c3-audit-dir\") pod \"oauth-openshift-558db77b4-mc59z\" (UID: \"b6e3c195-a333-4c4f-8b75-71edbfcc26c3\") " pod="openshift-authentication/oauth-openshift-558db77b4-mc59z" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.547989 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c57e08c1-826c-4332-8a9e-50f419231d73-serving-cert\") pod \"console-operator-58897d9998-fxhqt\" (UID: \"c57e08c1-826c-4332-8a9e-50f419231d73\") " pod="openshift-console-operator/console-operator-58897d9998-fxhqt" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.548013 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c57e08c1-826c-4332-8a9e-50f419231d73-trusted-ca\") pod \"console-operator-58897d9998-fxhqt\" (UID: \"c57e08c1-826c-4332-8a9e-50f419231d73\") " pod="openshift-console-operator/console-operator-58897d9998-fxhqt" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.548028 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/075528cd-eab2-47bb-8aa0-3d39002fb3d1-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-vs6d8\" (UID: \"075528cd-eab2-47bb-8aa0-3d39002fb3d1\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vs6d8" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.548049 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5feb1fb5-73be-4ce5-a6d5-d426e2eada2e-auth-proxy-config\") pod \"machine-approver-56656f9798-smtqm\" (UID: \"5feb1fb5-73be-4ce5-a6d5-d426e2eada2e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-smtqm" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.548067 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-shk8x\" (UniqueName: \"kubernetes.io/projected/5feb1fb5-73be-4ce5-a6d5-d426e2eada2e-kube-api-access-shk8x\") pod \"machine-approver-56656f9798-smtqm\" (UID: \"5feb1fb5-73be-4ce5-a6d5-d426e2eada2e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-smtqm" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.548087 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2qr2q\" (UniqueName: \"kubernetes.io/projected/fd852137-cff1-4fae-a921-42be5c511db8-kube-api-access-2qr2q\") pod \"downloads-7954f5f757-5kn8g\" (UID: \"fd852137-cff1-4fae-a921-42be5c511db8\") " pod="openshift-console/downloads-7954f5f757-5kn8g" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.548158 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/243fb6e4-1bf5-4c4a-b471-87f50d5e147d-serving-cert\") pod \"openshift-config-operator-7777fb866f-gg9lz\" (UID: \"243fb6e4-1bf5-4c4a-b471-87f50d5e147d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-gg9lz" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.548194 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/075528cd-eab2-47bb-8aa0-3d39002fb3d1-config\") pod \"controller-manager-879f6c89f-vs6d8\" (UID: \"075528cd-eab2-47bb-8aa0-3d39002fb3d1\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vs6d8" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.548237 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/f0062618-a397-45c0-a4f5-df3ebe244228-node-pullsecrets\") pod \"apiserver-76f77b778f-hg5rx\" (UID: \"f0062618-a397-45c0-a4f5-df3ebe244228\") " pod="openshift-apiserver/apiserver-76f77b778f-hg5rx" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.548299 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/b6e3c195-a333-4c4f-8b75-71edbfcc26c3-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-mc59z\" (UID: \"b6e3c195-a333-4c4f-8b75-71edbfcc26c3\") " pod="openshift-authentication/oauth-openshift-558db77b4-mc59z" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.548331 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l5s4w\" (UniqueName: \"kubernetes.io/projected/b6e3c195-a333-4c4f-8b75-71edbfcc26c3-kube-api-access-l5s4w\") pod \"oauth-openshift-558db77b4-mc59z\" (UID: \"b6e3c195-a333-4c4f-8b75-71edbfcc26c3\") " pod="openshift-authentication/oauth-openshift-558db77b4-mc59z" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.548389 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f0062618-a397-45c0-a4f5-df3ebe244228-trusted-ca-bundle\") pod \"apiserver-76f77b778f-hg5rx\" (UID: \"f0062618-a397-45c0-a4f5-df3ebe244228\") " pod="openshift-apiserver/apiserver-76f77b778f-hg5rx" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.548411 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/f0062618-a397-45c0-a4f5-df3ebe244228-etcd-serving-ca\") pod \"apiserver-76f77b778f-hg5rx\" (UID: \"f0062618-a397-45c0-a4f5-df3ebe244228\") " pod="openshift-apiserver/apiserver-76f77b778f-hg5rx" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.548426 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3a80ea92-4f96-4916-a00d-3f010959194d-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-8d6k4\" (UID: \"3a80ea92-4f96-4916-a00d-3f010959194d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8d6k4" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.548459 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/b6e3c195-a333-4c4f-8b75-71edbfcc26c3-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-mc59z\" (UID: \"b6e3c195-a333-4c4f-8b75-71edbfcc26c3\") " pod="openshift-authentication/oauth-openshift-558db77b4-mc59z" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.548481 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/b6e3c195-a333-4c4f-8b75-71edbfcc26c3-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-mc59z\" (UID: \"b6e3c195-a333-4c4f-8b75-71edbfcc26c3\") " pod="openshift-authentication/oauth-openshift-558db77b4-mc59z" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.548498 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/b6e3c195-a333-4c4f-8b75-71edbfcc26c3-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-mc59z\" (UID: \"b6e3c195-a333-4c4f-8b75-71edbfcc26c3\") " pod="openshift-authentication/oauth-openshift-558db77b4-mc59z" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.548515 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/b6e3c195-a333-4c4f-8b75-71edbfcc26c3-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-mc59z\" (UID: \"b6e3c195-a333-4c4f-8b75-71edbfcc26c3\") " pod="openshift-authentication/oauth-openshift-558db77b4-mc59z" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.548552 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b6e3c195-a333-4c4f-8b75-71edbfcc26c3-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-mc59z\" (UID: \"b6e3c195-a333-4c4f-8b75-71edbfcc26c3\") " pod="openshift-authentication/oauth-openshift-558db77b4-mc59z" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.548569 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/b6e3c195-a333-4c4f-8b75-71edbfcc26c3-audit-policies\") pod \"oauth-openshift-558db77b4-mc59z\" (UID: \"b6e3c195-a333-4c4f-8b75-71edbfcc26c3\") " pod="openshift-authentication/oauth-openshift-558db77b4-mc59z" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.548579 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.548583 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rq2s2\" (UniqueName: \"kubernetes.io/projected/c57e08c1-826c-4332-8a9e-50f419231d73-kube-api-access-rq2s2\") pod \"console-operator-58897d9998-fxhqt\" (UID: \"c57e08c1-826c-4332-8a9e-50f419231d73\") " pod="openshift-console-operator/console-operator-58897d9998-fxhqt" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.548713 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f0062618-a397-45c0-a4f5-df3ebe244228-serving-cert\") pod \"apiserver-76f77b778f-hg5rx\" (UID: \"f0062618-a397-45c0-a4f5-df3ebe244228\") " pod="openshift-apiserver/apiserver-76f77b778f-hg5rx" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.548725 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-vbm65"] Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.548747 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/5feb1fb5-73be-4ce5-a6d5-d426e2eada2e-machine-approver-tls\") pod \"machine-approver-56656f9798-smtqm\" (UID: \"5feb1fb5-73be-4ce5-a6d5-d426e2eada2e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-smtqm" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.548772 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/075528cd-eab2-47bb-8aa0-3d39002fb3d1-serving-cert\") pod \"controller-manager-879f6c89f-vs6d8\" (UID: \"075528cd-eab2-47bb-8aa0-3d39002fb3d1\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vs6d8" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.548794 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vw9xq\" (UniqueName: \"kubernetes.io/projected/075528cd-eab2-47bb-8aa0-3d39002fb3d1-kube-api-access-vw9xq\") pod \"controller-manager-879f6c89f-vs6d8\" (UID: \"075528cd-eab2-47bb-8aa0-3d39002fb3d1\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vs6d8" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.548816 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-55ssq\" (UniqueName: \"kubernetes.io/projected/3a80ea92-4f96-4916-a00d-3f010959194d-kube-api-access-55ssq\") pod \"cluster-image-registry-operator-dc59b4c8b-8d6k4\" (UID: \"3a80ea92-4f96-4916-a00d-3f010959194d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8d6k4" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.548836 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mpmq4\" (UniqueName: \"kubernetes.io/projected/45e90d9b-8520-4b1f-8f16-170787f4c8b6-kube-api-access-mpmq4\") pod \"route-controller-manager-6576b87f9c-g4vj4\" (UID: \"45e90d9b-8520-4b1f-8f16-170787f4c8b6\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-g4vj4" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.548879 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/f0062618-a397-45c0-a4f5-df3ebe244228-encryption-config\") pod \"apiserver-76f77b778f-hg5rx\" (UID: \"f0062618-a397-45c0-a4f5-df3ebe244228\") " pod="openshift-apiserver/apiserver-76f77b778f-hg5rx" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.548906 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/b6e3c195-a333-4c4f-8b75-71edbfcc26c3-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-mc59z\" (UID: \"b6e3c195-a333-4c4f-8b75-71edbfcc26c3\") " pod="openshift-authentication/oauth-openshift-558db77b4-mc59z" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.548936 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/b6e3c195-a333-4c4f-8b75-71edbfcc26c3-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-mc59z\" (UID: \"b6e3c195-a333-4c4f-8b75-71edbfcc26c3\") " pod="openshift-authentication/oauth-openshift-558db77b4-mc59z" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.549014 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/45e90d9b-8520-4b1f-8f16-170787f4c8b6-client-ca\") pod \"route-controller-manager-6576b87f9c-g4vj4\" (UID: \"45e90d9b-8520-4b1f-8f16-170787f4c8b6\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-g4vj4" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.549055 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/243fb6e4-1bf5-4c4a-b471-87f50d5e147d-available-featuregates\") pod \"openshift-config-operator-7777fb866f-gg9lz\" (UID: \"243fb6e4-1bf5-4c4a-b471-87f50d5e147d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-gg9lz" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.549106 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.549167 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.549152 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-vbm65" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.549110 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f0062618-a397-45c0-a4f5-df3ebe244228-audit-dir\") pod \"apiserver-76f77b778f-hg5rx\" (UID: \"f0062618-a397-45c0-a4f5-df3ebe244228\") " pod="openshift-apiserver/apiserver-76f77b778f-hg5rx" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.549385 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/3a80ea92-4f96-4916-a00d-3f010959194d-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-8d6k4\" (UID: \"3a80ea92-4f96-4916-a00d-3f010959194d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8d6k4" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.550820 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.551594 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.554616 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.558355 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-zxz72"] Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.558842 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-qgdp5"] Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.560656 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-zxz72" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.561175 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-8nb8p"] Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.561573 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8nb8p" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.561712 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-qgdp5" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.562066 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-kgxn6"] Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.562578 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-kgxn6" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.563898 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-xd4hd"] Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.564357 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-xd4hd" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.564449 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cm9ql"] Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.564881 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cm9ql" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.574179 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8d6k4"] Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.574777 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.574943 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-z4rlr"] Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.575450 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-z4rlr" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.576174 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-5lbxn"] Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.576818 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-5lbxn" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.579300 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-gg9lz"] Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.579620 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-4wjj8"] Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.590637 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.593854 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xjfdv"] Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.594047 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-4wjj8" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.594494 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-c94sm"] Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.594740 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xjfdv" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.594770 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-5kn8g"] Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.594785 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322825-b7zrt"] Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.594831 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-c94sm" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.595731 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-k8r8q"] Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.605036 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322825-b7zrt" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.610488 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-7hk2v"] Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.610857 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.610965 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-k8r8q" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.611626 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-vs6d8"] Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.611645 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-n7bs2"] Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.611958 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-7hk2v" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.617919 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-hg5rx"] Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.617960 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-g4vj4"] Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.617989 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-swrb9"] Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.618268 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-n7bs2" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.618344 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-5wldz"] Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.624740 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-mc59z"] Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.626995 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-66jkw"] Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.629189 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5vkgw"] Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.629282 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-98cfh"] Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.630547 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.630940 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dpqkj"] Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.632415 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-fxhqt"] Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.636099 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-8j5cm"] Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.637749 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-xft2d"] Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.638486 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-xft2d" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.639623 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-xd4hd"] Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.641024 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-kckm8"] Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.643696 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-z4rlr"] Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.645161 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-4wjj8"] Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.646152 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-hkwl9"] Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.648877 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-f6gkd"] Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.649990 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/243fb6e4-1bf5-4c4a-b471-87f50d5e147d-serving-cert\") pod \"openshift-config-operator-7777fb866f-gg9lz\" (UID: \"243fb6e4-1bf5-4c4a-b471-87f50d5e147d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-gg9lz" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.650022 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/075528cd-eab2-47bb-8aa0-3d39002fb3d1-config\") pod \"controller-manager-879f6c89f-vs6d8\" (UID: \"075528cd-eab2-47bb-8aa0-3d39002fb3d1\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vs6d8" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.650042 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/f0062618-a397-45c0-a4f5-df3ebe244228-node-pullsecrets\") pod \"apiserver-76f77b778f-hg5rx\" (UID: \"f0062618-a397-45c0-a4f5-df3ebe244228\") " pod="openshift-apiserver/apiserver-76f77b778f-hg5rx" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.650061 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6vhnl\" (UniqueName: \"kubernetes.io/projected/3de23ab3-e037-4163-b744-81ecbb340b13-kube-api-access-6vhnl\") pod \"openshift-apiserver-operator-796bbdcf4f-5vkgw\" (UID: \"3de23ab3-e037-4163-b744-81ecbb340b13\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5vkgw" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.650082 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/76fb6dce-01dd-4aef-8213-e03be7a5e055-images\") pod \"machine-api-operator-5694c8668f-5wldz\" (UID: \"76fb6dce-01dd-4aef-8213-e03be7a5e055\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5wldz" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.650103 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/b6e3c195-a333-4c4f-8b75-71edbfcc26c3-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-mc59z\" (UID: \"b6e3c195-a333-4c4f-8b75-71edbfcc26c3\") " pod="openshift-authentication/oauth-openshift-558db77b4-mc59z" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.650119 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c3a7de1e-09f7-41eb-8bf2-3737eee4ae74-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-8j5cm\" (UID: \"c3a7de1e-09f7-41eb-8bf2-3737eee4ae74\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8j5cm" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.650148 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/4a781c56-3979-49df-a7f1-340c44bcc9b5-srv-cert\") pod \"olm-operator-6b444d44fb-cm9ql\" (UID: \"4a781c56-3979-49df-a7f1-340c44bcc9b5\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cm9ql" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.650165 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/e925f407-5010-4a6f-bcb5-b555f2f7f8b1-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-6w2pm\" (UID: \"e925f407-5010-4a6f-bcb5-b555f2f7f8b1\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6w2pm" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.650183 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zh7sn\" (UniqueName: \"kubernetes.io/projected/88149eca-e1e7-4dd4-8cdf-170db9460b30-kube-api-access-zh7sn\") pod \"apiserver-7bbb656c7d-hkwl9\" (UID: \"88149eca-e1e7-4dd4-8cdf-170db9460b30\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hkwl9" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.650201 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l5s4w\" (UniqueName: \"kubernetes.io/projected/b6e3c195-a333-4c4f-8b75-71edbfcc26c3-kube-api-access-l5s4w\") pod \"oauth-openshift-558db77b4-mc59z\" (UID: \"b6e3c195-a333-4c4f-8b75-71edbfcc26c3\") " pod="openshift-authentication/oauth-openshift-558db77b4-mc59z" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.650215 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pm4kq\" (UniqueName: \"kubernetes.io/projected/472cd80c-0597-40b4-babb-89a3fce346b0-kube-api-access-pm4kq\") pod \"dns-operator-744455d44c-swrb9\" (UID: \"472cd80c-0597-40b4-babb-89a3fce346b0\") " pod="openshift-dns-operator/dns-operator-744455d44c-swrb9" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.650238 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/88149eca-e1e7-4dd4-8cdf-170db9460b30-audit-dir\") pod \"apiserver-7bbb656c7d-hkwl9\" (UID: \"88149eca-e1e7-4dd4-8cdf-170db9460b30\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hkwl9" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.650252 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/028a80a9-9747-4856-a2e8-f721abd82683-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-zqxvm\" (UID: \"028a80a9-9747-4856-a2e8-f721abd82683\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-zqxvm" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.650294 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/f0062618-a397-45c0-a4f5-df3ebe244228-etcd-serving-ca\") pod \"apiserver-76f77b778f-hg5rx\" (UID: \"f0062618-a397-45c0-a4f5-df3ebe244228\") " pod="openshift-apiserver/apiserver-76f77b778f-hg5rx" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.650310 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f0062618-a397-45c0-a4f5-df3ebe244228-trusted-ca-bundle\") pod \"apiserver-76f77b778f-hg5rx\" (UID: \"f0062618-a397-45c0-a4f5-df3ebe244228\") " pod="openshift-apiserver/apiserver-76f77b778f-hg5rx" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.650324 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c3a7de1e-09f7-41eb-8bf2-3737eee4ae74-config\") pod \"authentication-operator-69f744f599-8j5cm\" (UID: \"c3a7de1e-09f7-41eb-8bf2-3737eee4ae74\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8j5cm" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.650341 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c3a7de1e-09f7-41eb-8bf2-3737eee4ae74-serving-cert\") pod \"authentication-operator-69f744f599-8j5cm\" (UID: \"c3a7de1e-09f7-41eb-8bf2-3737eee4ae74\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8j5cm" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.650356 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tmkv7\" (UniqueName: \"kubernetes.io/projected/4a781c56-3979-49df-a7f1-340c44bcc9b5-kube-api-access-tmkv7\") pod \"olm-operator-6b444d44fb-cm9ql\" (UID: \"4a781c56-3979-49df-a7f1-340c44bcc9b5\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cm9ql" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.650373 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3a80ea92-4f96-4916-a00d-3f010959194d-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-8d6k4\" (UID: \"3a80ea92-4f96-4916-a00d-3f010959194d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8d6k4" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.650680 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/b6e3c195-a333-4c4f-8b75-71edbfcc26c3-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-mc59z\" (UID: \"b6e3c195-a333-4c4f-8b75-71edbfcc26c3\") " pod="openshift-authentication/oauth-openshift-558db77b4-mc59z" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.650807 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/f0062618-a397-45c0-a4f5-df3ebe244228-node-pullsecrets\") pod \"apiserver-76f77b778f-hg5rx\" (UID: \"f0062618-a397-45c0-a4f5-df3ebe244228\") " pod="openshift-apiserver/apiserver-76f77b778f-hg5rx" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.650807 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/2c17c274-7853-4717-a68e-9fa6467ee199-proxy-tls\") pod \"machine-config-operator-74547568cd-8nb8p\" (UID: \"2c17c274-7853-4717-a68e-9fa6467ee199\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8nb8p" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.650849 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/88149eca-e1e7-4dd4-8cdf-170db9460b30-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-hkwl9\" (UID: \"88149eca-e1e7-4dd4-8cdf-170db9460b30\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hkwl9" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.650886 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/88149eca-e1e7-4dd4-8cdf-170db9460b30-encryption-config\") pod \"apiserver-7bbb656c7d-hkwl9\" (UID: \"88149eca-e1e7-4dd4-8cdf-170db9460b30\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hkwl9" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.650990 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3de23ab3-e037-4163-b744-81ecbb340b13-config\") pod \"openshift-apiserver-operator-796bbdcf4f-5vkgw\" (UID: \"3de23ab3-e037-4163-b744-81ecbb340b13\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5vkgw" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.651017 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/b6e3c195-a333-4c4f-8b75-71edbfcc26c3-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-mc59z\" (UID: \"b6e3c195-a333-4c4f-8b75-71edbfcc26c3\") " pod="openshift-authentication/oauth-openshift-558db77b4-mc59z" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.651120 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/b6e3c195-a333-4c4f-8b75-71edbfcc26c3-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-mc59z\" (UID: \"b6e3c195-a333-4c4f-8b75-71edbfcc26c3\") " pod="openshift-authentication/oauth-openshift-558db77b4-mc59z" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.651144 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/b6e3c195-a333-4c4f-8b75-71edbfcc26c3-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-mc59z\" (UID: \"b6e3c195-a333-4c4f-8b75-71edbfcc26c3\") " pod="openshift-authentication/oauth-openshift-558db77b4-mc59z" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.651298 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/075528cd-eab2-47bb-8aa0-3d39002fb3d1-config\") pod \"controller-manager-879f6c89f-vs6d8\" (UID: \"075528cd-eab2-47bb-8aa0-3d39002fb3d1\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vs6d8" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.651327 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/f0062618-a397-45c0-a4f5-df3ebe244228-etcd-serving-ca\") pod \"apiserver-76f77b778f-hg5rx\" (UID: \"f0062618-a397-45c0-a4f5-df3ebe244228\") " pod="openshift-apiserver/apiserver-76f77b778f-hg5rx" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.651368 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-6bfsg"] Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.651433 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.652069 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-6bfsg" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.652369 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/b6e3c195-a333-4c4f-8b75-71edbfcc26c3-audit-policies\") pod \"oauth-openshift-558db77b4-mc59z\" (UID: \"b6e3c195-a333-4c4f-8b75-71edbfcc26c3\") " pod="openshift-authentication/oauth-openshift-558db77b4-mc59z" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.652409 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b6e3c195-a333-4c4f-8b75-71edbfcc26c3-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-mc59z\" (UID: \"b6e3c195-a333-4c4f-8b75-71edbfcc26c3\") " pod="openshift-authentication/oauth-openshift-558db77b4-mc59z" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.652432 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/472cd80c-0597-40b4-babb-89a3fce346b0-metrics-tls\") pod \"dns-operator-744455d44c-swrb9\" (UID: \"472cd80c-0597-40b4-babb-89a3fce346b0\") " pod="openshift-dns-operator/dns-operator-744455d44c-swrb9" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.652470 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/028a80a9-9747-4856-a2e8-f721abd82683-config\") pod \"kube-controller-manager-operator-78b949d7b-zqxvm\" (UID: \"028a80a9-9747-4856-a2e8-f721abd82683\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-zqxvm" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.652687 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cm9ql"] Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.652890 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f0062618-a397-45c0-a4f5-df3ebe244228-trusted-ca-bundle\") pod \"apiserver-76f77b778f-hg5rx\" (UID: \"f0062618-a397-45c0-a4f5-df3ebe244228\") " pod="openshift-apiserver/apiserver-76f77b778f-hg5rx" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.653006 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/b6e3c195-a333-4c4f-8b75-71edbfcc26c3-audit-policies\") pod \"oauth-openshift-558db77b4-mc59z\" (UID: \"b6e3c195-a333-4c4f-8b75-71edbfcc26c3\") " pod="openshift-authentication/oauth-openshift-558db77b4-mc59z" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.653072 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rq2s2\" (UniqueName: \"kubernetes.io/projected/c57e08c1-826c-4332-8a9e-50f419231d73-kube-api-access-rq2s2\") pod \"console-operator-58897d9998-fxhqt\" (UID: \"c57e08c1-826c-4332-8a9e-50f419231d73\") " pod="openshift-console-operator/console-operator-58897d9998-fxhqt" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.653111 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/88149eca-e1e7-4dd4-8cdf-170db9460b30-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-hkwl9\" (UID: \"88149eca-e1e7-4dd4-8cdf-170db9460b30\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hkwl9" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.653226 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/075528cd-eab2-47bb-8aa0-3d39002fb3d1-serving-cert\") pod \"controller-manager-879f6c89f-vs6d8\" (UID: \"075528cd-eab2-47bb-8aa0-3d39002fb3d1\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vs6d8" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.653317 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vw9xq\" (UniqueName: \"kubernetes.io/projected/075528cd-eab2-47bb-8aa0-3d39002fb3d1-kube-api-access-vw9xq\") pod \"controller-manager-879f6c89f-vs6d8\" (UID: \"075528cd-eab2-47bb-8aa0-3d39002fb3d1\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vs6d8" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.653358 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f0062618-a397-45c0-a4f5-df3ebe244228-serving-cert\") pod \"apiserver-76f77b778f-hg5rx\" (UID: \"f0062618-a397-45c0-a4f5-df3ebe244228\") " pod="openshift-apiserver/apiserver-76f77b778f-hg5rx" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.653377 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/5feb1fb5-73be-4ce5-a6d5-d426e2eada2e-machine-approver-tls\") pod \"machine-approver-56656f9798-smtqm\" (UID: \"5feb1fb5-73be-4ce5-a6d5-d426e2eada2e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-smtqm" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.653480 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/88149eca-e1e7-4dd4-8cdf-170db9460b30-serving-cert\") pod \"apiserver-7bbb656c7d-hkwl9\" (UID: \"88149eca-e1e7-4dd4-8cdf-170db9460b30\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hkwl9" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.653688 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b6e3c195-a333-4c4f-8b75-71edbfcc26c3-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-mc59z\" (UID: \"b6e3c195-a333-4c4f-8b75-71edbfcc26c3\") " pod="openshift-authentication/oauth-openshift-558db77b4-mc59z" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.653760 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-55ssq\" (UniqueName: \"kubernetes.io/projected/3a80ea92-4f96-4916-a00d-3f010959194d-kube-api-access-55ssq\") pod \"cluster-image-registry-operator-dc59b4c8b-8d6k4\" (UID: \"3a80ea92-4f96-4916-a00d-3f010959194d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8d6k4" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.653824 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mpmq4\" (UniqueName: \"kubernetes.io/projected/45e90d9b-8520-4b1f-8f16-170787f4c8b6-kube-api-access-mpmq4\") pod \"route-controller-manager-6576b87f9c-g4vj4\" (UID: \"45e90d9b-8520-4b1f-8f16-170787f4c8b6\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-g4vj4" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.653853 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d8p2l\" (UniqueName: \"kubernetes.io/projected/c3a7de1e-09f7-41eb-8bf2-3737eee4ae74-kube-api-access-d8p2l\") pod \"authentication-operator-69f744f599-8j5cm\" (UID: \"c3a7de1e-09f7-41eb-8bf2-3737eee4ae74\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8j5cm" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.653915 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/76fb6dce-01dd-4aef-8213-e03be7a5e055-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-5wldz\" (UID: \"76fb6dce-01dd-4aef-8213-e03be7a5e055\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5wldz" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.653942 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/f0062618-a397-45c0-a4f5-df3ebe244228-encryption-config\") pod \"apiserver-76f77b778f-hg5rx\" (UID: \"f0062618-a397-45c0-a4f5-df3ebe244228\") " pod="openshift-apiserver/apiserver-76f77b778f-hg5rx" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.653983 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/b6e3c195-a333-4c4f-8b75-71edbfcc26c3-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-mc59z\" (UID: \"b6e3c195-a333-4c4f-8b75-71edbfcc26c3\") " pod="openshift-authentication/oauth-openshift-558db77b4-mc59z" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.654024 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/b6e3c195-a333-4c4f-8b75-71edbfcc26c3-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-mc59z\" (UID: \"b6e3c195-a333-4c4f-8b75-71edbfcc26c3\") " pod="openshift-authentication/oauth-openshift-558db77b4-mc59z" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.654084 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/88149eca-e1e7-4dd4-8cdf-170db9460b30-etcd-client\") pod \"apiserver-7bbb656c7d-hkwl9\" (UID: \"88149eca-e1e7-4dd4-8cdf-170db9460b30\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hkwl9" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.654104 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/45e90d9b-8520-4b1f-8f16-170787f4c8b6-client-ca\") pod \"route-controller-manager-6576b87f9c-g4vj4\" (UID: \"45e90d9b-8520-4b1f-8f16-170787f4c8b6\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-g4vj4" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.654120 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/028a80a9-9747-4856-a2e8-f721abd82683-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-zqxvm\" (UID: \"028a80a9-9747-4856-a2e8-f721abd82683\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-zqxvm" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.654172 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/243fb6e4-1bf5-4c4a-b471-87f50d5e147d-available-featuregates\") pod \"openshift-config-operator-7777fb866f-gg9lz\" (UID: \"243fb6e4-1bf5-4c4a-b471-87f50d5e147d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-gg9lz" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.654189 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rkk59\" (UniqueName: \"kubernetes.io/projected/cc6bb484-0230-46d9-9507-b641d2a5f330-kube-api-access-rkk59\") pod \"control-plane-machine-set-operator-78cbb6b69f-dpqkj\" (UID: \"cc6bb484-0230-46d9-9507-b641d2a5f330\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dpqkj" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.654616 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/b6e3c195-a333-4c4f-8b75-71edbfcc26c3-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-mc59z\" (UID: \"b6e3c195-a333-4c4f-8b75-71edbfcc26c3\") " pod="openshift-authentication/oauth-openshift-558db77b4-mc59z" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.655806 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/b6e3c195-a333-4c4f-8b75-71edbfcc26c3-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-mc59z\" (UID: \"b6e3c195-a333-4c4f-8b75-71edbfcc26c3\") " pod="openshift-authentication/oauth-openshift-558db77b4-mc59z" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.655992 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/243fb6e4-1bf5-4c4a-b471-87f50d5e147d-serving-cert\") pod \"openshift-config-operator-7777fb866f-gg9lz\" (UID: \"243fb6e4-1bf5-4c4a-b471-87f50d5e147d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-gg9lz" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.656044 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/b6e3c195-a333-4c4f-8b75-71edbfcc26c3-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-mc59z\" (UID: \"b6e3c195-a333-4c4f-8b75-71edbfcc26c3\") " pod="openshift-authentication/oauth-openshift-558db77b4-mc59z" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.656077 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6w2pm"] Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.656091 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/075528cd-eab2-47bb-8aa0-3d39002fb3d1-serving-cert\") pod \"controller-manager-879f6c89f-vs6d8\" (UID: \"075528cd-eab2-47bb-8aa0-3d39002fb3d1\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vs6d8" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.656105 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-gdxhh"] Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.656637 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-vbm65"] Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.656686 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h4fbp\" (UniqueName: \"kubernetes.io/projected/2c17c274-7853-4717-a68e-9fa6467ee199-kube-api-access-h4fbp\") pod \"machine-config-operator-74547568cd-8nb8p\" (UID: \"2c17c274-7853-4717-a68e-9fa6467ee199\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8nb8p" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.656717 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/76fb6dce-01dd-4aef-8213-e03be7a5e055-config\") pod \"machine-api-operator-5694c8668f-5wldz\" (UID: \"76fb6dce-01dd-4aef-8213-e03be7a5e055\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5wldz" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.656726 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/45e90d9b-8520-4b1f-8f16-170787f4c8b6-client-ca\") pod \"route-controller-manager-6576b87f9c-g4vj4\" (UID: \"45e90d9b-8520-4b1f-8f16-170787f4c8b6\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-g4vj4" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.656866 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/3a80ea92-4f96-4916-a00d-3f010959194d-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-8d6k4\" (UID: \"3a80ea92-4f96-4916-a00d-3f010959194d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8d6k4" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.656920 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f0062618-a397-45c0-a4f5-df3ebe244228-audit-dir\") pod \"apiserver-76f77b778f-hg5rx\" (UID: \"f0062618-a397-45c0-a4f5-df3ebe244228\") " pod="openshift-apiserver/apiserver-76f77b778f-hg5rx" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.657027 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f0062618-a397-45c0-a4f5-df3ebe244228-audit-dir\") pod \"apiserver-76f77b778f-hg5rx\" (UID: \"f0062618-a397-45c0-a4f5-df3ebe244228\") " pod="openshift-apiserver/apiserver-76f77b778f-hg5rx" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.657089 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/45e90d9b-8520-4b1f-8f16-170787f4c8b6-serving-cert\") pod \"route-controller-manager-6576b87f9c-g4vj4\" (UID: \"45e90d9b-8520-4b1f-8f16-170787f4c8b6\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-g4vj4" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.657691 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/b6e3c195-a333-4c4f-8b75-71edbfcc26c3-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-mc59z\" (UID: \"b6e3c195-a333-4c4f-8b75-71edbfcc26c3\") " pod="openshift-authentication/oauth-openshift-558db77b4-mc59z" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.657736 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xjfdv"] Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.657757 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f0062618-a397-45c0-a4f5-df3ebe244228-etcd-client\") pod \"apiserver-76f77b778f-hg5rx\" (UID: \"f0062618-a397-45c0-a4f5-df3ebe244228\") " pod="openshift-apiserver/apiserver-76f77b778f-hg5rx" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.657790 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/f0062618-a397-45c0-a4f5-df3ebe244228-image-import-ca\") pod \"apiserver-76f77b778f-hg5rx\" (UID: \"f0062618-a397-45c0-a4f5-df3ebe244228\") " pod="openshift-apiserver/apiserver-76f77b778f-hg5rx" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.657820 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8qtsm\" (UniqueName: \"kubernetes.io/projected/76fb6dce-01dd-4aef-8213-e03be7a5e055-kube-api-access-8qtsm\") pod \"machine-api-operator-5694c8668f-5wldz\" (UID: \"76fb6dce-01dd-4aef-8213-e03be7a5e055\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5wldz" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.658746 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/f0062618-a397-45c0-a4f5-df3ebe244228-image-import-ca\") pod \"apiserver-76f77b778f-hg5rx\" (UID: \"f0062618-a397-45c0-a4f5-df3ebe244228\") " pod="openshift-apiserver/apiserver-76f77b778f-hg5rx" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.658759 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-8nb8p"] Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.658787 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c57e08c1-826c-4332-8a9e-50f419231d73-config\") pod \"console-operator-58897d9998-fxhqt\" (UID: \"c57e08c1-826c-4332-8a9e-50f419231d73\") " pod="openshift-console-operator/console-operator-58897d9998-fxhqt" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.658822 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wfjz8\" (UniqueName: \"kubernetes.io/projected/f0062618-a397-45c0-a4f5-df3ebe244228-kube-api-access-wfjz8\") pod \"apiserver-76f77b778f-hg5rx\" (UID: \"f0062618-a397-45c0-a4f5-df3ebe244228\") " pod="openshift-apiserver/apiserver-76f77b778f-hg5rx" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.658838 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5feb1fb5-73be-4ce5-a6d5-d426e2eada2e-config\") pod \"machine-approver-56656f9798-smtqm\" (UID: \"5feb1fb5-73be-4ce5-a6d5-d426e2eada2e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-smtqm" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.658875 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sqkml\" (UniqueName: \"kubernetes.io/projected/e925f407-5010-4a6f-bcb5-b555f2f7f8b1-kube-api-access-sqkml\") pod \"cluster-samples-operator-665b6dd947-6w2pm\" (UID: \"e925f407-5010-4a6f-bcb5-b555f2f7f8b1\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6w2pm" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.659202 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/f0062618-a397-45c0-a4f5-df3ebe244228-encryption-config\") pod \"apiserver-76f77b778f-hg5rx\" (UID: \"f0062618-a397-45c0-a4f5-df3ebe244228\") " pod="openshift-apiserver/apiserver-76f77b778f-hg5rx" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.659241 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/3a80ea92-4f96-4916-a00d-3f010959194d-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-8d6k4\" (UID: \"3a80ea92-4f96-4916-a00d-3f010959194d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8d6k4" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.659582 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hfdzk\" (UniqueName: \"kubernetes.io/projected/243fb6e4-1bf5-4c4a-b471-87f50d5e147d-kube-api-access-hfdzk\") pod \"openshift-config-operator-7777fb866f-gg9lz\" (UID: \"243fb6e4-1bf5-4c4a-b471-87f50d5e147d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-gg9lz" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.659623 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3a80ea92-4f96-4916-a00d-3f010959194d-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-8d6k4\" (UID: \"3a80ea92-4f96-4916-a00d-3f010959194d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8d6k4" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.659644 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f0062618-a397-45c0-a4f5-df3ebe244228-config\") pod \"apiserver-76f77b778f-hg5rx\" (UID: \"f0062618-a397-45c0-a4f5-df3ebe244228\") " pod="openshift-apiserver/apiserver-76f77b778f-hg5rx" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.659921 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c57e08c1-826c-4332-8a9e-50f419231d73-config\") pod \"console-operator-58897d9998-fxhqt\" (UID: \"c57e08c1-826c-4332-8a9e-50f419231d73\") " pod="openshift-console-operator/console-operator-58897d9998-fxhqt" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.660113 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f0062618-a397-45c0-a4f5-df3ebe244228-etcd-client\") pod \"apiserver-76f77b778f-hg5rx\" (UID: \"f0062618-a397-45c0-a4f5-df3ebe244228\") " pod="openshift-apiserver/apiserver-76f77b778f-hg5rx" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.660134 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-qgdp5"] Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.660116 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/45e90d9b-8520-4b1f-8f16-170787f4c8b6-serving-cert\") pod \"route-controller-manager-6576b87f9c-g4vj4\" (UID: \"45e90d9b-8520-4b1f-8f16-170787f4c8b6\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-g4vj4" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.660190 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/b6e3c195-a333-4c4f-8b75-71edbfcc26c3-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-mc59z\" (UID: \"b6e3c195-a333-4c4f-8b75-71edbfcc26c3\") " pod="openshift-authentication/oauth-openshift-558db77b4-mc59z" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.660246 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/b6e3c195-a333-4c4f-8b75-71edbfcc26c3-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-mc59z\" (UID: \"b6e3c195-a333-4c4f-8b75-71edbfcc26c3\") " pod="openshift-authentication/oauth-openshift-558db77b4-mc59z" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.660308 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/f0062618-a397-45c0-a4f5-df3ebe244228-audit\") pod \"apiserver-76f77b778f-hg5rx\" (UID: \"f0062618-a397-45c0-a4f5-df3ebe244228\") " pod="openshift-apiserver/apiserver-76f77b778f-hg5rx" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.660345 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f0062618-a397-45c0-a4f5-df3ebe244228-config\") pod \"apiserver-76f77b778f-hg5rx\" (UID: \"f0062618-a397-45c0-a4f5-df3ebe244228\") " pod="openshift-apiserver/apiserver-76f77b778f-hg5rx" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.660490 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3a80ea92-4f96-4916-a00d-3f010959194d-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-8d6k4\" (UID: \"3a80ea92-4f96-4916-a00d-3f010959194d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8d6k4" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.660535 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/4a781c56-3979-49df-a7f1-340c44bcc9b5-profile-collector-cert\") pod \"olm-operator-6b444d44fb-cm9ql\" (UID: \"4a781c56-3979-49df-a7f1-340c44bcc9b5\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cm9ql" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.660766 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/b6e3c195-a333-4c4f-8b75-71edbfcc26c3-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-mc59z\" (UID: \"b6e3c195-a333-4c4f-8b75-71edbfcc26c3\") " pod="openshift-authentication/oauth-openshift-558db77b4-mc59z" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.660818 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-xft2d"] Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.660954 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/f0062618-a397-45c0-a4f5-df3ebe244228-audit\") pod \"apiserver-76f77b778f-hg5rx\" (UID: \"f0062618-a397-45c0-a4f5-df3ebe244228\") " pod="openshift-apiserver/apiserver-76f77b778f-hg5rx" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.661100 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5feb1fb5-73be-4ce5-a6d5-d426e2eada2e-config\") pod \"machine-approver-56656f9798-smtqm\" (UID: \"5feb1fb5-73be-4ce5-a6d5-d426e2eada2e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-smtqm" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.661107 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/f257ddf8-bc9c-4670-b8c3-943829b80433-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-vbm65\" (UID: \"f257ddf8-bc9c-4670-b8c3-943829b80433\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-vbm65" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.661152 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7lhrd\" (UniqueName: \"kubernetes.io/projected/f257ddf8-bc9c-4670-b8c3-943829b80433-kube-api-access-7lhrd\") pod \"multus-admission-controller-857f4d67dd-vbm65\" (UID: \"f257ddf8-bc9c-4670-b8c3-943829b80433\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-vbm65" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.661178 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/45e90d9b-8520-4b1f-8f16-170787f4c8b6-config\") pod \"route-controller-manager-6576b87f9c-g4vj4\" (UID: \"45e90d9b-8520-4b1f-8f16-170787f4c8b6\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-g4vj4" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.661193 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/b6e3c195-a333-4c4f-8b75-71edbfcc26c3-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-mc59z\" (UID: \"b6e3c195-a333-4c4f-8b75-71edbfcc26c3\") " pod="openshift-authentication/oauth-openshift-558db77b4-mc59z" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.661214 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/cc6bb484-0230-46d9-9507-b641d2a5f330-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-dpqkj\" (UID: \"cc6bb484-0230-46d9-9507-b641d2a5f330\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dpqkj" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.661235 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3de23ab3-e037-4163-b744-81ecbb340b13-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-5vkgw\" (UID: \"3de23ab3-e037-4163-b744-81ecbb340b13\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5vkgw" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.661251 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/88149eca-e1e7-4dd4-8cdf-170db9460b30-audit-policies\") pod \"apiserver-7bbb656c7d-hkwl9\" (UID: \"88149eca-e1e7-4dd4-8cdf-170db9460b30\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hkwl9" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.661305 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/2c17c274-7853-4717-a68e-9fa6467ee199-auth-proxy-config\") pod \"machine-config-operator-74547568cd-8nb8p\" (UID: \"2c17c274-7853-4717-a68e-9fa6467ee199\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8nb8p" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.661334 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/075528cd-eab2-47bb-8aa0-3d39002fb3d1-client-ca\") pod \"controller-manager-879f6c89f-vs6d8\" (UID: \"075528cd-eab2-47bb-8aa0-3d39002fb3d1\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vs6d8" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.661352 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b6e3c195-a333-4c4f-8b75-71edbfcc26c3-audit-dir\") pod \"oauth-openshift-558db77b4-mc59z\" (UID: \"b6e3c195-a333-4c4f-8b75-71edbfcc26c3\") " pod="openshift-authentication/oauth-openshift-558db77b4-mc59z" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.661370 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c3a7de1e-09f7-41eb-8bf2-3737eee4ae74-service-ca-bundle\") pod \"authentication-operator-69f744f599-8j5cm\" (UID: \"c3a7de1e-09f7-41eb-8bf2-3737eee4ae74\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8j5cm" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.661388 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c57e08c1-826c-4332-8a9e-50f419231d73-serving-cert\") pod \"console-operator-58897d9998-fxhqt\" (UID: \"c57e08c1-826c-4332-8a9e-50f419231d73\") " pod="openshift-console-operator/console-operator-58897d9998-fxhqt" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.661403 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c57e08c1-826c-4332-8a9e-50f419231d73-trusted-ca\") pod \"console-operator-58897d9998-fxhqt\" (UID: \"c57e08c1-826c-4332-8a9e-50f419231d73\") " pod="openshift-console-operator/console-operator-58897d9998-fxhqt" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.661419 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/075528cd-eab2-47bb-8aa0-3d39002fb3d1-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-vs6d8\" (UID: \"075528cd-eab2-47bb-8aa0-3d39002fb3d1\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vs6d8" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.661439 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2qr2q\" (UniqueName: \"kubernetes.io/projected/fd852137-cff1-4fae-a921-42be5c511db8-kube-api-access-2qr2q\") pod \"downloads-7954f5f757-5kn8g\" (UID: \"fd852137-cff1-4fae-a921-42be5c511db8\") " pod="openshift-console/downloads-7954f5f757-5kn8g" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.661456 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5feb1fb5-73be-4ce5-a6d5-d426e2eada2e-auth-proxy-config\") pod \"machine-approver-56656f9798-smtqm\" (UID: \"5feb1fb5-73be-4ce5-a6d5-d426e2eada2e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-smtqm" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.661473 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-shk8x\" (UniqueName: \"kubernetes.io/projected/5feb1fb5-73be-4ce5-a6d5-d426e2eada2e-kube-api-access-shk8x\") pod \"machine-approver-56656f9798-smtqm\" (UID: \"5feb1fb5-73be-4ce5-a6d5-d426e2eada2e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-smtqm" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.661501 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/2c17c274-7853-4717-a68e-9fa6467ee199-images\") pod \"machine-config-operator-74547568cd-8nb8p\" (UID: \"2c17c274-7853-4717-a68e-9fa6467ee199\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8nb8p" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.661584 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b6e3c195-a333-4c4f-8b75-71edbfcc26c3-audit-dir\") pod \"oauth-openshift-558db77b4-mc59z\" (UID: \"b6e3c195-a333-4c4f-8b75-71edbfcc26c3\") " pod="openshift-authentication/oauth-openshift-558db77b4-mc59z" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.662005 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/075528cd-eab2-47bb-8aa0-3d39002fb3d1-client-ca\") pod \"controller-manager-879f6c89f-vs6d8\" (UID: \"075528cd-eab2-47bb-8aa0-3d39002fb3d1\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vs6d8" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.662235 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/45e90d9b-8520-4b1f-8f16-170787f4c8b6-config\") pod \"route-controller-manager-6576b87f9c-g4vj4\" (UID: \"45e90d9b-8520-4b1f-8f16-170787f4c8b6\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-g4vj4" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.662754 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5feb1fb5-73be-4ce5-a6d5-d426e2eada2e-auth-proxy-config\") pod \"machine-approver-56656f9798-smtqm\" (UID: \"5feb1fb5-73be-4ce5-a6d5-d426e2eada2e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-smtqm" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.662945 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c57e08c1-826c-4332-8a9e-50f419231d73-trusted-ca\") pod \"console-operator-58897d9998-fxhqt\" (UID: \"c57e08c1-826c-4332-8a9e-50f419231d73\") " pod="openshift-console-operator/console-operator-58897d9998-fxhqt" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.663004 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/b6e3c195-a333-4c4f-8b75-71edbfcc26c3-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-mc59z\" (UID: \"b6e3c195-a333-4c4f-8b75-71edbfcc26c3\") " pod="openshift-authentication/oauth-openshift-558db77b4-mc59z" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.663047 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-7nhrb"] Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.663221 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-6bfsg"] Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.663568 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/b6e3c195-a333-4c4f-8b75-71edbfcc26c3-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-mc59z\" (UID: \"b6e3c195-a333-4c4f-8b75-71edbfcc26c3\") " pod="openshift-authentication/oauth-openshift-558db77b4-mc59z" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.663694 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/b6e3c195-a333-4c4f-8b75-71edbfcc26c3-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-mc59z\" (UID: \"b6e3c195-a333-4c4f-8b75-71edbfcc26c3\") " pod="openshift-authentication/oauth-openshift-558db77b4-mc59z" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.664383 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c57e08c1-826c-4332-8a9e-50f419231d73-serving-cert\") pod \"console-operator-58897d9998-fxhqt\" (UID: \"c57e08c1-826c-4332-8a9e-50f419231d73\") " pod="openshift-console-operator/console-operator-58897d9998-fxhqt" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.664670 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/075528cd-eab2-47bb-8aa0-3d39002fb3d1-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-vs6d8\" (UID: \"075528cd-eab2-47bb-8aa0-3d39002fb3d1\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vs6d8" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.664679 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/b6e3c195-a333-4c4f-8b75-71edbfcc26c3-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-mc59z\" (UID: \"b6e3c195-a333-4c4f-8b75-71edbfcc26c3\") " pod="openshift-authentication/oauth-openshift-558db77b4-mc59z" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.664808 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-kgxn6"] Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.665311 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/b6e3c195-a333-4c4f-8b75-71edbfcc26c3-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-mc59z\" (UID: \"b6e3c195-a333-4c4f-8b75-71edbfcc26c3\") " pod="openshift-authentication/oauth-openshift-558db77b4-mc59z" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.665963 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-zqxvm"] Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.666942 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-5lbxn"] Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.667939 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322825-b7zrt"] Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.668721 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f0062618-a397-45c0-a4f5-df3ebe244228-serving-cert\") pod \"apiserver-76f77b778f-hg5rx\" (UID: \"f0062618-a397-45c0-a4f5-df3ebe244228\") " pod="openshift-apiserver/apiserver-76f77b778f-hg5rx" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.668900 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-c94sm"] Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.669514 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/243fb6e4-1bf5-4c4a-b471-87f50d5e147d-available-featuregates\") pod \"openshift-config-operator-7777fb866f-gg9lz\" (UID: \"243fb6e4-1bf5-4c4a-b471-87f50d5e147d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-gg9lz" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.669899 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-k8r8q"] Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.669997 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/5feb1fb5-73be-4ce5-a6d5-d426e2eada2e-machine-approver-tls\") pod \"machine-approver-56656f9798-smtqm\" (UID: \"5feb1fb5-73be-4ce5-a6d5-d426e2eada2e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-smtqm" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.670942 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-7hk2v"] Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.671926 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-n7bs2"] Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.672903 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-qztpv"] Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.673457 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-qztpv" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.676272 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.710906 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.738129 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.751616 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.762032 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/e925f407-5010-4a6f-bcb5-b555f2f7f8b1-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-6w2pm\" (UID: \"e925f407-5010-4a6f-bcb5-b555f2f7f8b1\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6w2pm" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.762059 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zh7sn\" (UniqueName: \"kubernetes.io/projected/88149eca-e1e7-4dd4-8cdf-170db9460b30-kube-api-access-zh7sn\") pod \"apiserver-7bbb656c7d-hkwl9\" (UID: \"88149eca-e1e7-4dd4-8cdf-170db9460b30\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hkwl9" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.762076 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/028a80a9-9747-4856-a2e8-f721abd82683-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-zqxvm\" (UID: \"028a80a9-9747-4856-a2e8-f721abd82683\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-zqxvm" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.762104 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pm4kq\" (UniqueName: \"kubernetes.io/projected/472cd80c-0597-40b4-babb-89a3fce346b0-kube-api-access-pm4kq\") pod \"dns-operator-744455d44c-swrb9\" (UID: \"472cd80c-0597-40b4-babb-89a3fce346b0\") " pod="openshift-dns-operator/dns-operator-744455d44c-swrb9" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.762119 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/88149eca-e1e7-4dd4-8cdf-170db9460b30-audit-dir\") pod \"apiserver-7bbb656c7d-hkwl9\" (UID: \"88149eca-e1e7-4dd4-8cdf-170db9460b30\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hkwl9" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.762135 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tmkv7\" (UniqueName: \"kubernetes.io/projected/4a781c56-3979-49df-a7f1-340c44bcc9b5-kube-api-access-tmkv7\") pod \"olm-operator-6b444d44fb-cm9ql\" (UID: \"4a781c56-3979-49df-a7f1-340c44bcc9b5\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cm9ql" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.762150 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c3a7de1e-09f7-41eb-8bf2-3737eee4ae74-config\") pod \"authentication-operator-69f744f599-8j5cm\" (UID: \"c3a7de1e-09f7-41eb-8bf2-3737eee4ae74\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8j5cm" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.762168 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c3a7de1e-09f7-41eb-8bf2-3737eee4ae74-serving-cert\") pod \"authentication-operator-69f744f599-8j5cm\" (UID: \"c3a7de1e-09f7-41eb-8bf2-3737eee4ae74\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8j5cm" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.762184 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/2c17c274-7853-4717-a68e-9fa6467ee199-proxy-tls\") pod \"machine-config-operator-74547568cd-8nb8p\" (UID: \"2c17c274-7853-4717-a68e-9fa6467ee199\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8nb8p" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.762210 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/88149eca-e1e7-4dd4-8cdf-170db9460b30-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-hkwl9\" (UID: \"88149eca-e1e7-4dd4-8cdf-170db9460b30\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hkwl9" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.762226 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/88149eca-e1e7-4dd4-8cdf-170db9460b30-encryption-config\") pod \"apiserver-7bbb656c7d-hkwl9\" (UID: \"88149eca-e1e7-4dd4-8cdf-170db9460b30\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hkwl9" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.762243 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3de23ab3-e037-4163-b744-81ecbb340b13-config\") pod \"openshift-apiserver-operator-796bbdcf4f-5vkgw\" (UID: \"3de23ab3-e037-4163-b744-81ecbb340b13\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5vkgw" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.762318 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/472cd80c-0597-40b4-babb-89a3fce346b0-metrics-tls\") pod \"dns-operator-744455d44c-swrb9\" (UID: \"472cd80c-0597-40b4-babb-89a3fce346b0\") " pod="openshift-dns-operator/dns-operator-744455d44c-swrb9" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.762336 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/028a80a9-9747-4856-a2e8-f721abd82683-config\") pod \"kube-controller-manager-operator-78b949d7b-zqxvm\" (UID: \"028a80a9-9747-4856-a2e8-f721abd82683\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-zqxvm" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.762360 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/88149eca-e1e7-4dd4-8cdf-170db9460b30-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-hkwl9\" (UID: \"88149eca-e1e7-4dd4-8cdf-170db9460b30\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hkwl9" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.762387 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/88149eca-e1e7-4dd4-8cdf-170db9460b30-audit-dir\") pod \"apiserver-7bbb656c7d-hkwl9\" (UID: \"88149eca-e1e7-4dd4-8cdf-170db9460b30\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hkwl9" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.762505 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/88149eca-e1e7-4dd4-8cdf-170db9460b30-serving-cert\") pod \"apiserver-7bbb656c7d-hkwl9\" (UID: \"88149eca-e1e7-4dd4-8cdf-170db9460b30\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hkwl9" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.762527 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/76fb6dce-01dd-4aef-8213-e03be7a5e055-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-5wldz\" (UID: \"76fb6dce-01dd-4aef-8213-e03be7a5e055\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5wldz" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.762573 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d8p2l\" (UniqueName: \"kubernetes.io/projected/c3a7de1e-09f7-41eb-8bf2-3737eee4ae74-kube-api-access-d8p2l\") pod \"authentication-operator-69f744f599-8j5cm\" (UID: \"c3a7de1e-09f7-41eb-8bf2-3737eee4ae74\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8j5cm" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.763286 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/88149eca-e1e7-4dd4-8cdf-170db9460b30-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-hkwl9\" (UID: \"88149eca-e1e7-4dd4-8cdf-170db9460b30\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hkwl9" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.762590 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/88149eca-e1e7-4dd4-8cdf-170db9460b30-etcd-client\") pod \"apiserver-7bbb656c7d-hkwl9\" (UID: \"88149eca-e1e7-4dd4-8cdf-170db9460b30\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hkwl9" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.763363 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/028a80a9-9747-4856-a2e8-f721abd82683-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-zqxvm\" (UID: \"028a80a9-9747-4856-a2e8-f721abd82683\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-zqxvm" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.763383 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rkk59\" (UniqueName: \"kubernetes.io/projected/cc6bb484-0230-46d9-9507-b641d2a5f330-kube-api-access-rkk59\") pod \"control-plane-machine-set-operator-78cbb6b69f-dpqkj\" (UID: \"cc6bb484-0230-46d9-9507-b641d2a5f330\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dpqkj" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.763409 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h4fbp\" (UniqueName: \"kubernetes.io/projected/2c17c274-7853-4717-a68e-9fa6467ee199-kube-api-access-h4fbp\") pod \"machine-config-operator-74547568cd-8nb8p\" (UID: \"2c17c274-7853-4717-a68e-9fa6467ee199\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8nb8p" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.763424 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/76fb6dce-01dd-4aef-8213-e03be7a5e055-config\") pod \"machine-api-operator-5694c8668f-5wldz\" (UID: \"76fb6dce-01dd-4aef-8213-e03be7a5e055\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5wldz" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.763448 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8qtsm\" (UniqueName: \"kubernetes.io/projected/76fb6dce-01dd-4aef-8213-e03be7a5e055-kube-api-access-8qtsm\") pod \"machine-api-operator-5694c8668f-5wldz\" (UID: \"76fb6dce-01dd-4aef-8213-e03be7a5e055\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5wldz" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.763473 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sqkml\" (UniqueName: \"kubernetes.io/projected/e925f407-5010-4a6f-bcb5-b555f2f7f8b1-kube-api-access-sqkml\") pod \"cluster-samples-operator-665b6dd947-6w2pm\" (UID: \"e925f407-5010-4a6f-bcb5-b555f2f7f8b1\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6w2pm" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.763505 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/4a781c56-3979-49df-a7f1-340c44bcc9b5-profile-collector-cert\") pod \"olm-operator-6b444d44fb-cm9ql\" (UID: \"4a781c56-3979-49df-a7f1-340c44bcc9b5\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cm9ql" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.763528 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/f257ddf8-bc9c-4670-b8c3-943829b80433-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-vbm65\" (UID: \"f257ddf8-bc9c-4670-b8c3-943829b80433\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-vbm65" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.763553 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7lhrd\" (UniqueName: \"kubernetes.io/projected/f257ddf8-bc9c-4670-b8c3-943829b80433-kube-api-access-7lhrd\") pod \"multus-admission-controller-857f4d67dd-vbm65\" (UID: \"f257ddf8-bc9c-4670-b8c3-943829b80433\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-vbm65" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.763554 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/88149eca-e1e7-4dd4-8cdf-170db9460b30-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-hkwl9\" (UID: \"88149eca-e1e7-4dd4-8cdf-170db9460b30\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hkwl9" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.763572 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/cc6bb484-0230-46d9-9507-b641d2a5f330-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-dpqkj\" (UID: \"cc6bb484-0230-46d9-9507-b641d2a5f330\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dpqkj" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.763591 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3de23ab3-e037-4163-b744-81ecbb340b13-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-5vkgw\" (UID: \"3de23ab3-e037-4163-b744-81ecbb340b13\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5vkgw" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.763607 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/88149eca-e1e7-4dd4-8cdf-170db9460b30-audit-policies\") pod \"apiserver-7bbb656c7d-hkwl9\" (UID: \"88149eca-e1e7-4dd4-8cdf-170db9460b30\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hkwl9" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.763624 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/2c17c274-7853-4717-a68e-9fa6467ee199-auth-proxy-config\") pod \"machine-config-operator-74547568cd-8nb8p\" (UID: \"2c17c274-7853-4717-a68e-9fa6467ee199\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8nb8p" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.763641 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c3a7de1e-09f7-41eb-8bf2-3737eee4ae74-service-ca-bundle\") pod \"authentication-operator-69f744f599-8j5cm\" (UID: \"c3a7de1e-09f7-41eb-8bf2-3737eee4ae74\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8j5cm" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.763670 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/2c17c274-7853-4717-a68e-9fa6467ee199-images\") pod \"machine-config-operator-74547568cd-8nb8p\" (UID: \"2c17c274-7853-4717-a68e-9fa6467ee199\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8nb8p" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.763687 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6vhnl\" (UniqueName: \"kubernetes.io/projected/3de23ab3-e037-4163-b744-81ecbb340b13-kube-api-access-6vhnl\") pod \"openshift-apiserver-operator-796bbdcf4f-5vkgw\" (UID: \"3de23ab3-e037-4163-b744-81ecbb340b13\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5vkgw" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.763700 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/76fb6dce-01dd-4aef-8213-e03be7a5e055-images\") pod \"machine-api-operator-5694c8668f-5wldz\" (UID: \"76fb6dce-01dd-4aef-8213-e03be7a5e055\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5wldz" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.763717 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c3a7de1e-09f7-41eb-8bf2-3737eee4ae74-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-8j5cm\" (UID: \"c3a7de1e-09f7-41eb-8bf2-3737eee4ae74\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8j5cm" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.763731 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/4a781c56-3979-49df-a7f1-340c44bcc9b5-srv-cert\") pod \"olm-operator-6b444d44fb-cm9ql\" (UID: \"4a781c56-3979-49df-a7f1-340c44bcc9b5\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cm9ql" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.763516 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c3a7de1e-09f7-41eb-8bf2-3737eee4ae74-config\") pod \"authentication-operator-69f744f599-8j5cm\" (UID: \"c3a7de1e-09f7-41eb-8bf2-3737eee4ae74\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8j5cm" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.763900 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3de23ab3-e037-4163-b744-81ecbb340b13-config\") pod \"openshift-apiserver-operator-796bbdcf4f-5vkgw\" (UID: \"3de23ab3-e037-4163-b744-81ecbb340b13\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5vkgw" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.764749 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/76fb6dce-01dd-4aef-8213-e03be7a5e055-config\") pod \"machine-api-operator-5694c8668f-5wldz\" (UID: \"76fb6dce-01dd-4aef-8213-e03be7a5e055\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5wldz" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.765069 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/88149eca-e1e7-4dd4-8cdf-170db9460b30-audit-policies\") pod \"apiserver-7bbb656c7d-hkwl9\" (UID: \"88149eca-e1e7-4dd4-8cdf-170db9460b30\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hkwl9" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.765500 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/e925f407-5010-4a6f-bcb5-b555f2f7f8b1-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-6w2pm\" (UID: \"e925f407-5010-4a6f-bcb5-b555f2f7f8b1\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6w2pm" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.766291 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c3a7de1e-09f7-41eb-8bf2-3737eee4ae74-service-ca-bundle\") pod \"authentication-operator-69f744f599-8j5cm\" (UID: \"c3a7de1e-09f7-41eb-8bf2-3737eee4ae74\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8j5cm" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.766483 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/2c17c274-7853-4717-a68e-9fa6467ee199-auth-proxy-config\") pod \"machine-config-operator-74547568cd-8nb8p\" (UID: \"2c17c274-7853-4717-a68e-9fa6467ee199\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8nb8p" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.766721 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/88149eca-e1e7-4dd4-8cdf-170db9460b30-etcd-client\") pod \"apiserver-7bbb656c7d-hkwl9\" (UID: \"88149eca-e1e7-4dd4-8cdf-170db9460b30\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hkwl9" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.767064 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c3a7de1e-09f7-41eb-8bf2-3737eee4ae74-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-8j5cm\" (UID: \"c3a7de1e-09f7-41eb-8bf2-3737eee4ae74\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8j5cm" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.767183 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/76fb6dce-01dd-4aef-8213-e03be7a5e055-images\") pod \"machine-api-operator-5694c8668f-5wldz\" (UID: \"76fb6dce-01dd-4aef-8213-e03be7a5e055\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5wldz" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.767339 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/88149eca-e1e7-4dd4-8cdf-170db9460b30-serving-cert\") pod \"apiserver-7bbb656c7d-hkwl9\" (UID: \"88149eca-e1e7-4dd4-8cdf-170db9460b30\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hkwl9" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.767641 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3de23ab3-e037-4163-b744-81ecbb340b13-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-5vkgw\" (UID: \"3de23ab3-e037-4163-b744-81ecbb340b13\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5vkgw" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.768649 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c3a7de1e-09f7-41eb-8bf2-3737eee4ae74-serving-cert\") pod \"authentication-operator-69f744f599-8j5cm\" (UID: \"c3a7de1e-09f7-41eb-8bf2-3737eee4ae74\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8j5cm" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.768767 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/472cd80c-0597-40b4-babb-89a3fce346b0-metrics-tls\") pod \"dns-operator-744455d44c-swrb9\" (UID: \"472cd80c-0597-40b4-babb-89a3fce346b0\") " pod="openshift-dns-operator/dns-operator-744455d44c-swrb9" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.769737 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/88149eca-e1e7-4dd4-8cdf-170db9460b30-encryption-config\") pod \"apiserver-7bbb656c7d-hkwl9\" (UID: \"88149eca-e1e7-4dd4-8cdf-170db9460b30\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hkwl9" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.770994 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.781753 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/76fb6dce-01dd-4aef-8213-e03be7a5e055-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-5wldz\" (UID: \"76fb6dce-01dd-4aef-8213-e03be7a5e055\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5wldz" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.790891 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.798954 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/cc6bb484-0230-46d9-9507-b641d2a5f330-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-dpqkj\" (UID: \"cc6bb484-0230-46d9-9507-b641d2a5f330\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dpqkj" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.811208 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.831297 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.850973 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.871251 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.891546 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.910766 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.932388 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.951079 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.981007 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Oct 02 01:49:11 crc kubenswrapper[4885]: I1002 01:49:11.992749 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Oct 02 01:49:12 crc kubenswrapper[4885]: I1002 01:49:12.010991 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Oct 02 01:49:12 crc kubenswrapper[4885]: I1002 01:49:12.031638 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Oct 02 01:49:12 crc kubenswrapper[4885]: I1002 01:49:12.051694 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Oct 02 01:49:12 crc kubenswrapper[4885]: I1002 01:49:12.061620 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/028a80a9-9747-4856-a2e8-f721abd82683-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-zqxvm\" (UID: \"028a80a9-9747-4856-a2e8-f721abd82683\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-zqxvm" Oct 02 01:49:12 crc kubenswrapper[4885]: I1002 01:49:12.071989 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Oct 02 01:49:12 crc kubenswrapper[4885]: I1002 01:49:12.091838 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Oct 02 01:49:12 crc kubenswrapper[4885]: I1002 01:49:12.094332 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/028a80a9-9747-4856-a2e8-f721abd82683-config\") pod \"kube-controller-manager-operator-78b949d7b-zqxvm\" (UID: \"028a80a9-9747-4856-a2e8-f721abd82683\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-zqxvm" Oct 02 01:49:12 crc kubenswrapper[4885]: I1002 01:49:12.111393 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Oct 02 01:49:12 crc kubenswrapper[4885]: I1002 01:49:12.131192 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Oct 02 01:49:12 crc kubenswrapper[4885]: I1002 01:49:12.152193 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Oct 02 01:49:12 crc kubenswrapper[4885]: I1002 01:49:12.171382 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Oct 02 01:49:12 crc kubenswrapper[4885]: I1002 01:49:12.191447 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Oct 02 01:49:12 crc kubenswrapper[4885]: I1002 01:49:12.199246 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/f257ddf8-bc9c-4670-b8c3-943829b80433-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-vbm65\" (UID: \"f257ddf8-bc9c-4670-b8c3-943829b80433\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-vbm65" Oct 02 01:49:12 crc kubenswrapper[4885]: I1002 01:49:12.212122 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Oct 02 01:49:12 crc kubenswrapper[4885]: I1002 01:49:12.231902 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Oct 02 01:49:12 crc kubenswrapper[4885]: I1002 01:49:12.251648 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Oct 02 01:49:12 crc kubenswrapper[4885]: I1002 01:49:12.272091 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Oct 02 01:49:12 crc kubenswrapper[4885]: I1002 01:49:12.291357 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Oct 02 01:49:12 crc kubenswrapper[4885]: I1002 01:49:12.312756 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Oct 02 01:49:12 crc kubenswrapper[4885]: I1002 01:49:12.332022 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Oct 02 01:49:12 crc kubenswrapper[4885]: I1002 01:49:12.352191 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Oct 02 01:49:12 crc kubenswrapper[4885]: I1002 01:49:12.372386 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Oct 02 01:49:12 crc kubenswrapper[4885]: I1002 01:49:12.377168 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/2c17c274-7853-4717-a68e-9fa6467ee199-proxy-tls\") pod \"machine-config-operator-74547568cd-8nb8p\" (UID: \"2c17c274-7853-4717-a68e-9fa6467ee199\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8nb8p" Oct 02 01:49:12 crc kubenswrapper[4885]: I1002 01:49:12.391934 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Oct 02 01:49:12 crc kubenswrapper[4885]: I1002 01:49:12.395987 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/2c17c274-7853-4717-a68e-9fa6467ee199-images\") pod \"machine-config-operator-74547568cd-8nb8p\" (UID: \"2c17c274-7853-4717-a68e-9fa6467ee199\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8nb8p" Oct 02 01:49:12 crc kubenswrapper[4885]: I1002 01:49:12.412690 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Oct 02 01:49:12 crc kubenswrapper[4885]: I1002 01:49:12.432420 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Oct 02 01:49:12 crc kubenswrapper[4885]: I1002 01:49:12.452437 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Oct 02 01:49:12 crc kubenswrapper[4885]: I1002 01:49:12.471686 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Oct 02 01:49:12 crc kubenswrapper[4885]: I1002 01:49:12.492119 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Oct 02 01:49:12 crc kubenswrapper[4885]: I1002 01:49:12.499665 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/4a781c56-3979-49df-a7f1-340c44bcc9b5-profile-collector-cert\") pod \"olm-operator-6b444d44fb-cm9ql\" (UID: \"4a781c56-3979-49df-a7f1-340c44bcc9b5\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cm9ql" Oct 02 01:49:12 crc kubenswrapper[4885]: I1002 01:49:12.513135 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Oct 02 01:49:12 crc kubenswrapper[4885]: I1002 01:49:12.532541 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Oct 02 01:49:12 crc kubenswrapper[4885]: I1002 01:49:12.551737 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Oct 02 01:49:12 crc kubenswrapper[4885]: I1002 01:49:12.570289 4885 request.go:700] Waited for 1.005653538s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-storage-version-migrator-operator/secrets?fieldSelector=metadata.name%3Dkube-storage-version-migrator-operator-dockercfg-2bh8d&limit=500&resourceVersion=0 Oct 02 01:49:12 crc kubenswrapper[4885]: I1002 01:49:12.572670 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Oct 02 01:49:12 crc kubenswrapper[4885]: I1002 01:49:12.593157 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Oct 02 01:49:12 crc kubenswrapper[4885]: I1002 01:49:12.612191 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Oct 02 01:49:12 crc kubenswrapper[4885]: I1002 01:49:12.631882 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Oct 02 01:49:12 crc kubenswrapper[4885]: I1002 01:49:12.651994 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Oct 02 01:49:12 crc kubenswrapper[4885]: I1002 01:49:12.660239 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/4a781c56-3979-49df-a7f1-340c44bcc9b5-srv-cert\") pod \"olm-operator-6b444d44fb-cm9ql\" (UID: \"4a781c56-3979-49df-a7f1-340c44bcc9b5\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cm9ql" Oct 02 01:49:12 crc kubenswrapper[4885]: I1002 01:49:12.672306 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Oct 02 01:49:12 crc kubenswrapper[4885]: I1002 01:49:12.692152 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Oct 02 01:49:12 crc kubenswrapper[4885]: I1002 01:49:12.711649 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Oct 02 01:49:12 crc kubenswrapper[4885]: I1002 01:49:12.744083 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Oct 02 01:49:12 crc kubenswrapper[4885]: I1002 01:49:12.751510 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Oct 02 01:49:12 crc kubenswrapper[4885]: I1002 01:49:12.772345 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Oct 02 01:49:12 crc kubenswrapper[4885]: I1002 01:49:12.792851 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Oct 02 01:49:12 crc kubenswrapper[4885]: I1002 01:49:12.812042 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Oct 02 01:49:12 crc kubenswrapper[4885]: I1002 01:49:12.851961 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Oct 02 01:49:12 crc kubenswrapper[4885]: I1002 01:49:12.871589 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Oct 02 01:49:12 crc kubenswrapper[4885]: I1002 01:49:12.892294 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Oct 02 01:49:12 crc kubenswrapper[4885]: I1002 01:49:12.911205 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Oct 02 01:49:12 crc kubenswrapper[4885]: I1002 01:49:12.931443 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Oct 02 01:49:12 crc kubenswrapper[4885]: I1002 01:49:12.952062 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Oct 02 01:49:12 crc kubenswrapper[4885]: I1002 01:49:12.973313 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Oct 02 01:49:12 crc kubenswrapper[4885]: I1002 01:49:12.991712 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 02 01:49:13 crc kubenswrapper[4885]: I1002 01:49:13.012238 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 02 01:49:13 crc kubenswrapper[4885]: I1002 01:49:13.032994 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Oct 02 01:49:13 crc kubenswrapper[4885]: I1002 01:49:13.051249 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Oct 02 01:49:13 crc kubenswrapper[4885]: I1002 01:49:13.071628 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Oct 02 01:49:13 crc kubenswrapper[4885]: I1002 01:49:13.094184 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Oct 02 01:49:13 crc kubenswrapper[4885]: I1002 01:49:13.107112 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" Oct 02 01:49:13 crc kubenswrapper[4885]: I1002 01:49:13.111703 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Oct 02 01:49:13 crc kubenswrapper[4885]: I1002 01:49:13.132155 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Oct 02 01:49:13 crc kubenswrapper[4885]: I1002 01:49:13.151905 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Oct 02 01:49:13 crc kubenswrapper[4885]: I1002 01:49:13.172901 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Oct 02 01:49:13 crc kubenswrapper[4885]: I1002 01:49:13.191745 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Oct 02 01:49:13 crc kubenswrapper[4885]: I1002 01:49:13.212298 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Oct 02 01:49:13 crc kubenswrapper[4885]: I1002 01:49:13.231940 4885 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Oct 02 01:49:13 crc kubenswrapper[4885]: I1002 01:49:13.251772 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Oct 02 01:49:13 crc kubenswrapper[4885]: I1002 01:49:13.271643 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Oct 02 01:49:13 crc kubenswrapper[4885]: I1002 01:49:13.292115 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Oct 02 01:49:13 crc kubenswrapper[4885]: I1002 01:49:13.311577 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Oct 02 01:49:13 crc kubenswrapper[4885]: I1002 01:49:13.332201 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Oct 02 01:49:13 crc kubenswrapper[4885]: I1002 01:49:13.378300 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3a80ea92-4f96-4916-a00d-3f010959194d-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-8d6k4\" (UID: \"3a80ea92-4f96-4916-a00d-3f010959194d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8d6k4" Oct 02 01:49:13 crc kubenswrapper[4885]: I1002 01:49:13.392416 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Oct 02 01:49:13 crc kubenswrapper[4885]: I1002 01:49:13.399251 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l5s4w\" (UniqueName: \"kubernetes.io/projected/b6e3c195-a333-4c4f-8b75-71edbfcc26c3-kube-api-access-l5s4w\") pod \"oauth-openshift-558db77b4-mc59z\" (UID: \"b6e3c195-a333-4c4f-8b75-71edbfcc26c3\") " pod="openshift-authentication/oauth-openshift-558db77b4-mc59z" Oct 02 01:49:13 crc kubenswrapper[4885]: I1002 01:49:13.411535 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Oct 02 01:49:13 crc kubenswrapper[4885]: I1002 01:49:13.431469 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Oct 02 01:49:13 crc kubenswrapper[4885]: I1002 01:49:13.480465 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rq2s2\" (UniqueName: \"kubernetes.io/projected/c57e08c1-826c-4332-8a9e-50f419231d73-kube-api-access-rq2s2\") pod \"console-operator-58897d9998-fxhqt\" (UID: \"c57e08c1-826c-4332-8a9e-50f419231d73\") " pod="openshift-console-operator/console-operator-58897d9998-fxhqt" Oct 02 01:49:13 crc kubenswrapper[4885]: I1002 01:49:13.499898 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vw9xq\" (UniqueName: \"kubernetes.io/projected/075528cd-eab2-47bb-8aa0-3d39002fb3d1-kube-api-access-vw9xq\") pod \"controller-manager-879f6c89f-vs6d8\" (UID: \"075528cd-eab2-47bb-8aa0-3d39002fb3d1\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vs6d8" Oct 02 01:49:13 crc kubenswrapper[4885]: I1002 01:49:13.519349 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-55ssq\" (UniqueName: \"kubernetes.io/projected/3a80ea92-4f96-4916-a00d-3f010959194d-kube-api-access-55ssq\") pod \"cluster-image-registry-operator-dc59b4c8b-8d6k4\" (UID: \"3a80ea92-4f96-4916-a00d-3f010959194d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8d6k4" Oct 02 01:49:13 crc kubenswrapper[4885]: I1002 01:49:13.529833 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mpmq4\" (UniqueName: \"kubernetes.io/projected/45e90d9b-8520-4b1f-8f16-170787f4c8b6-kube-api-access-mpmq4\") pod \"route-controller-manager-6576b87f9c-g4vj4\" (UID: \"45e90d9b-8520-4b1f-8f16-170787f4c8b6\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-g4vj4" Oct 02 01:49:13 crc kubenswrapper[4885]: I1002 01:49:13.529850 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-vs6d8" Oct 02 01:49:13 crc kubenswrapper[4885]: I1002 01:49:13.554135 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wfjz8\" (UniqueName: \"kubernetes.io/projected/f0062618-a397-45c0-a4f5-df3ebe244228-kube-api-access-wfjz8\") pod \"apiserver-76f77b778f-hg5rx\" (UID: \"f0062618-a397-45c0-a4f5-df3ebe244228\") " pod="openshift-apiserver/apiserver-76f77b778f-hg5rx" Oct 02 01:49:13 crc kubenswrapper[4885]: I1002 01:49:13.577658 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hfdzk\" (UniqueName: \"kubernetes.io/projected/243fb6e4-1bf5-4c4a-b471-87f50d5e147d-kube-api-access-hfdzk\") pod \"openshift-config-operator-7777fb866f-gg9lz\" (UID: \"243fb6e4-1bf5-4c4a-b471-87f50d5e147d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-gg9lz" Oct 02 01:49:13 crc kubenswrapper[4885]: I1002 01:49:13.582458 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-mc59z" Oct 02 01:49:13 crc kubenswrapper[4885]: I1002 01:49:13.590620 4885 request.go:700] Waited for 1.927515443s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-cluster-machine-approver/serviceaccounts/machine-approver-sa/token Oct 02 01:49:13 crc kubenswrapper[4885]: I1002 01:49:13.598845 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2qr2q\" (UniqueName: \"kubernetes.io/projected/fd852137-cff1-4fae-a921-42be5c511db8-kube-api-access-2qr2q\") pod \"downloads-7954f5f757-5kn8g\" (UID: \"fd852137-cff1-4fae-a921-42be5c511db8\") " pod="openshift-console/downloads-7954f5f757-5kn8g" Oct 02 01:49:13 crc kubenswrapper[4885]: I1002 01:49:13.605965 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-gg9lz" Oct 02 01:49:13 crc kubenswrapper[4885]: I1002 01:49:13.611477 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Oct 02 01:49:13 crc kubenswrapper[4885]: I1002 01:49:13.613915 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-shk8x\" (UniqueName: \"kubernetes.io/projected/5feb1fb5-73be-4ce5-a6d5-d426e2eada2e-kube-api-access-shk8x\") pod \"machine-approver-56656f9798-smtqm\" (UID: \"5feb1fb5-73be-4ce5-a6d5-d426e2eada2e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-smtqm" Oct 02 01:49:13 crc kubenswrapper[4885]: I1002 01:49:13.632208 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Oct 02 01:49:13 crc kubenswrapper[4885]: I1002 01:49:13.648088 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-g4vj4" Oct 02 01:49:13 crc kubenswrapper[4885]: I1002 01:49:13.652141 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Oct 02 01:49:13 crc kubenswrapper[4885]: I1002 01:49:13.662716 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8d6k4" Oct 02 01:49:13 crc kubenswrapper[4885]: I1002 01:49:13.681062 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-hg5rx" Oct 02 01:49:13 crc kubenswrapper[4885]: I1002 01:49:13.711490 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-5kn8g" Oct 02 01:49:13 crc kubenswrapper[4885]: I1002 01:49:13.719002 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-fxhqt" Oct 02 01:49:13 crc kubenswrapper[4885]: I1002 01:49:13.735153 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/028a80a9-9747-4856-a2e8-f721abd82683-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-zqxvm\" (UID: \"028a80a9-9747-4856-a2e8-f721abd82683\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-zqxvm" Oct 02 01:49:13 crc kubenswrapper[4885]: I1002 01:49:13.744218 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zh7sn\" (UniqueName: \"kubernetes.io/projected/88149eca-e1e7-4dd4-8cdf-170db9460b30-kube-api-access-zh7sn\") pod \"apiserver-7bbb656c7d-hkwl9\" (UID: \"88149eca-e1e7-4dd4-8cdf-170db9460b30\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hkwl9" Oct 02 01:49:13 crc kubenswrapper[4885]: I1002 01:49:13.754912 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tmkv7\" (UniqueName: \"kubernetes.io/projected/4a781c56-3979-49df-a7f1-340c44bcc9b5-kube-api-access-tmkv7\") pod \"olm-operator-6b444d44fb-cm9ql\" (UID: \"4a781c56-3979-49df-a7f1-340c44bcc9b5\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cm9ql" Oct 02 01:49:13 crc kubenswrapper[4885]: I1002 01:49:13.772325 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pm4kq\" (UniqueName: \"kubernetes.io/projected/472cd80c-0597-40b4-babb-89a3fce346b0-kube-api-access-pm4kq\") pod \"dns-operator-744455d44c-swrb9\" (UID: \"472cd80c-0597-40b4-babb-89a3fce346b0\") " pod="openshift-dns-operator/dns-operator-744455d44c-swrb9" Oct 02 01:49:13 crc kubenswrapper[4885]: I1002 01:49:13.788734 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d8p2l\" (UniqueName: \"kubernetes.io/projected/c3a7de1e-09f7-41eb-8bf2-3737eee4ae74-kube-api-access-d8p2l\") pod \"authentication-operator-69f744f599-8j5cm\" (UID: \"c3a7de1e-09f7-41eb-8bf2-3737eee4ae74\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8j5cm" Oct 02 01:49:13 crc kubenswrapper[4885]: I1002 01:49:13.812703 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-swrb9" Oct 02 01:49:13 crc kubenswrapper[4885]: I1002 01:49:13.818738 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8qtsm\" (UniqueName: \"kubernetes.io/projected/76fb6dce-01dd-4aef-8213-e03be7a5e055-kube-api-access-8qtsm\") pod \"machine-api-operator-5694c8668f-5wldz\" (UID: \"76fb6dce-01dd-4aef-8213-e03be7a5e055\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5wldz" Oct 02 01:49:13 crc kubenswrapper[4885]: I1002 01:49:13.822349 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-8j5cm" Oct 02 01:49:13 crc kubenswrapper[4885]: I1002 01:49:13.830687 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sqkml\" (UniqueName: \"kubernetes.io/projected/e925f407-5010-4a6f-bcb5-b555f2f7f8b1-kube-api-access-sqkml\") pod \"cluster-samples-operator-665b6dd947-6w2pm\" (UID: \"e925f407-5010-4a6f-bcb5-b555f2f7f8b1\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6w2pm" Oct 02 01:49:13 crc kubenswrapper[4885]: I1002 01:49:13.846315 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-smtqm" Oct 02 01:49:13 crc kubenswrapper[4885]: I1002 01:49:13.851373 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7lhrd\" (UniqueName: \"kubernetes.io/projected/f257ddf8-bc9c-4670-b8c3-943829b80433-kube-api-access-7lhrd\") pod \"multus-admission-controller-857f4d67dd-vbm65\" (UID: \"f257ddf8-bc9c-4670-b8c3-943829b80433\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-vbm65" Oct 02 01:49:13 crc kubenswrapper[4885]: I1002 01:49:13.860837 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-zqxvm" Oct 02 01:49:13 crc kubenswrapper[4885]: I1002 01:49:13.875185 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-vs6d8"] Oct 02 01:49:13 crc kubenswrapper[4885]: I1002 01:49:13.885692 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-smtqm" event={"ID":"5feb1fb5-73be-4ce5-a6d5-d426e2eada2e","Type":"ContainerStarted","Data":"c11526763dcb4d778e7f37d36076aaf6a2397af0b80514aacebfe97e9f2ea9c2"} Oct 02 01:49:13 crc kubenswrapper[4885]: I1002 01:49:13.895437 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-mc59z"] Oct 02 01:49:13 crc kubenswrapper[4885]: I1002 01:49:13.895447 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-vbm65" Oct 02 01:49:13 crc kubenswrapper[4885]: I1002 01:49:13.916077 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-gg9lz"] Oct 02 01:49:13 crc kubenswrapper[4885]: I1002 01:49:13.924807 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cm9ql" Oct 02 01:49:13 crc kubenswrapper[4885]: I1002 01:49:13.960029 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h4fbp\" (UniqueName: \"kubernetes.io/projected/2c17c274-7853-4717-a68e-9fa6467ee199-kube-api-access-h4fbp\") pod \"machine-config-operator-74547568cd-8nb8p\" (UID: \"2c17c274-7853-4717-a68e-9fa6467ee199\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8nb8p" Oct 02 01:49:13 crc kubenswrapper[4885]: I1002 01:49:13.961204 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6vhnl\" (UniqueName: \"kubernetes.io/projected/3de23ab3-e037-4163-b744-81ecbb340b13-kube-api-access-6vhnl\") pod \"openshift-apiserver-operator-796bbdcf4f-5vkgw\" (UID: \"3de23ab3-e037-4163-b744-81ecbb340b13\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5vkgw" Oct 02 01:49:13 crc kubenswrapper[4885]: I1002 01:49:13.963227 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rkk59\" (UniqueName: \"kubernetes.io/projected/cc6bb484-0230-46d9-9507-b641d2a5f330-kube-api-access-rkk59\") pod \"control-plane-machine-set-operator-78cbb6b69f-dpqkj\" (UID: \"cc6bb484-0230-46d9-9507-b641d2a5f330\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dpqkj" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.001681 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/35380309-be1e-4d61-a2f2-e40a09ff4814-trusted-ca\") pod \"image-registry-697d97f7c8-gdxhh\" (UID: \"35380309-be1e-4d61-a2f2-e40a09ff4814\") " pod="openshift-image-registry/image-registry-697d97f7c8-gdxhh" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.001961 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/941951e3-26bd-4bb1-a5ae-a7f90813666a-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-66jkw\" (UID: \"941951e3-26bd-4bb1-a5ae-a7f90813666a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-66jkw" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.001984 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/54328f81-6f98-4576-b16e-4a1ce2bb481b-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-98cfh\" (UID: \"54328f81-6f98-4576-b16e-4a1ce2bb481b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-98cfh" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.002003 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ns269\" (UniqueName: \"kubernetes.io/projected/a2bf9cd6-f368-4eb0-ac04-cd818c43c469-kube-api-access-ns269\") pod \"catalog-operator-68c6474976-kgxn6\" (UID: \"a2bf9cd6-f368-4eb0-ac04-cd818c43c469\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-kgxn6" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.002019 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fg475\" (UniqueName: \"kubernetes.io/projected/f0d19318-e77c-40d4-89b6-da97f7b1f91a-kube-api-access-fg475\") pod \"router-default-5444994796-zxz72\" (UID: \"f0d19318-e77c-40d4-89b6-da97f7b1f91a\") " pod="openshift-ingress/router-default-5444994796-zxz72" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.002119 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gdxhh\" (UID: \"35380309-be1e-4d61-a2f2-e40a09ff4814\") " pod="openshift-image-registry/image-registry-697d97f7c8-gdxhh" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.002162 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/f0d19318-e77c-40d4-89b6-da97f7b1f91a-stats-auth\") pod \"router-default-5444994796-zxz72\" (UID: \"f0d19318-e77c-40d4-89b6-da97f7b1f91a\") " pod="openshift-ingress/router-default-5444994796-zxz72" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.002191 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/a86af5ef-01e3-47e4-b324-208537e5ec71-console-oauth-config\") pod \"console-f9d7485db-kckm8\" (UID: \"a86af5ef-01e3-47e4-b324-208537e5ec71\") " pod="openshift-console/console-f9d7485db-kckm8" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.002215 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c5b65af5-6c3f-42db-88a4-b5a73090a390-bound-sa-token\") pod \"ingress-operator-5b745b69d9-f6gkd\" (UID: \"c5b65af5-6c3f-42db-88a4-b5a73090a390\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-f6gkd" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.002238 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5cp9p\" (UniqueName: \"kubernetes.io/projected/408ea460-05a8-4c98-a372-8eb9b3e43334-kube-api-access-5cp9p\") pod \"migrator-59844c95c7-5lbxn\" (UID: \"408ea460-05a8-4c98-a372-8eb9b3e43334\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-5lbxn" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.002285 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/54328f81-6f98-4576-b16e-4a1ce2bb481b-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-98cfh\" (UID: \"54328f81-6f98-4576-b16e-4a1ce2bb481b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-98cfh" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.002313 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c5b65af5-6c3f-42db-88a4-b5a73090a390-metrics-tls\") pod \"ingress-operator-5b745b69d9-f6gkd\" (UID: \"c5b65af5-6c3f-42db-88a4-b5a73090a390\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-f6gkd" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.002332 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-98nfr\" (UniqueName: \"kubernetes.io/projected/35380309-be1e-4d61-a2f2-e40a09ff4814-kube-api-access-98nfr\") pod \"image-registry-697d97f7c8-gdxhh\" (UID: \"35380309-be1e-4d61-a2f2-e40a09ff4814\") " pod="openshift-image-registry/image-registry-697d97f7c8-gdxhh" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.002469 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/35380309-be1e-4d61-a2f2-e40a09ff4814-registry-tls\") pod \"image-registry-697d97f7c8-gdxhh\" (UID: \"35380309-be1e-4d61-a2f2-e40a09ff4814\") " pod="openshift-image-registry/image-registry-697d97f7c8-gdxhh" Oct 02 01:49:14 crc kubenswrapper[4885]: E1002 01:49:14.002786 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 01:49:14.50276843 +0000 UTC m=+143.314515829 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gdxhh" (UID: "35380309-be1e-4d61-a2f2-e40a09ff4814") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.002937 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lthnf\" (UniqueName: \"kubernetes.io/projected/1b7f5b51-ddb5-4889-9d85-6a4940b8908e-kube-api-access-lthnf\") pod \"etcd-operator-b45778765-7nhrb\" (UID: \"1b7f5b51-ddb5-4889-9d85-6a4940b8908e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7nhrb" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.003059 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/35380309-be1e-4d61-a2f2-e40a09ff4814-installation-pull-secrets\") pod \"image-registry-697d97f7c8-gdxhh\" (UID: \"35380309-be1e-4d61-a2f2-e40a09ff4814\") " pod="openshift-image-registry/image-registry-697d97f7c8-gdxhh" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.003114 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1b7f5b51-ddb5-4889-9d85-6a4940b8908e-serving-cert\") pod \"etcd-operator-b45778765-7nhrb\" (UID: \"1b7f5b51-ddb5-4889-9d85-6a4940b8908e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7nhrb" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.003141 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1b7f5b51-ddb5-4889-9d85-6a4940b8908e-etcd-client\") pod \"etcd-operator-b45778765-7nhrb\" (UID: \"1b7f5b51-ddb5-4889-9d85-6a4940b8908e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7nhrb" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.003207 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/aa89bc07-6d38-40ba-8e90-9ecae726246d-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-z4rlr\" (UID: \"aa89bc07-6d38-40ba-8e90-9ecae726246d\") " pod="openshift-marketplace/marketplace-operator-79b997595-z4rlr" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.003230 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/1b7f5b51-ddb5-4889-9d85-6a4940b8908e-etcd-ca\") pod \"etcd-operator-b45778765-7nhrb\" (UID: \"1b7f5b51-ddb5-4889-9d85-6a4940b8908e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7nhrb" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.003248 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/1b7f5b51-ddb5-4889-9d85-6a4940b8908e-etcd-service-ca\") pod \"etcd-operator-b45778765-7nhrb\" (UID: \"1b7f5b51-ddb5-4889-9d85-6a4940b8908e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7nhrb" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.003323 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/941951e3-26bd-4bb1-a5ae-a7f90813666a-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-66jkw\" (UID: \"941951e3-26bd-4bb1-a5ae-a7f90813666a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-66jkw" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.003367 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/35380309-be1e-4d61-a2f2-e40a09ff4814-bound-sa-token\") pod \"image-registry-697d97f7c8-gdxhh\" (UID: \"35380309-be1e-4d61-a2f2-e40a09ff4814\") " pod="openshift-image-registry/image-registry-697d97f7c8-gdxhh" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.003388 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/aa89bc07-6d38-40ba-8e90-9ecae726246d-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-z4rlr\" (UID: \"aa89bc07-6d38-40ba-8e90-9ecae726246d\") " pod="openshift-marketplace/marketplace-operator-79b997595-z4rlr" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.003429 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/a86af5ef-01e3-47e4-b324-208537e5ec71-console-config\") pod \"console-f9d7485db-kckm8\" (UID: \"a86af5ef-01e3-47e4-b324-208537e5ec71\") " pod="openshift-console/console-f9d7485db-kckm8" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.003465 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f0d19318-e77c-40d4-89b6-da97f7b1f91a-service-ca-bundle\") pod \"router-default-5444994796-zxz72\" (UID: \"f0d19318-e77c-40d4-89b6-da97f7b1f91a\") " pod="openshift-ingress/router-default-5444994796-zxz72" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.003497 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/607e06a5-d8aa-403e-8976-e48c794bb041-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-qgdp5\" (UID: \"607e06a5-d8aa-403e-8976-e48c794bb041\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-qgdp5" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.003539 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/941951e3-26bd-4bb1-a5ae-a7f90813666a-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-66jkw\" (UID: \"941951e3-26bd-4bb1-a5ae-a7f90813666a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-66jkw" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.003560 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fzmz6\" (UniqueName: \"kubernetes.io/projected/607e06a5-d8aa-403e-8976-e48c794bb041-kube-api-access-fzmz6\") pod \"machine-config-controller-84d6567774-qgdp5\" (UID: \"607e06a5-d8aa-403e-8976-e48c794bb041\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-qgdp5" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.003598 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2dp4w\" (UniqueName: \"kubernetes.io/projected/54328f81-6f98-4576-b16e-4a1ce2bb481b-kube-api-access-2dp4w\") pod \"openshift-controller-manager-operator-756b6f6bc6-98cfh\" (UID: \"54328f81-6f98-4576-b16e-4a1ce2bb481b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-98cfh" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.003623 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/a86af5ef-01e3-47e4-b324-208537e5ec71-console-serving-cert\") pod \"console-f9d7485db-kckm8\" (UID: \"a86af5ef-01e3-47e4-b324-208537e5ec71\") " pod="openshift-console/console-f9d7485db-kckm8" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.003657 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nc9tk\" (UniqueName: \"kubernetes.io/projected/aa89bc07-6d38-40ba-8e90-9ecae726246d-kube-api-access-nc9tk\") pod \"marketplace-operator-79b997595-z4rlr\" (UID: \"aa89bc07-6d38-40ba-8e90-9ecae726246d\") " pod="openshift-marketplace/marketplace-operator-79b997595-z4rlr" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.003678 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/35380309-be1e-4d61-a2f2-e40a09ff4814-registry-certificates\") pod \"image-registry-697d97f7c8-gdxhh\" (UID: \"35380309-be1e-4d61-a2f2-e40a09ff4814\") " pod="openshift-image-registry/image-registry-697d97f7c8-gdxhh" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.003700 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/96d8ec00-6314-4955-97bd-2d00f28b6f19-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-xd4hd\" (UID: \"96d8ec00-6314-4955-97bd-2d00f28b6f19\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-xd4hd" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.003724 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qbmxg\" (UniqueName: \"kubernetes.io/projected/a86af5ef-01e3-47e4-b324-208537e5ec71-kube-api-access-qbmxg\") pod \"console-f9d7485db-kckm8\" (UID: \"a86af5ef-01e3-47e4-b324-208537e5ec71\") " pod="openshift-console/console-f9d7485db-kckm8" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.003751 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cc4tf\" (UniqueName: \"kubernetes.io/projected/96d8ec00-6314-4955-97bd-2d00f28b6f19-kube-api-access-cc4tf\") pod \"kube-storage-version-migrator-operator-b67b599dd-xd4hd\" (UID: \"96d8ec00-6314-4955-97bd-2d00f28b6f19\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-xd4hd" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.003789 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a86af5ef-01e3-47e4-b324-208537e5ec71-service-ca\") pod \"console-f9d7485db-kckm8\" (UID: \"a86af5ef-01e3-47e4-b324-208537e5ec71\") " pod="openshift-console/console-f9d7485db-kckm8" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.003818 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/35380309-be1e-4d61-a2f2-e40a09ff4814-ca-trust-extracted\") pod \"image-registry-697d97f7c8-gdxhh\" (UID: \"35380309-be1e-4d61-a2f2-e40a09ff4814\") " pod="openshift-image-registry/image-registry-697d97f7c8-gdxhh" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.003839 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c5b65af5-6c3f-42db-88a4-b5a73090a390-trusted-ca\") pod \"ingress-operator-5b745b69d9-f6gkd\" (UID: \"c5b65af5-6c3f-42db-88a4-b5a73090a390\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-f6gkd" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.003863 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/a86af5ef-01e3-47e4-b324-208537e5ec71-oauth-serving-cert\") pod \"console-f9d7485db-kckm8\" (UID: \"a86af5ef-01e3-47e4-b324-208537e5ec71\") " pod="openshift-console/console-f9d7485db-kckm8" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.003905 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/a2bf9cd6-f368-4eb0-ac04-cd818c43c469-profile-collector-cert\") pod \"catalog-operator-68c6474976-kgxn6\" (UID: \"a2bf9cd6-f368-4eb0-ac04-cd818c43c469\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-kgxn6" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.003952 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f0d19318-e77c-40d4-89b6-da97f7b1f91a-metrics-certs\") pod \"router-default-5444994796-zxz72\" (UID: \"f0d19318-e77c-40d4-89b6-da97f7b1f91a\") " pod="openshift-ingress/router-default-5444994796-zxz72" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.003982 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/607e06a5-d8aa-403e-8976-e48c794bb041-proxy-tls\") pod \"machine-config-controller-84d6567774-qgdp5\" (UID: \"607e06a5-d8aa-403e-8976-e48c794bb041\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-qgdp5" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.004037 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s6thl\" (UniqueName: \"kubernetes.io/projected/c5b65af5-6c3f-42db-88a4-b5a73090a390-kube-api-access-s6thl\") pod \"ingress-operator-5b745b69d9-f6gkd\" (UID: \"c5b65af5-6c3f-42db-88a4-b5a73090a390\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-f6gkd" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.004061 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/96d8ec00-6314-4955-97bd-2d00f28b6f19-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-xd4hd\" (UID: \"96d8ec00-6314-4955-97bd-2d00f28b6f19\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-xd4hd" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.004087 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a86af5ef-01e3-47e4-b324-208537e5ec71-trusted-ca-bundle\") pod \"console-f9d7485db-kckm8\" (UID: \"a86af5ef-01e3-47e4-b324-208537e5ec71\") " pod="openshift-console/console-f9d7485db-kckm8" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.004110 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/f0d19318-e77c-40d4-89b6-da97f7b1f91a-default-certificate\") pod \"router-default-5444994796-zxz72\" (UID: \"f0d19318-e77c-40d4-89b6-da97f7b1f91a\") " pod="openshift-ingress/router-default-5444994796-zxz72" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.004133 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/a2bf9cd6-f368-4eb0-ac04-cd818c43c469-srv-cert\") pod \"catalog-operator-68c6474976-kgxn6\" (UID: \"a2bf9cd6-f368-4eb0-ac04-cd818c43c469\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-kgxn6" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.004151 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b7f5b51-ddb5-4889-9d85-6a4940b8908e-config\") pod \"etcd-operator-b45778765-7nhrb\" (UID: \"1b7f5b51-ddb5-4889-9d85-6a4940b8908e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7nhrb" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.027647 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-5wldz" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.029937 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5vkgw" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.037096 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hkwl9" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.049396 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6w2pm" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.054840 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-g4vj4"] Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.105062 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:49:14 crc kubenswrapper[4885]: E1002 01:49:14.105233 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:49:14.605207259 +0000 UTC m=+143.416954658 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.105292 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lthnf\" (UniqueName: \"kubernetes.io/projected/1b7f5b51-ddb5-4889-9d85-6a4940b8908e-kube-api-access-lthnf\") pod \"etcd-operator-b45778765-7nhrb\" (UID: \"1b7f5b51-ddb5-4889-9d85-6a4940b8908e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7nhrb" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.105329 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/d2e53137-c638-4ec1-8177-00f07a5fa604-signing-key\") pod \"service-ca-9c57cc56f-c94sm\" (UID: \"d2e53137-c638-4ec1-8177-00f07a5fa604\") " pod="openshift-service-ca/service-ca-9c57cc56f-c94sm" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.105349 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/35380309-be1e-4d61-a2f2-e40a09ff4814-installation-pull-secrets\") pod \"image-registry-697d97f7c8-gdxhh\" (UID: \"35380309-be1e-4d61-a2f2-e40a09ff4814\") " pod="openshift-image-registry/image-registry-697d97f7c8-gdxhh" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.105376 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1b7f5b51-ddb5-4889-9d85-6a4940b8908e-serving-cert\") pod \"etcd-operator-b45778765-7nhrb\" (UID: \"1b7f5b51-ddb5-4889-9d85-6a4940b8908e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7nhrb" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.105390 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1b7f5b51-ddb5-4889-9d85-6a4940b8908e-etcd-client\") pod \"etcd-operator-b45778765-7nhrb\" (UID: \"1b7f5b51-ddb5-4889-9d85-6a4940b8908e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7nhrb" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.105418 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/559ee4a2-19e7-4c8a-a9a2-75e559f5517a-node-bootstrap-token\") pod \"machine-config-server-qztpv\" (UID: \"559ee4a2-19e7-4c8a-a9a2-75e559f5517a\") " pod="openshift-machine-config-operator/machine-config-server-qztpv" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.105453 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/aa89bc07-6d38-40ba-8e90-9ecae726246d-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-z4rlr\" (UID: \"aa89bc07-6d38-40ba-8e90-9ecae726246d\") " pod="openshift-marketplace/marketplace-operator-79b997595-z4rlr" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.105479 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/1b7f5b51-ddb5-4889-9d85-6a4940b8908e-etcd-ca\") pod \"etcd-operator-b45778765-7nhrb\" (UID: \"1b7f5b51-ddb5-4889-9d85-6a4940b8908e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7nhrb" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.105493 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/1b7f5b51-ddb5-4889-9d85-6a4940b8908e-etcd-service-ca\") pod \"etcd-operator-b45778765-7nhrb\" (UID: \"1b7f5b51-ddb5-4889-9d85-6a4940b8908e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7nhrb" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.105509 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/941951e3-26bd-4bb1-a5ae-a7f90813666a-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-66jkw\" (UID: \"941951e3-26bd-4bb1-a5ae-a7f90813666a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-66jkw" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.105536 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sq68j\" (UniqueName: \"kubernetes.io/projected/559ee4a2-19e7-4c8a-a9a2-75e559f5517a-kube-api-access-sq68j\") pod \"machine-config-server-qztpv\" (UID: \"559ee4a2-19e7-4c8a-a9a2-75e559f5517a\") " pod="openshift-machine-config-operator/machine-config-server-qztpv" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.105557 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rpjmm\" (UniqueName: \"kubernetes.io/projected/3b3cd833-9da7-47ca-979d-464cde398589-kube-api-access-rpjmm\") pod \"dns-default-6bfsg\" (UID: \"3b3cd833-9da7-47ca-979d-464cde398589\") " pod="openshift-dns/dns-default-6bfsg" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.105574 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/35380309-be1e-4d61-a2f2-e40a09ff4814-bound-sa-token\") pod \"image-registry-697d97f7c8-gdxhh\" (UID: \"35380309-be1e-4d61-a2f2-e40a09ff4814\") " pod="openshift-image-registry/image-registry-697d97f7c8-gdxhh" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.105600 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/a86af5ef-01e3-47e4-b324-208537e5ec71-console-config\") pod \"console-f9d7485db-kckm8\" (UID: \"a86af5ef-01e3-47e4-b324-208537e5ec71\") " pod="openshift-console/console-f9d7485db-kckm8" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.105615 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/aa89bc07-6d38-40ba-8e90-9ecae726246d-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-z4rlr\" (UID: \"aa89bc07-6d38-40ba-8e90-9ecae726246d\") " pod="openshift-marketplace/marketplace-operator-79b997595-z4rlr" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.105630 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0b1aface-d1e4-4d8b-a0e6-08f9032c3554-webhook-cert\") pod \"packageserver-d55dfcdfc-4wjj8\" (UID: \"0b1aface-d1e4-4d8b-a0e6-08f9032c3554\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-4wjj8" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.105646 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f0d19318-e77c-40d4-89b6-da97f7b1f91a-service-ca-bundle\") pod \"router-default-5444994796-zxz72\" (UID: \"f0d19318-e77c-40d4-89b6-da97f7b1f91a\") " pod="openshift-ingress/router-default-5444994796-zxz72" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.105662 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e8f0ddce-4638-4e1c-b27b-00b991807f23-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-k8r8q\" (UID: \"e8f0ddce-4638-4e1c-b27b-00b991807f23\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-k8r8q" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.105680 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jjdjh\" (UniqueName: \"kubernetes.io/projected/d2e53137-c638-4ec1-8177-00f07a5fa604-kube-api-access-jjdjh\") pod \"service-ca-9c57cc56f-c94sm\" (UID: \"d2e53137-c638-4ec1-8177-00f07a5fa604\") " pod="openshift-service-ca/service-ca-9c57cc56f-c94sm" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.105703 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/607e06a5-d8aa-403e-8976-e48c794bb041-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-qgdp5\" (UID: \"607e06a5-d8aa-403e-8976-e48c794bb041\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-qgdp5" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.105718 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/941951e3-26bd-4bb1-a5ae-a7f90813666a-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-66jkw\" (UID: \"941951e3-26bd-4bb1-a5ae-a7f90813666a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-66jkw" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.105733 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fzmz6\" (UniqueName: \"kubernetes.io/projected/607e06a5-d8aa-403e-8976-e48c794bb041-kube-api-access-fzmz6\") pod \"machine-config-controller-84d6567774-qgdp5\" (UID: \"607e06a5-d8aa-403e-8976-e48c794bb041\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-qgdp5" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.105750 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s4bc5\" (UniqueName: \"kubernetes.io/projected/858faa4e-3646-4a78-b9a8-0fdb8b2a54df-kube-api-access-s4bc5\") pod \"ingress-canary-xft2d\" (UID: \"858faa4e-3646-4a78-b9a8-0fdb8b2a54df\") " pod="openshift-ingress-canary/ingress-canary-xft2d" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.105778 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2dp4w\" (UniqueName: \"kubernetes.io/projected/54328f81-6f98-4576-b16e-4a1ce2bb481b-kube-api-access-2dp4w\") pod \"openshift-controller-manager-operator-756b6f6bc6-98cfh\" (UID: \"54328f81-6f98-4576-b16e-4a1ce2bb481b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-98cfh" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.105794 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zhzzc\" (UniqueName: \"kubernetes.io/projected/0b1aface-d1e4-4d8b-a0e6-08f9032c3554-kube-api-access-zhzzc\") pod \"packageserver-d55dfcdfc-4wjj8\" (UID: \"0b1aface-d1e4-4d8b-a0e6-08f9032c3554\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-4wjj8" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.105819 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/691b626a-d098-4351-9d7b-8bacf7803327-socket-dir\") pod \"csi-hostpathplugin-n7bs2\" (UID: \"691b626a-d098-4351-9d7b-8bacf7803327\") " pod="hostpath-provisioner/csi-hostpathplugin-n7bs2" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.105834 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/0b1aface-d1e4-4d8b-a0e6-08f9032c3554-tmpfs\") pod \"packageserver-d55dfcdfc-4wjj8\" (UID: \"0b1aface-d1e4-4d8b-a0e6-08f9032c3554\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-4wjj8" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.105850 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b2hdq\" (UniqueName: \"kubernetes.io/projected/8f81f79b-3fa8-430a-a21b-df5fd127e485-kube-api-access-b2hdq\") pod \"service-ca-operator-777779d784-7hk2v\" (UID: \"8f81f79b-3fa8-430a-a21b-df5fd127e485\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-7hk2v" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.105874 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/a86af5ef-01e3-47e4-b324-208537e5ec71-console-serving-cert\") pod \"console-f9d7485db-kckm8\" (UID: \"a86af5ef-01e3-47e4-b324-208537e5ec71\") " pod="openshift-console/console-f9d7485db-kckm8" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.105891 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7rqt5\" (UniqueName: \"kubernetes.io/projected/4bc961f3-ae85-4c19-8a57-be0af8e7bd10-kube-api-access-7rqt5\") pod \"collect-profiles-29322825-b7zrt\" (UID: \"4bc961f3-ae85-4c19-8a57-be0af8e7bd10\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322825-b7zrt" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.105905 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3b3cd833-9da7-47ca-979d-464cde398589-config-volume\") pod \"dns-default-6bfsg\" (UID: \"3b3cd833-9da7-47ca-979d-464cde398589\") " pod="openshift-dns/dns-default-6bfsg" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.105922 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nc9tk\" (UniqueName: \"kubernetes.io/projected/aa89bc07-6d38-40ba-8e90-9ecae726246d-kube-api-access-nc9tk\") pod \"marketplace-operator-79b997595-z4rlr\" (UID: \"aa89bc07-6d38-40ba-8e90-9ecae726246d\") " pod="openshift-marketplace/marketplace-operator-79b997595-z4rlr" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.105947 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/35380309-be1e-4d61-a2f2-e40a09ff4814-registry-certificates\") pod \"image-registry-697d97f7c8-gdxhh\" (UID: \"35380309-be1e-4d61-a2f2-e40a09ff4814\") " pod="openshift-image-registry/image-registry-697d97f7c8-gdxhh" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.105963 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/691b626a-d098-4351-9d7b-8bacf7803327-registration-dir\") pod \"csi-hostpathplugin-n7bs2\" (UID: \"691b626a-d098-4351-9d7b-8bacf7803327\") " pod="hostpath-provisioner/csi-hostpathplugin-n7bs2" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.105982 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/96d8ec00-6314-4955-97bd-2d00f28b6f19-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-xd4hd\" (UID: \"96d8ec00-6314-4955-97bd-2d00f28b6f19\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-xd4hd" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.105997 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qbmxg\" (UniqueName: \"kubernetes.io/projected/a86af5ef-01e3-47e4-b324-208537e5ec71-kube-api-access-qbmxg\") pod \"console-f9d7485db-kckm8\" (UID: \"a86af5ef-01e3-47e4-b324-208537e5ec71\") " pod="openshift-console/console-f9d7485db-kckm8" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.106013 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cc4tf\" (UniqueName: \"kubernetes.io/projected/96d8ec00-6314-4955-97bd-2d00f28b6f19-kube-api-access-cc4tf\") pod \"kube-storage-version-migrator-operator-b67b599dd-xd4hd\" (UID: \"96d8ec00-6314-4955-97bd-2d00f28b6f19\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-xd4hd" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.106056 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a86af5ef-01e3-47e4-b324-208537e5ec71-service-ca\") pod \"console-f9d7485db-kckm8\" (UID: \"a86af5ef-01e3-47e4-b324-208537e5ec71\") " pod="openshift-console/console-f9d7485db-kckm8" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.106071 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/3b3cd833-9da7-47ca-979d-464cde398589-metrics-tls\") pod \"dns-default-6bfsg\" (UID: \"3b3cd833-9da7-47ca-979d-464cde398589\") " pod="openshift-dns/dns-default-6bfsg" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.106111 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8f0ddce-4638-4e1c-b27b-00b991807f23-config\") pod \"kube-apiserver-operator-766d6c64bb-k8r8q\" (UID: \"e8f0ddce-4638-4e1c-b27b-00b991807f23\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-k8r8q" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.106128 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/35380309-be1e-4d61-a2f2-e40a09ff4814-ca-trust-extracted\") pod \"image-registry-697d97f7c8-gdxhh\" (UID: \"35380309-be1e-4d61-a2f2-e40a09ff4814\") " pod="openshift-image-registry/image-registry-697d97f7c8-gdxhh" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.106145 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c5b65af5-6c3f-42db-88a4-b5a73090a390-trusted-ca\") pod \"ingress-operator-5b745b69d9-f6gkd\" (UID: \"c5b65af5-6c3f-42db-88a4-b5a73090a390\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-f6gkd" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.106160 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/a86af5ef-01e3-47e4-b324-208537e5ec71-oauth-serving-cert\") pod \"console-f9d7485db-kckm8\" (UID: \"a86af5ef-01e3-47e4-b324-208537e5ec71\") " pod="openshift-console/console-f9d7485db-kckm8" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.106175 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/559ee4a2-19e7-4c8a-a9a2-75e559f5517a-certs\") pod \"machine-config-server-qztpv\" (UID: \"559ee4a2-19e7-4c8a-a9a2-75e559f5517a\") " pod="openshift-machine-config-operator/machine-config-server-qztpv" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.106213 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/a2bf9cd6-f368-4eb0-ac04-cd818c43c469-profile-collector-cert\") pod \"catalog-operator-68c6474976-kgxn6\" (UID: \"a2bf9cd6-f368-4eb0-ac04-cd818c43c469\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-kgxn6" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.106271 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f0d19318-e77c-40d4-89b6-da97f7b1f91a-metrics-certs\") pod \"router-default-5444994796-zxz72\" (UID: \"f0d19318-e77c-40d4-89b6-da97f7b1f91a\") " pod="openshift-ingress/router-default-5444994796-zxz72" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.106288 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4bc961f3-ae85-4c19-8a57-be0af8e7bd10-config-volume\") pod \"collect-profiles-29322825-b7zrt\" (UID: \"4bc961f3-ae85-4c19-8a57-be0af8e7bd10\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322825-b7zrt" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.106313 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/1d4a35f8-1702-438b-ad5c-e48eacc16666-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-xjfdv\" (UID: \"1d4a35f8-1702-438b-ad5c-e48eacc16666\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xjfdv" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.106339 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/607e06a5-d8aa-403e-8976-e48c794bb041-proxy-tls\") pod \"machine-config-controller-84d6567774-qgdp5\" (UID: \"607e06a5-d8aa-403e-8976-e48c794bb041\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-qgdp5" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.106373 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e8f0ddce-4638-4e1c-b27b-00b991807f23-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-k8r8q\" (UID: \"e8f0ddce-4638-4e1c-b27b-00b991807f23\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-k8r8q" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.106395 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/691b626a-d098-4351-9d7b-8bacf7803327-mountpoint-dir\") pod \"csi-hostpathplugin-n7bs2\" (UID: \"691b626a-d098-4351-9d7b-8bacf7803327\") " pod="hostpath-provisioner/csi-hostpathplugin-n7bs2" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.106411 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s6thl\" (UniqueName: \"kubernetes.io/projected/c5b65af5-6c3f-42db-88a4-b5a73090a390-kube-api-access-s6thl\") pod \"ingress-operator-5b745b69d9-f6gkd\" (UID: \"c5b65af5-6c3f-42db-88a4-b5a73090a390\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-f6gkd" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.106427 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/d2e53137-c638-4ec1-8177-00f07a5fa604-signing-cabundle\") pod \"service-ca-9c57cc56f-c94sm\" (UID: \"d2e53137-c638-4ec1-8177-00f07a5fa604\") " pod="openshift-service-ca/service-ca-9c57cc56f-c94sm" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.106455 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/96d8ec00-6314-4955-97bd-2d00f28b6f19-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-xd4hd\" (UID: \"96d8ec00-6314-4955-97bd-2d00f28b6f19\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-xd4hd" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.106471 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/691b626a-d098-4351-9d7b-8bacf7803327-csi-data-dir\") pod \"csi-hostpathplugin-n7bs2\" (UID: \"691b626a-d098-4351-9d7b-8bacf7803327\") " pod="hostpath-provisioner/csi-hostpathplugin-n7bs2" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.106486 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a86af5ef-01e3-47e4-b324-208537e5ec71-trusted-ca-bundle\") pod \"console-f9d7485db-kckm8\" (UID: \"a86af5ef-01e3-47e4-b324-208537e5ec71\") " pod="openshift-console/console-f9d7485db-kckm8" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.106501 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/f0d19318-e77c-40d4-89b6-da97f7b1f91a-default-certificate\") pod \"router-default-5444994796-zxz72\" (UID: \"f0d19318-e77c-40d4-89b6-da97f7b1f91a\") " pod="openshift-ingress/router-default-5444994796-zxz72" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.106516 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0b1aface-d1e4-4d8b-a0e6-08f9032c3554-apiservice-cert\") pod \"packageserver-d55dfcdfc-4wjj8\" (UID: \"0b1aface-d1e4-4d8b-a0e6-08f9032c3554\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-4wjj8" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.106552 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4bc961f3-ae85-4c19-8a57-be0af8e7bd10-secret-volume\") pod \"collect-profiles-29322825-b7zrt\" (UID: \"4bc961f3-ae85-4c19-8a57-be0af8e7bd10\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322825-b7zrt" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.106578 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/a2bf9cd6-f368-4eb0-ac04-cd818c43c469-srv-cert\") pod \"catalog-operator-68c6474976-kgxn6\" (UID: \"a2bf9cd6-f368-4eb0-ac04-cd818c43c469\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-kgxn6" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.106593 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b7f5b51-ddb5-4889-9d85-6a4940b8908e-config\") pod \"etcd-operator-b45778765-7nhrb\" (UID: \"1b7f5b51-ddb5-4889-9d85-6a4940b8908e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7nhrb" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.106628 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/35380309-be1e-4d61-a2f2-e40a09ff4814-trusted-ca\") pod \"image-registry-697d97f7c8-gdxhh\" (UID: \"35380309-be1e-4d61-a2f2-e40a09ff4814\") " pod="openshift-image-registry/image-registry-697d97f7c8-gdxhh" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.106642 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g6qqs\" (UniqueName: \"kubernetes.io/projected/691b626a-d098-4351-9d7b-8bacf7803327-kube-api-access-g6qqs\") pod \"csi-hostpathplugin-n7bs2\" (UID: \"691b626a-d098-4351-9d7b-8bacf7803327\") " pod="hostpath-provisioner/csi-hostpathplugin-n7bs2" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.106669 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/941951e3-26bd-4bb1-a5ae-a7f90813666a-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-66jkw\" (UID: \"941951e3-26bd-4bb1-a5ae-a7f90813666a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-66jkw" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.106686 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/54328f81-6f98-4576-b16e-4a1ce2bb481b-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-98cfh\" (UID: \"54328f81-6f98-4576-b16e-4a1ce2bb481b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-98cfh" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.106702 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ns269\" (UniqueName: \"kubernetes.io/projected/a2bf9cd6-f368-4eb0-ac04-cd818c43c469-kube-api-access-ns269\") pod \"catalog-operator-68c6474976-kgxn6\" (UID: \"a2bf9cd6-f368-4eb0-ac04-cd818c43c469\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-kgxn6" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.106782 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fg475\" (UniqueName: \"kubernetes.io/projected/f0d19318-e77c-40d4-89b6-da97f7b1f91a-kube-api-access-fg475\") pod \"router-default-5444994796-zxz72\" (UID: \"f0d19318-e77c-40d4-89b6-da97f7b1f91a\") " pod="openshift-ingress/router-default-5444994796-zxz72" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.106814 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/691b626a-d098-4351-9d7b-8bacf7803327-plugins-dir\") pod \"csi-hostpathplugin-n7bs2\" (UID: \"691b626a-d098-4351-9d7b-8bacf7803327\") " pod="hostpath-provisioner/csi-hostpathplugin-n7bs2" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.106842 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gdxhh\" (UID: \"35380309-be1e-4d61-a2f2-e40a09ff4814\") " pod="openshift-image-registry/image-registry-697d97f7c8-gdxhh" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.106858 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fqhmp\" (UniqueName: \"kubernetes.io/projected/1d4a35f8-1702-438b-ad5c-e48eacc16666-kube-api-access-fqhmp\") pod \"package-server-manager-789f6589d5-xjfdv\" (UID: \"1d4a35f8-1702-438b-ad5c-e48eacc16666\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xjfdv" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.106893 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/f0d19318-e77c-40d4-89b6-da97f7b1f91a-stats-auth\") pod \"router-default-5444994796-zxz72\" (UID: \"f0d19318-e77c-40d4-89b6-da97f7b1f91a\") " pod="openshift-ingress/router-default-5444994796-zxz72" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.106917 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/858faa4e-3646-4a78-b9a8-0fdb8b2a54df-cert\") pod \"ingress-canary-xft2d\" (UID: \"858faa4e-3646-4a78-b9a8-0fdb8b2a54df\") " pod="openshift-ingress-canary/ingress-canary-xft2d" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.106954 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/a86af5ef-01e3-47e4-b324-208537e5ec71-console-oauth-config\") pod \"console-f9d7485db-kckm8\" (UID: \"a86af5ef-01e3-47e4-b324-208537e5ec71\") " pod="openshift-console/console-f9d7485db-kckm8" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.106971 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c5b65af5-6c3f-42db-88a4-b5a73090a390-bound-sa-token\") pod \"ingress-operator-5b745b69d9-f6gkd\" (UID: \"c5b65af5-6c3f-42db-88a4-b5a73090a390\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-f6gkd" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.106986 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f81f79b-3fa8-430a-a21b-df5fd127e485-config\") pod \"service-ca-operator-777779d784-7hk2v\" (UID: \"8f81f79b-3fa8-430a-a21b-df5fd127e485\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-7hk2v" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.107022 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5cp9p\" (UniqueName: \"kubernetes.io/projected/408ea460-05a8-4c98-a372-8eb9b3e43334-kube-api-access-5cp9p\") pod \"migrator-59844c95c7-5lbxn\" (UID: \"408ea460-05a8-4c98-a372-8eb9b3e43334\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-5lbxn" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.107038 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/54328f81-6f98-4576-b16e-4a1ce2bb481b-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-98cfh\" (UID: \"54328f81-6f98-4576-b16e-4a1ce2bb481b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-98cfh" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.107074 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-98nfr\" (UniqueName: \"kubernetes.io/projected/35380309-be1e-4d61-a2f2-e40a09ff4814-kube-api-access-98nfr\") pod \"image-registry-697d97f7c8-gdxhh\" (UID: \"35380309-be1e-4d61-a2f2-e40a09ff4814\") " pod="openshift-image-registry/image-registry-697d97f7c8-gdxhh" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.107411 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c5b65af5-6c3f-42db-88a4-b5a73090a390-metrics-tls\") pod \"ingress-operator-5b745b69d9-f6gkd\" (UID: \"c5b65af5-6c3f-42db-88a4-b5a73090a390\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-f6gkd" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.107435 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8f81f79b-3fa8-430a-a21b-df5fd127e485-serving-cert\") pod \"service-ca-operator-777779d784-7hk2v\" (UID: \"8f81f79b-3fa8-430a-a21b-df5fd127e485\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-7hk2v" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.107585 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/35380309-be1e-4d61-a2f2-e40a09ff4814-registry-tls\") pod \"image-registry-697d97f7c8-gdxhh\" (UID: \"35380309-be1e-4d61-a2f2-e40a09ff4814\") " pod="openshift-image-registry/image-registry-697d97f7c8-gdxhh" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.112192 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/35380309-be1e-4d61-a2f2-e40a09ff4814-installation-pull-secrets\") pod \"image-registry-697d97f7c8-gdxhh\" (UID: \"35380309-be1e-4d61-a2f2-e40a09ff4814\") " pod="openshift-image-registry/image-registry-697d97f7c8-gdxhh" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.113119 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b7f5b51-ddb5-4889-9d85-6a4940b8908e-config\") pod \"etcd-operator-b45778765-7nhrb\" (UID: \"1b7f5b51-ddb5-4889-9d85-6a4940b8908e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7nhrb" Oct 02 01:49:14 crc kubenswrapper[4885]: E1002 01:49:14.114226 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 01:49:14.61421285 +0000 UTC m=+143.425960249 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gdxhh" (UID: "35380309-be1e-4d61-a2f2-e40a09ff4814") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.114751 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/54328f81-6f98-4576-b16e-4a1ce2bb481b-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-98cfh\" (UID: \"54328f81-6f98-4576-b16e-4a1ce2bb481b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-98cfh" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.116150 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/a86af5ef-01e3-47e4-b324-208537e5ec71-console-config\") pod \"console-f9d7485db-kckm8\" (UID: \"a86af5ef-01e3-47e4-b324-208537e5ec71\") " pod="openshift-console/console-f9d7485db-kckm8" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.117491 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/1b7f5b51-ddb5-4889-9d85-6a4940b8908e-etcd-ca\") pod \"etcd-operator-b45778765-7nhrb\" (UID: \"1b7f5b51-ddb5-4889-9d85-6a4940b8908e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7nhrb" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.119024 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f0d19318-e77c-40d4-89b6-da97f7b1f91a-service-ca-bundle\") pod \"router-default-5444994796-zxz72\" (UID: \"f0d19318-e77c-40d4-89b6-da97f7b1f91a\") " pod="openshift-ingress/router-default-5444994796-zxz72" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.119432 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/607e06a5-d8aa-403e-8976-e48c794bb041-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-qgdp5\" (UID: \"607e06a5-d8aa-403e-8976-e48c794bb041\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-qgdp5" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.119546 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/941951e3-26bd-4bb1-a5ae-a7f90813666a-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-66jkw\" (UID: \"941951e3-26bd-4bb1-a5ae-a7f90813666a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-66jkw" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.119906 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/1b7f5b51-ddb5-4889-9d85-6a4940b8908e-etcd-service-ca\") pod \"etcd-operator-b45778765-7nhrb\" (UID: \"1b7f5b51-ddb5-4889-9d85-6a4940b8908e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7nhrb" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.120366 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/35380309-be1e-4d61-a2f2-e40a09ff4814-ca-trust-extracted\") pod \"image-registry-697d97f7c8-gdxhh\" (UID: \"35380309-be1e-4d61-a2f2-e40a09ff4814\") " pod="openshift-image-registry/image-registry-697d97f7c8-gdxhh" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.121543 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/35380309-be1e-4d61-a2f2-e40a09ff4814-registry-tls\") pod \"image-registry-697d97f7c8-gdxhh\" (UID: \"35380309-be1e-4d61-a2f2-e40a09ff4814\") " pod="openshift-image-registry/image-registry-697d97f7c8-gdxhh" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.121731 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a86af5ef-01e3-47e4-b324-208537e5ec71-trusted-ca-bundle\") pod \"console-f9d7485db-kckm8\" (UID: \"a86af5ef-01e3-47e4-b324-208537e5ec71\") " pod="openshift-console/console-f9d7485db-kckm8" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.122758 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c5b65af5-6c3f-42db-88a4-b5a73090a390-trusted-ca\") pod \"ingress-operator-5b745b69d9-f6gkd\" (UID: \"c5b65af5-6c3f-42db-88a4-b5a73090a390\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-f6gkd" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.123862 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1b7f5b51-ddb5-4889-9d85-6a4940b8908e-serving-cert\") pod \"etcd-operator-b45778765-7nhrb\" (UID: \"1b7f5b51-ddb5-4889-9d85-6a4940b8908e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7nhrb" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.125250 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/35380309-be1e-4d61-a2f2-e40a09ff4814-trusted-ca\") pod \"image-registry-697d97f7c8-gdxhh\" (UID: \"35380309-be1e-4d61-a2f2-e40a09ff4814\") " pod="openshift-image-registry/image-registry-697d97f7c8-gdxhh" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.126034 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/aa89bc07-6d38-40ba-8e90-9ecae726246d-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-z4rlr\" (UID: \"aa89bc07-6d38-40ba-8e90-9ecae726246d\") " pod="openshift-marketplace/marketplace-operator-79b997595-z4rlr" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.126125 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/35380309-be1e-4d61-a2f2-e40a09ff4814-registry-certificates\") pod \"image-registry-697d97f7c8-gdxhh\" (UID: \"35380309-be1e-4d61-a2f2-e40a09ff4814\") " pod="openshift-image-registry/image-registry-697d97f7c8-gdxhh" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.126895 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/a2bf9cd6-f368-4eb0-ac04-cd818c43c469-profile-collector-cert\") pod \"catalog-operator-68c6474976-kgxn6\" (UID: \"a2bf9cd6-f368-4eb0-ac04-cd818c43c469\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-kgxn6" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.144589 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a86af5ef-01e3-47e4-b324-208537e5ec71-service-ca\") pod \"console-f9d7485db-kckm8\" (UID: \"a86af5ef-01e3-47e4-b324-208537e5ec71\") " pod="openshift-console/console-f9d7485db-kckm8" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.145980 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/a2bf9cd6-f368-4eb0-ac04-cd818c43c469-srv-cert\") pod \"catalog-operator-68c6474976-kgxn6\" (UID: \"a2bf9cd6-f368-4eb0-ac04-cd818c43c469\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-kgxn6" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.148459 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/f0d19318-e77c-40d4-89b6-da97f7b1f91a-default-certificate\") pod \"router-default-5444994796-zxz72\" (UID: \"f0d19318-e77c-40d4-89b6-da97f7b1f91a\") " pod="openshift-ingress/router-default-5444994796-zxz72" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.151186 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/96d8ec00-6314-4955-97bd-2d00f28b6f19-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-xd4hd\" (UID: \"96d8ec00-6314-4955-97bd-2d00f28b6f19\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-xd4hd" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.152776 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/aa89bc07-6d38-40ba-8e90-9ecae726246d-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-z4rlr\" (UID: \"aa89bc07-6d38-40ba-8e90-9ecae726246d\") " pod="openshift-marketplace/marketplace-operator-79b997595-z4rlr" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.153458 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f0d19318-e77c-40d4-89b6-da97f7b1f91a-metrics-certs\") pod \"router-default-5444994796-zxz72\" (UID: \"f0d19318-e77c-40d4-89b6-da97f7b1f91a\") " pod="openshift-ingress/router-default-5444994796-zxz72" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.153484 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/f0d19318-e77c-40d4-89b6-da97f7b1f91a-stats-auth\") pod \"router-default-5444994796-zxz72\" (UID: \"f0d19318-e77c-40d4-89b6-da97f7b1f91a\") " pod="openshift-ingress/router-default-5444994796-zxz72" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.154331 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dpqkj" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.156638 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/941951e3-26bd-4bb1-a5ae-a7f90813666a-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-66jkw\" (UID: \"941951e3-26bd-4bb1-a5ae-a7f90813666a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-66jkw" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.160879 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/607e06a5-d8aa-403e-8976-e48c794bb041-proxy-tls\") pod \"machine-config-controller-84d6567774-qgdp5\" (UID: \"607e06a5-d8aa-403e-8976-e48c794bb041\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-qgdp5" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.162392 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/96d8ec00-6314-4955-97bd-2d00f28b6f19-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-xd4hd\" (UID: \"96d8ec00-6314-4955-97bd-2d00f28b6f19\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-xd4hd" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.162509 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/a86af5ef-01e3-47e4-b324-208537e5ec71-console-oauth-config\") pod \"console-f9d7485db-kckm8\" (UID: \"a86af5ef-01e3-47e4-b324-208537e5ec71\") " pod="openshift-console/console-f9d7485db-kckm8" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.162691 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/54328f81-6f98-4576-b16e-4a1ce2bb481b-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-98cfh\" (UID: \"54328f81-6f98-4576-b16e-4a1ce2bb481b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-98cfh" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.162698 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/a86af5ef-01e3-47e4-b324-208537e5ec71-console-serving-cert\") pod \"console-f9d7485db-kckm8\" (UID: \"a86af5ef-01e3-47e4-b324-208537e5ec71\") " pod="openshift-console/console-f9d7485db-kckm8" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.162876 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c5b65af5-6c3f-42db-88a4-b5a73090a390-metrics-tls\") pod \"ingress-operator-5b745b69d9-f6gkd\" (UID: \"c5b65af5-6c3f-42db-88a4-b5a73090a390\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-f6gkd" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.163428 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/a86af5ef-01e3-47e4-b324-208537e5ec71-oauth-serving-cert\") pod \"console-f9d7485db-kckm8\" (UID: \"a86af5ef-01e3-47e4-b324-208537e5ec71\") " pod="openshift-console/console-f9d7485db-kckm8" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.167976 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1b7f5b51-ddb5-4889-9d85-6a4940b8908e-etcd-client\") pod \"etcd-operator-b45778765-7nhrb\" (UID: \"1b7f5b51-ddb5-4889-9d85-6a4940b8908e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7nhrb" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.180996 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fzmz6\" (UniqueName: \"kubernetes.io/projected/607e06a5-d8aa-403e-8976-e48c794bb041-kube-api-access-fzmz6\") pod \"machine-config-controller-84d6567774-qgdp5\" (UID: \"607e06a5-d8aa-403e-8976-e48c794bb041\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-qgdp5" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.183848 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8d6k4"] Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.193248 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lthnf\" (UniqueName: \"kubernetes.io/projected/1b7f5b51-ddb5-4889-9d85-6a4940b8908e-kube-api-access-lthnf\") pod \"etcd-operator-b45778765-7nhrb\" (UID: \"1b7f5b51-ddb5-4889-9d85-6a4940b8908e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7nhrb" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.205867 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8nb8p" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.208548 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:49:14 crc kubenswrapper[4885]: E1002 01:49:14.208673 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:49:14.70865431 +0000 UTC m=+143.520401709 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.208814 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e8f0ddce-4638-4e1c-b27b-00b991807f23-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-k8r8q\" (UID: \"e8f0ddce-4638-4e1c-b27b-00b991807f23\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-k8r8q" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.208843 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/691b626a-d098-4351-9d7b-8bacf7803327-mountpoint-dir\") pod \"csi-hostpathplugin-n7bs2\" (UID: \"691b626a-d098-4351-9d7b-8bacf7803327\") " pod="hostpath-provisioner/csi-hostpathplugin-n7bs2" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.208875 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/d2e53137-c638-4ec1-8177-00f07a5fa604-signing-cabundle\") pod \"service-ca-9c57cc56f-c94sm\" (UID: \"d2e53137-c638-4ec1-8177-00f07a5fa604\") " pod="openshift-service-ca/service-ca-9c57cc56f-c94sm" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.208891 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/691b626a-d098-4351-9d7b-8bacf7803327-csi-data-dir\") pod \"csi-hostpathplugin-n7bs2\" (UID: \"691b626a-d098-4351-9d7b-8bacf7803327\") " pod="hostpath-provisioner/csi-hostpathplugin-n7bs2" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.208907 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0b1aface-d1e4-4d8b-a0e6-08f9032c3554-apiservice-cert\") pod \"packageserver-d55dfcdfc-4wjj8\" (UID: \"0b1aface-d1e4-4d8b-a0e6-08f9032c3554\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-4wjj8" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.208927 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4bc961f3-ae85-4c19-8a57-be0af8e7bd10-secret-volume\") pod \"collect-profiles-29322825-b7zrt\" (UID: \"4bc961f3-ae85-4c19-8a57-be0af8e7bd10\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322825-b7zrt" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.208944 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g6qqs\" (UniqueName: \"kubernetes.io/projected/691b626a-d098-4351-9d7b-8bacf7803327-kube-api-access-g6qqs\") pod \"csi-hostpathplugin-n7bs2\" (UID: \"691b626a-d098-4351-9d7b-8bacf7803327\") " pod="hostpath-provisioner/csi-hostpathplugin-n7bs2" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.208972 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/691b626a-d098-4351-9d7b-8bacf7803327-plugins-dir\") pod \"csi-hostpathplugin-n7bs2\" (UID: \"691b626a-d098-4351-9d7b-8bacf7803327\") " pod="hostpath-provisioner/csi-hostpathplugin-n7bs2" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.208992 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gdxhh\" (UID: \"35380309-be1e-4d61-a2f2-e40a09ff4814\") " pod="openshift-image-registry/image-registry-697d97f7c8-gdxhh" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.209010 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fqhmp\" (UniqueName: \"kubernetes.io/projected/1d4a35f8-1702-438b-ad5c-e48eacc16666-kube-api-access-fqhmp\") pod \"package-server-manager-789f6589d5-xjfdv\" (UID: \"1d4a35f8-1702-438b-ad5c-e48eacc16666\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xjfdv" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.209026 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/858faa4e-3646-4a78-b9a8-0fdb8b2a54df-cert\") pod \"ingress-canary-xft2d\" (UID: \"858faa4e-3646-4a78-b9a8-0fdb8b2a54df\") " pod="openshift-ingress-canary/ingress-canary-xft2d" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.209059 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f81f79b-3fa8-430a-a21b-df5fd127e485-config\") pod \"service-ca-operator-777779d784-7hk2v\" (UID: \"8f81f79b-3fa8-430a-a21b-df5fd127e485\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-7hk2v" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.209087 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8f81f79b-3fa8-430a-a21b-df5fd127e485-serving-cert\") pod \"service-ca-operator-777779d784-7hk2v\" (UID: \"8f81f79b-3fa8-430a-a21b-df5fd127e485\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-7hk2v" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.209106 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/d2e53137-c638-4ec1-8177-00f07a5fa604-signing-key\") pod \"service-ca-9c57cc56f-c94sm\" (UID: \"d2e53137-c638-4ec1-8177-00f07a5fa604\") " pod="openshift-service-ca/service-ca-9c57cc56f-c94sm" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.209130 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/559ee4a2-19e7-4c8a-a9a2-75e559f5517a-node-bootstrap-token\") pod \"machine-config-server-qztpv\" (UID: \"559ee4a2-19e7-4c8a-a9a2-75e559f5517a\") " pod="openshift-machine-config-operator/machine-config-server-qztpv" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.209181 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sq68j\" (UniqueName: \"kubernetes.io/projected/559ee4a2-19e7-4c8a-a9a2-75e559f5517a-kube-api-access-sq68j\") pod \"machine-config-server-qztpv\" (UID: \"559ee4a2-19e7-4c8a-a9a2-75e559f5517a\") " pod="openshift-machine-config-operator/machine-config-server-qztpv" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.209198 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rpjmm\" (UniqueName: \"kubernetes.io/projected/3b3cd833-9da7-47ca-979d-464cde398589-kube-api-access-rpjmm\") pod \"dns-default-6bfsg\" (UID: \"3b3cd833-9da7-47ca-979d-464cde398589\") " pod="openshift-dns/dns-default-6bfsg" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.209219 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0b1aface-d1e4-4d8b-a0e6-08f9032c3554-webhook-cert\") pod \"packageserver-d55dfcdfc-4wjj8\" (UID: \"0b1aface-d1e4-4d8b-a0e6-08f9032c3554\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-4wjj8" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.209241 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e8f0ddce-4638-4e1c-b27b-00b991807f23-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-k8r8q\" (UID: \"e8f0ddce-4638-4e1c-b27b-00b991807f23\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-k8r8q" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.209267 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jjdjh\" (UniqueName: \"kubernetes.io/projected/d2e53137-c638-4ec1-8177-00f07a5fa604-kube-api-access-jjdjh\") pod \"service-ca-9c57cc56f-c94sm\" (UID: \"d2e53137-c638-4ec1-8177-00f07a5fa604\") " pod="openshift-service-ca/service-ca-9c57cc56f-c94sm" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.209286 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s4bc5\" (UniqueName: \"kubernetes.io/projected/858faa4e-3646-4a78-b9a8-0fdb8b2a54df-kube-api-access-s4bc5\") pod \"ingress-canary-xft2d\" (UID: \"858faa4e-3646-4a78-b9a8-0fdb8b2a54df\") " pod="openshift-ingress-canary/ingress-canary-xft2d" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.209309 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zhzzc\" (UniqueName: \"kubernetes.io/projected/0b1aface-d1e4-4d8b-a0e6-08f9032c3554-kube-api-access-zhzzc\") pod \"packageserver-d55dfcdfc-4wjj8\" (UID: \"0b1aface-d1e4-4d8b-a0e6-08f9032c3554\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-4wjj8" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.209325 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/691b626a-d098-4351-9d7b-8bacf7803327-socket-dir\") pod \"csi-hostpathplugin-n7bs2\" (UID: \"691b626a-d098-4351-9d7b-8bacf7803327\") " pod="hostpath-provisioner/csi-hostpathplugin-n7bs2" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.209342 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/0b1aface-d1e4-4d8b-a0e6-08f9032c3554-tmpfs\") pod \"packageserver-d55dfcdfc-4wjj8\" (UID: \"0b1aface-d1e4-4d8b-a0e6-08f9032c3554\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-4wjj8" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.209471 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/691b626a-d098-4351-9d7b-8bacf7803327-mountpoint-dir\") pod \"csi-hostpathplugin-n7bs2\" (UID: \"691b626a-d098-4351-9d7b-8bacf7803327\") " pod="hostpath-provisioner/csi-hostpathplugin-n7bs2" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.209541 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/691b626a-d098-4351-9d7b-8bacf7803327-csi-data-dir\") pod \"csi-hostpathplugin-n7bs2\" (UID: \"691b626a-d098-4351-9d7b-8bacf7803327\") " pod="hostpath-provisioner/csi-hostpathplugin-n7bs2" Oct 02 01:49:14 crc kubenswrapper[4885]: E1002 01:49:14.209590 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 01:49:14.70956865 +0000 UTC m=+143.521316139 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gdxhh" (UID: "35380309-be1e-4d61-a2f2-e40a09ff4814") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.209688 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7rqt5\" (UniqueName: \"kubernetes.io/projected/4bc961f3-ae85-4c19-8a57-be0af8e7bd10-kube-api-access-7rqt5\") pod \"collect-profiles-29322825-b7zrt\" (UID: \"4bc961f3-ae85-4c19-8a57-be0af8e7bd10\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322825-b7zrt" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.209716 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3b3cd833-9da7-47ca-979d-464cde398589-config-volume\") pod \"dns-default-6bfsg\" (UID: \"3b3cd833-9da7-47ca-979d-464cde398589\") " pod="openshift-dns/dns-default-6bfsg" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.209733 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b2hdq\" (UniqueName: \"kubernetes.io/projected/8f81f79b-3fa8-430a-a21b-df5fd127e485-kube-api-access-b2hdq\") pod \"service-ca-operator-777779d784-7hk2v\" (UID: \"8f81f79b-3fa8-430a-a21b-df5fd127e485\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-7hk2v" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.209752 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/691b626a-d098-4351-9d7b-8bacf7803327-registration-dir\") pod \"csi-hostpathplugin-n7bs2\" (UID: \"691b626a-d098-4351-9d7b-8bacf7803327\") " pod="hostpath-provisioner/csi-hostpathplugin-n7bs2" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.209780 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/3b3cd833-9da7-47ca-979d-464cde398589-metrics-tls\") pod \"dns-default-6bfsg\" (UID: \"3b3cd833-9da7-47ca-979d-464cde398589\") " pod="openshift-dns/dns-default-6bfsg" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.209803 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8f0ddce-4638-4e1c-b27b-00b991807f23-config\") pod \"kube-apiserver-operator-766d6c64bb-k8r8q\" (UID: \"e8f0ddce-4638-4e1c-b27b-00b991807f23\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-k8r8q" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.209819 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/559ee4a2-19e7-4c8a-a9a2-75e559f5517a-certs\") pod \"machine-config-server-qztpv\" (UID: \"559ee4a2-19e7-4c8a-a9a2-75e559f5517a\") " pod="openshift-machine-config-operator/machine-config-server-qztpv" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.209839 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4bc961f3-ae85-4c19-8a57-be0af8e7bd10-config-volume\") pod \"collect-profiles-29322825-b7zrt\" (UID: \"4bc961f3-ae85-4c19-8a57-be0af8e7bd10\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322825-b7zrt" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.209859 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/1d4a35f8-1702-438b-ad5c-e48eacc16666-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-xjfdv\" (UID: \"1d4a35f8-1702-438b-ad5c-e48eacc16666\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xjfdv" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.210183 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/691b626a-d098-4351-9d7b-8bacf7803327-plugins-dir\") pod \"csi-hostpathplugin-n7bs2\" (UID: \"691b626a-d098-4351-9d7b-8bacf7803327\") " pod="hostpath-provisioner/csi-hostpathplugin-n7bs2" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.210300 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/d2e53137-c638-4ec1-8177-00f07a5fa604-signing-cabundle\") pod \"service-ca-9c57cc56f-c94sm\" (UID: \"d2e53137-c638-4ec1-8177-00f07a5fa604\") " pod="openshift-service-ca/service-ca-9c57cc56f-c94sm" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.211291 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/691b626a-d098-4351-9d7b-8bacf7803327-socket-dir\") pod \"csi-hostpathplugin-n7bs2\" (UID: \"691b626a-d098-4351-9d7b-8bacf7803327\") " pod="hostpath-provisioner/csi-hostpathplugin-n7bs2" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.211372 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/691b626a-d098-4351-9d7b-8bacf7803327-registration-dir\") pod \"csi-hostpathplugin-n7bs2\" (UID: \"691b626a-d098-4351-9d7b-8bacf7803327\") " pod="hostpath-provisioner/csi-hostpathplugin-n7bs2" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.211646 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3b3cd833-9da7-47ca-979d-464cde398589-config-volume\") pod \"dns-default-6bfsg\" (UID: \"3b3cd833-9da7-47ca-979d-464cde398589\") " pod="openshift-dns/dns-default-6bfsg" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.212404 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/0b1aface-d1e4-4d8b-a0e6-08f9032c3554-tmpfs\") pod \"packageserver-d55dfcdfc-4wjj8\" (UID: \"0b1aface-d1e4-4d8b-a0e6-08f9032c3554\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-4wjj8" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.212665 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-qgdp5" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.212732 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4bc961f3-ae85-4c19-8a57-be0af8e7bd10-config-volume\") pod \"collect-profiles-29322825-b7zrt\" (UID: \"4bc961f3-ae85-4c19-8a57-be0af8e7bd10\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322825-b7zrt" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.214398 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f81f79b-3fa8-430a-a21b-df5fd127e485-config\") pod \"service-ca-operator-777779d784-7hk2v\" (UID: \"8f81f79b-3fa8-430a-a21b-df5fd127e485\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-7hk2v" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.215793 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8f0ddce-4638-4e1c-b27b-00b991807f23-config\") pod \"kube-apiserver-operator-766d6c64bb-k8r8q\" (UID: \"e8f0ddce-4638-4e1c-b27b-00b991807f23\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-k8r8q" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.216329 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2dp4w\" (UniqueName: \"kubernetes.io/projected/54328f81-6f98-4576-b16e-4a1ce2bb481b-kube-api-access-2dp4w\") pod \"openshift-controller-manager-operator-756b6f6bc6-98cfh\" (UID: \"54328f81-6f98-4576-b16e-4a1ce2bb481b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-98cfh" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.222899 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s6thl\" (UniqueName: \"kubernetes.io/projected/c5b65af5-6c3f-42db-88a4-b5a73090a390-kube-api-access-s6thl\") pod \"ingress-operator-5b745b69d9-f6gkd\" (UID: \"c5b65af5-6c3f-42db-88a4-b5a73090a390\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-f6gkd" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.226080 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/3b3cd833-9da7-47ca-979d-464cde398589-metrics-tls\") pod \"dns-default-6bfsg\" (UID: \"3b3cd833-9da7-47ca-979d-464cde398589\") " pod="openshift-dns/dns-default-6bfsg" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.226291 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4bc961f3-ae85-4c19-8a57-be0af8e7bd10-secret-volume\") pod \"collect-profiles-29322825-b7zrt\" (UID: \"4bc961f3-ae85-4c19-8a57-be0af8e7bd10\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322825-b7zrt" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.226325 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/1d4a35f8-1702-438b-ad5c-e48eacc16666-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-xjfdv\" (UID: \"1d4a35f8-1702-438b-ad5c-e48eacc16666\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xjfdv" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.226439 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/858faa4e-3646-4a78-b9a8-0fdb8b2a54df-cert\") pod \"ingress-canary-xft2d\" (UID: \"858faa4e-3646-4a78-b9a8-0fdb8b2a54df\") " pod="openshift-ingress-canary/ingress-canary-xft2d" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.226547 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/d2e53137-c638-4ec1-8177-00f07a5fa604-signing-key\") pod \"service-ca-9c57cc56f-c94sm\" (UID: \"d2e53137-c638-4ec1-8177-00f07a5fa604\") " pod="openshift-service-ca/service-ca-9c57cc56f-c94sm" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.226603 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8f81f79b-3fa8-430a-a21b-df5fd127e485-serving-cert\") pod \"service-ca-operator-777779d784-7hk2v\" (UID: \"8f81f79b-3fa8-430a-a21b-df5fd127e485\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-7hk2v" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.226754 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0b1aface-d1e4-4d8b-a0e6-08f9032c3554-apiservice-cert\") pod \"packageserver-d55dfcdfc-4wjj8\" (UID: \"0b1aface-d1e4-4d8b-a0e6-08f9032c3554\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-4wjj8" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.227097 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e8f0ddce-4638-4e1c-b27b-00b991807f23-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-k8r8q\" (UID: \"e8f0ddce-4638-4e1c-b27b-00b991807f23\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-k8r8q" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.227360 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0b1aface-d1e4-4d8b-a0e6-08f9032c3554-webhook-cert\") pod \"packageserver-d55dfcdfc-4wjj8\" (UID: \"0b1aface-d1e4-4d8b-a0e6-08f9032c3554\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-4wjj8" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.227671 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/559ee4a2-19e7-4c8a-a9a2-75e559f5517a-node-bootstrap-token\") pod \"machine-config-server-qztpv\" (UID: \"559ee4a2-19e7-4c8a-a9a2-75e559f5517a\") " pod="openshift-machine-config-operator/machine-config-server-qztpv" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.227766 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/559ee4a2-19e7-4c8a-a9a2-75e559f5517a-certs\") pod \"machine-config-server-qztpv\" (UID: \"559ee4a2-19e7-4c8a-a9a2-75e559f5517a\") " pod="openshift-machine-config-operator/machine-config-server-qztpv" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.232363 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ns269\" (UniqueName: \"kubernetes.io/projected/a2bf9cd6-f368-4eb0-ac04-cd818c43c469-kube-api-access-ns269\") pod \"catalog-operator-68c6474976-kgxn6\" (UID: \"a2bf9cd6-f368-4eb0-ac04-cd818c43c469\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-kgxn6" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.244331 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fg475\" (UniqueName: \"kubernetes.io/projected/f0d19318-e77c-40d4-89b6-da97f7b1f91a-kube-api-access-fg475\") pod \"router-default-5444994796-zxz72\" (UID: \"f0d19318-e77c-40d4-89b6-da97f7b1f91a\") " pod="openshift-ingress/router-default-5444994796-zxz72" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.269350 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/35380309-be1e-4d61-a2f2-e40a09ff4814-bound-sa-token\") pod \"image-registry-697d97f7c8-gdxhh\" (UID: \"35380309-be1e-4d61-a2f2-e40a09ff4814\") " pod="openshift-image-registry/image-registry-697d97f7c8-gdxhh" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.288093 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nc9tk\" (UniqueName: \"kubernetes.io/projected/aa89bc07-6d38-40ba-8e90-9ecae726246d-kube-api-access-nc9tk\") pod \"marketplace-operator-79b997595-z4rlr\" (UID: \"aa89bc07-6d38-40ba-8e90-9ecae726246d\") " pod="openshift-marketplace/marketplace-operator-79b997595-z4rlr" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.290153 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-kgxn6" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.313429 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:49:14 crc kubenswrapper[4885]: E1002 01:49:14.313837 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:49:14.813820127 +0000 UTC m=+143.625567526 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.313967 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/941951e3-26bd-4bb1-a5ae-a7f90813666a-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-66jkw\" (UID: \"941951e3-26bd-4bb1-a5ae-a7f90813666a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-66jkw" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.318099 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-hg5rx"] Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.321056 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-fxhqt"] Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.335633 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cc4tf\" (UniqueName: \"kubernetes.io/projected/96d8ec00-6314-4955-97bd-2d00f28b6f19-kube-api-access-cc4tf\") pod \"kube-storage-version-migrator-operator-b67b599dd-xd4hd\" (UID: \"96d8ec00-6314-4955-97bd-2d00f28b6f19\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-xd4hd" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.352025 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qbmxg\" (UniqueName: \"kubernetes.io/projected/a86af5ef-01e3-47e4-b324-208537e5ec71-kube-api-access-qbmxg\") pod \"console-f9d7485db-kckm8\" (UID: \"a86af5ef-01e3-47e4-b324-208537e5ec71\") " pod="openshift-console/console-f9d7485db-kckm8" Oct 02 01:49:14 crc kubenswrapper[4885]: W1002 01:49:14.353115 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf0062618_a397_45c0_a4f5_df3ebe244228.slice/crio-1685784ab80a03c4ed851c34238d6ffc240d5d59698879302ce1cf3a908e6b96 WatchSource:0}: Error finding container 1685784ab80a03c4ed851c34238d6ffc240d5d59698879302ce1cf3a908e6b96: Status 404 returned error can't find the container with id 1685784ab80a03c4ed851c34238d6ffc240d5d59698879302ce1cf3a908e6b96 Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.358794 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-kckm8" Oct 02 01:49:14 crc kubenswrapper[4885]: W1002 01:49:14.359756 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc57e08c1_826c_4332_8a9e_50f419231d73.slice/crio-e428e3d0ff468f63d405d9a5298fbcdac93584bdb9b26dc072176ba47d2680d4 WatchSource:0}: Error finding container e428e3d0ff468f63d405d9a5298fbcdac93584bdb9b26dc072176ba47d2680d4: Status 404 returned error can't find the container with id e428e3d0ff468f63d405d9a5298fbcdac93584bdb9b26dc072176ba47d2680d4 Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.375228 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c5b65af5-6c3f-42db-88a4-b5a73090a390-bound-sa-token\") pod \"ingress-operator-5b745b69d9-f6gkd\" (UID: \"c5b65af5-6c3f-42db-88a4-b5a73090a390\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-f6gkd" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.397232 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-98cfh" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.421271 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gdxhh\" (UID: \"35380309-be1e-4d61-a2f2-e40a09ff4814\") " pod="openshift-image-registry/image-registry-697d97f7c8-gdxhh" Oct 02 01:49:14 crc kubenswrapper[4885]: E1002 01:49:14.421877 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 01:49:14.921865427 +0000 UTC m=+143.733612826 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gdxhh" (UID: "35380309-be1e-4d61-a2f2-e40a09ff4814") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.423442 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5cp9p\" (UniqueName: \"kubernetes.io/projected/408ea460-05a8-4c98-a372-8eb9b3e43334-kube-api-access-5cp9p\") pod \"migrator-59844c95c7-5lbxn\" (UID: \"408ea460-05a8-4c98-a372-8eb9b3e43334\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-5lbxn" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.427722 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-66jkw" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.441523 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-swrb9"] Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.441786 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-7nhrb" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.449315 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-f6gkd" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.450415 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-8j5cm"] Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.452098 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-5kn8g"] Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.455999 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-98nfr\" (UniqueName: \"kubernetes.io/projected/35380309-be1e-4d61-a2f2-e40a09ff4814-kube-api-access-98nfr\") pod \"image-registry-697d97f7c8-gdxhh\" (UID: \"35380309-be1e-4d61-a2f2-e40a09ff4814\") " pod="openshift-image-registry/image-registry-697d97f7c8-gdxhh" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.460805 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g6qqs\" (UniqueName: \"kubernetes.io/projected/691b626a-d098-4351-9d7b-8bacf7803327-kube-api-access-g6qqs\") pod \"csi-hostpathplugin-n7bs2\" (UID: \"691b626a-d098-4351-9d7b-8bacf7803327\") " pod="hostpath-provisioner/csi-hostpathplugin-n7bs2" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.472562 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fqhmp\" (UniqueName: \"kubernetes.io/projected/1d4a35f8-1702-438b-ad5c-e48eacc16666-kube-api-access-fqhmp\") pod \"package-server-manager-789f6589d5-xjfdv\" (UID: \"1d4a35f8-1702-438b-ad5c-e48eacc16666\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xjfdv" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.499478 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-zxz72" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.500446 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rpjmm\" (UniqueName: \"kubernetes.io/projected/3b3cd833-9da7-47ca-979d-464cde398589-kube-api-access-rpjmm\") pod \"dns-default-6bfsg\" (UID: \"3b3cd833-9da7-47ca-979d-464cde398589\") " pod="openshift-dns/dns-default-6bfsg" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.506929 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7rqt5\" (UniqueName: \"kubernetes.io/projected/4bc961f3-ae85-4c19-8a57-be0af8e7bd10-kube-api-access-7rqt5\") pod \"collect-profiles-29322825-b7zrt\" (UID: \"4bc961f3-ae85-4c19-8a57-be0af8e7bd10\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322825-b7zrt" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.519447 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-xd4hd" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.524279 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:49:14 crc kubenswrapper[4885]: E1002 01:49:14.524704 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:49:15.024684378 +0000 UTC m=+143.836431777 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.529409 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s4bc5\" (UniqueName: \"kubernetes.io/projected/858faa4e-3646-4a78-b9a8-0fdb8b2a54df-kube-api-access-s4bc5\") pod \"ingress-canary-xft2d\" (UID: \"858faa4e-3646-4a78-b9a8-0fdb8b2a54df\") " pod="openshift-ingress-canary/ingress-canary-xft2d" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.531745 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-z4rlr" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.541537 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-5lbxn" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.549513 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e8f0ddce-4638-4e1c-b27b-00b991807f23-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-k8r8q\" (UID: \"e8f0ddce-4638-4e1c-b27b-00b991807f23\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-k8r8q" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.552290 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xjfdv" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.564847 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322825-b7zrt" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.571804 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-k8r8q" Oct 02 01:49:14 crc kubenswrapper[4885]: W1002 01:49:14.574093 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod472cd80c_0597_40b4_babb_89a3fce346b0.slice/crio-1f40cdfb01d2e537b8616335226e19210e8d244ae9c14a8ac37306da4215356e WatchSource:0}: Error finding container 1f40cdfb01d2e537b8616335226e19210e8d244ae9c14a8ac37306da4215356e: Status 404 returned error can't find the container with id 1f40cdfb01d2e537b8616335226e19210e8d244ae9c14a8ac37306da4215356e Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.582002 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cm9ql"] Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.582621 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jjdjh\" (UniqueName: \"kubernetes.io/projected/d2e53137-c638-4ec1-8177-00f07a5fa604-kube-api-access-jjdjh\") pod \"service-ca-9c57cc56f-c94sm\" (UID: \"d2e53137-c638-4ec1-8177-00f07a5fa604\") " pod="openshift-service-ca/service-ca-9c57cc56f-c94sm" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.592550 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-vbm65"] Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.596605 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b2hdq\" (UniqueName: \"kubernetes.io/projected/8f81f79b-3fa8-430a-a21b-df5fd127e485-kube-api-access-b2hdq\") pod \"service-ca-operator-777779d784-7hk2v\" (UID: \"8f81f79b-3fa8-430a-a21b-df5fd127e485\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-7hk2v" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.606795 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-zqxvm"] Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.608152 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-5wldz"] Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.624052 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-n7bs2" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.625364 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gdxhh\" (UID: \"35380309-be1e-4d61-a2f2-e40a09ff4814\") " pod="openshift-image-registry/image-registry-697d97f7c8-gdxhh" Oct 02 01:49:14 crc kubenswrapper[4885]: E1002 01:49:14.625933 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 01:49:15.125804344 +0000 UTC m=+143.937551743 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gdxhh" (UID: "35380309-be1e-4d61-a2f2-e40a09ff4814") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.631450 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-xft2d" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.631565 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sq68j\" (UniqueName: \"kubernetes.io/projected/559ee4a2-19e7-4c8a-a9a2-75e559f5517a-kube-api-access-sq68j\") pod \"machine-config-server-qztpv\" (UID: \"559ee4a2-19e7-4c8a-a9a2-75e559f5517a\") " pod="openshift-machine-config-operator/machine-config-server-qztpv" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.631741 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zhzzc\" (UniqueName: \"kubernetes.io/projected/0b1aface-d1e4-4d8b-a0e6-08f9032c3554-kube-api-access-zhzzc\") pod \"packageserver-d55dfcdfc-4wjj8\" (UID: \"0b1aface-d1e4-4d8b-a0e6-08f9032c3554\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-4wjj8" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.643088 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-6bfsg" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.650506 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-qztpv" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.723118 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-hkwl9"] Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.727373 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:49:14 crc kubenswrapper[4885]: E1002 01:49:14.727788 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:49:15.227773427 +0000 UTC m=+144.039520826 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.738495 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6w2pm"] Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.740074 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5vkgw"] Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.828941 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gdxhh\" (UID: \"35380309-be1e-4d61-a2f2-e40a09ff4814\") " pod="openshift-image-registry/image-registry-697d97f7c8-gdxhh" Oct 02 01:49:14 crc kubenswrapper[4885]: E1002 01:49:14.829921 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 01:49:15.329909696 +0000 UTC m=+144.141657095 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gdxhh" (UID: "35380309-be1e-4d61-a2f2-e40a09ff4814") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.830176 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-kgxn6"] Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.846658 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-4wjj8" Oct 02 01:49:14 crc kubenswrapper[4885]: W1002 01:49:14.854892 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3de23ab3_e037_4163_b744_81ecbb340b13.slice/crio-7a81f6f8ca557489613edb15f2e33b3c6914c95753e80b78d37fc94bc625a28a WatchSource:0}: Error finding container 7a81f6f8ca557489613edb15f2e33b3c6914c95753e80b78d37fc94bc625a28a: Status 404 returned error can't find the container with id 7a81f6f8ca557489613edb15f2e33b3c6914c95753e80b78d37fc94bc625a28a Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.867556 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-c94sm" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.877223 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-qgdp5"] Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.878293 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-7hk2v" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.884698 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-8nb8p"] Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.891061 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dpqkj"] Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.903687 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-swrb9" event={"ID":"472cd80c-0597-40b4-babb-89a3fce346b0","Type":"ContainerStarted","Data":"1f40cdfb01d2e537b8616335226e19210e8d244ae9c14a8ac37306da4215356e"} Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.905247 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hkwl9" event={"ID":"88149eca-e1e7-4dd4-8cdf-170db9460b30","Type":"ContainerStarted","Data":"497f595c8914c6f1001615189b1989d160ef9b7e3fdad1b82f342540314fbee8"} Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.912686 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-hg5rx" event={"ID":"f0062618-a397-45c0-a4f5-df3ebe244228","Type":"ContainerStarted","Data":"1685784ab80a03c4ed851c34238d6ffc240d5d59698879302ce1cf3a908e6b96"} Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.920844 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-5kn8g" event={"ID":"fd852137-cff1-4fae-a921-42be5c511db8","Type":"ContainerStarted","Data":"d23cd93c3507405c8e2769966e1e3312c5ccdf89229819c9eba56baa789bd75b"} Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.933934 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.933946 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-mc59z" event={"ID":"b6e3c195-a333-4c4f-8b75-71edbfcc26c3","Type":"ContainerStarted","Data":"4199acd0bb43e86b61474db1810b27dd89d8d2bd044322f903c037b271703dc7"} Oct 02 01:49:14 crc kubenswrapper[4885]: E1002 01:49:14.934234 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:49:15.434207844 +0000 UTC m=+144.245955243 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.934387 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gdxhh\" (UID: \"35380309-be1e-4d61-a2f2-e40a09ff4814\") " pod="openshift-image-registry/image-registry-697d97f7c8-gdxhh" Oct 02 01:49:14 crc kubenswrapper[4885]: E1002 01:49:14.935102 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 01:49:15.435092973 +0000 UTC m=+144.246840372 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gdxhh" (UID: "35380309-be1e-4d61-a2f2-e40a09ff4814") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.936122 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-mc59z" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.936158 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-mc59z" event={"ID":"b6e3c195-a333-4c4f-8b75-71edbfcc26c3","Type":"ContainerStarted","Data":"240fb0c7b8e78aab6f935fa72be12ed1cd0c0d37c3359c5ce3772b71db33943d"} Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.939003 4885 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-mc59z container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.6:6443/healthz\": dial tcp 10.217.0.6:6443: connect: connection refused" start-of-body= Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.939042 4885 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-mc59z" podUID="b6e3c195-a333-4c4f-8b75-71edbfcc26c3" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.6:6443/healthz\": dial tcp 10.217.0.6:6443: connect: connection refused" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.952916 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-vs6d8" event={"ID":"075528cd-eab2-47bb-8aa0-3d39002fb3d1","Type":"ContainerStarted","Data":"0334e77510540e98449dacc7eddb4b6b9513daf0f37d0d7d39fc07f8b59b004c"} Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.952959 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-vs6d8" event={"ID":"075528cd-eab2-47bb-8aa0-3d39002fb3d1","Type":"ContainerStarted","Data":"9386f1ad17039deb3ae8e6c18844ef75711e75e80749203854287098ebd5f47b"} Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.953819 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-vs6d8" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.955636 4885 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-vs6d8 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" start-of-body= Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.955671 4885 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-vs6d8" podUID="075528cd-eab2-47bb-8aa0-3d39002fb3d1" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.959969 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cm9ql" event={"ID":"4a781c56-3979-49df-a7f1-340c44bcc9b5","Type":"ContainerStarted","Data":"306ecb0c1192d7908c6fcdde9a15e06ea75fcd5ce4e889c3d630dae99f4fde0f"} Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.971181 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-5wldz" event={"ID":"76fb6dce-01dd-4aef-8213-e03be7a5e055","Type":"ContainerStarted","Data":"f62460070b82206f1606a14bb0a8a4110307770c82c8bf9cde6bd1ae6f9095e4"} Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.978254 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-smtqm" event={"ID":"5feb1fb5-73be-4ce5-a6d5-d426e2eada2e","Type":"ContainerStarted","Data":"7eb842d60ba55c1a2abb6a4d78f6a5d0c81a2067613de7d0d50e954055e55d67"} Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.980382 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-g4vj4" event={"ID":"45e90d9b-8520-4b1f-8f16-170787f4c8b6","Type":"ContainerStarted","Data":"84fc3ecbcb85114c98bdfaf8649d09eab5466a1e07ec102855f9e750ec6196ba"} Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.980620 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-g4vj4" event={"ID":"45e90d9b-8520-4b1f-8f16-170787f4c8b6","Type":"ContainerStarted","Data":"7830569e1288c370ed713a70e5eee29f8266a7e5dc69fe9fc372ac47eb9e8061"} Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.981241 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-g4vj4" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.984422 4885 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-g4vj4 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.14:8443/healthz\": dial tcp 10.217.0.14:8443: connect: connection refused" start-of-body= Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.984450 4885 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-g4vj4" podUID="45e90d9b-8520-4b1f-8f16-170787f4c8b6" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.14:8443/healthz\": dial tcp 10.217.0.14:8443: connect: connection refused" Oct 02 01:49:14 crc kubenswrapper[4885]: I1002 01:49:14.989450 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-vbm65" event={"ID":"f257ddf8-bc9c-4670-b8c3-943829b80433","Type":"ContainerStarted","Data":"a71e1b9e17bdce05a71803e4d4416bba40de1947334bd724d2b57e72e26888c5"} Oct 02 01:49:15 crc kubenswrapper[4885]: I1002 01:49:15.005061 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8d6k4" event={"ID":"3a80ea92-4f96-4916-a00d-3f010959194d","Type":"ContainerStarted","Data":"63447ba63aafea7ce22afd38f5cf5414bede209d38eeb9099927c44d69e62344"} Oct 02 01:49:15 crc kubenswrapper[4885]: I1002 01:49:15.005096 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8d6k4" event={"ID":"3a80ea92-4f96-4916-a00d-3f010959194d","Type":"ContainerStarted","Data":"9fddb730f8555fa1259516ac993a51b08ab484a8f2f9a33e6d17886c2345e69b"} Oct 02 01:49:15 crc kubenswrapper[4885]: I1002 01:49:15.007649 4885 generic.go:334] "Generic (PLEG): container finished" podID="243fb6e4-1bf5-4c4a-b471-87f50d5e147d" containerID="3b69c73b1bbe98fe8a3b0853ee49153c3f7c8c088b038f917010327a534ea835" exitCode=0 Oct 02 01:49:15 crc kubenswrapper[4885]: I1002 01:49:15.007731 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-gg9lz" event={"ID":"243fb6e4-1bf5-4c4a-b471-87f50d5e147d","Type":"ContainerDied","Data":"3b69c73b1bbe98fe8a3b0853ee49153c3f7c8c088b038f917010327a534ea835"} Oct 02 01:49:15 crc kubenswrapper[4885]: I1002 01:49:15.007797 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-gg9lz" event={"ID":"243fb6e4-1bf5-4c4a-b471-87f50d5e147d","Type":"ContainerStarted","Data":"641f2729fad63aa56d3c90e73dd11ad778f7e5cd89bfdbaf38a8e5cb53fb4bf8"} Oct 02 01:49:15 crc kubenswrapper[4885]: I1002 01:49:15.011355 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-zxz72" event={"ID":"f0d19318-e77c-40d4-89b6-da97f7b1f91a","Type":"ContainerStarted","Data":"0d2de505364c3a90da568894844f0d9cc2dba4b05de4eec0f80b59ef2c721ede"} Oct 02 01:49:15 crc kubenswrapper[4885]: I1002 01:49:15.023210 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-8j5cm" event={"ID":"c3a7de1e-09f7-41eb-8bf2-3737eee4ae74","Type":"ContainerStarted","Data":"b2e88244d8454c979101b650249f2b045ecfe5ef344f5fb3a2b52a43c2a6a6d1"} Oct 02 01:49:15 crc kubenswrapper[4885]: I1002 01:49:15.024212 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5vkgw" event={"ID":"3de23ab3-e037-4163-b744-81ecbb340b13","Type":"ContainerStarted","Data":"7a81f6f8ca557489613edb15f2e33b3c6914c95753e80b78d37fc94bc625a28a"} Oct 02 01:49:15 crc kubenswrapper[4885]: I1002 01:49:15.032712 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-zqxvm" event={"ID":"028a80a9-9747-4856-a2e8-f721abd82683","Type":"ContainerStarted","Data":"bb5389901da1ed483982bf116d1ae2e8c8107c2bd7b997499e4b4a231416118c"} Oct 02 01:49:15 crc kubenswrapper[4885]: I1002 01:49:15.035290 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:49:15 crc kubenswrapper[4885]: E1002 01:49:15.036637 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:49:15.536606531 +0000 UTC m=+144.348353930 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:49:15 crc kubenswrapper[4885]: I1002 01:49:15.042540 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-fxhqt" event={"ID":"c57e08c1-826c-4332-8a9e-50f419231d73","Type":"ContainerStarted","Data":"fe25834f515b4995f500c168a2d830a6253538ccd6a3098055ae90983cc23355"} Oct 02 01:49:15 crc kubenswrapper[4885]: I1002 01:49:15.042580 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-fxhqt" event={"ID":"c57e08c1-826c-4332-8a9e-50f419231d73","Type":"ContainerStarted","Data":"e428e3d0ff468f63d405d9a5298fbcdac93584bdb9b26dc072176ba47d2680d4"} Oct 02 01:49:15 crc kubenswrapper[4885]: I1002 01:49:15.043129 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-fxhqt" Oct 02 01:49:15 crc kubenswrapper[4885]: I1002 01:49:15.047116 4885 patch_prober.go:28] interesting pod/console-operator-58897d9998-fxhqt container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.28:8443/readyz\": dial tcp 10.217.0.28:8443: connect: connection refused" start-of-body= Oct 02 01:49:15 crc kubenswrapper[4885]: I1002 01:49:15.047154 4885 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-fxhqt" podUID="c57e08c1-826c-4332-8a9e-50f419231d73" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.28:8443/readyz\": dial tcp 10.217.0.28:8443: connect: connection refused" Oct 02 01:49:15 crc kubenswrapper[4885]: I1002 01:49:15.100794 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-kckm8"] Oct 02 01:49:15 crc kubenswrapper[4885]: I1002 01:49:15.125769 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-66jkw"] Oct 02 01:49:15 crc kubenswrapper[4885]: I1002 01:49:15.126606 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-98cfh"] Oct 02 01:49:15 crc kubenswrapper[4885]: I1002 01:49:15.138204 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gdxhh\" (UID: \"35380309-be1e-4d61-a2f2-e40a09ff4814\") " pod="openshift-image-registry/image-registry-697d97f7c8-gdxhh" Oct 02 01:49:15 crc kubenswrapper[4885]: E1002 01:49:15.139827 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 01:49:15.639811395 +0000 UTC m=+144.451558794 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gdxhh" (UID: "35380309-be1e-4d61-a2f2-e40a09ff4814") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:49:15 crc kubenswrapper[4885]: I1002 01:49:15.160529 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-mc59z" podStartSLOduration=122.160515196 podStartE2EDuration="2m2.160515196s" podCreationTimestamp="2025-10-02 01:47:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:49:15.131242908 +0000 UTC m=+143.942990307" watchObservedRunningTime="2025-10-02 01:49:15.160515196 +0000 UTC m=+143.972262595" Oct 02 01:49:15 crc kubenswrapper[4885]: I1002 01:49:15.188797 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-xd4hd"] Oct 02 01:49:15 crc kubenswrapper[4885]: I1002 01:49:15.206791 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-5lbxn"] Oct 02 01:49:15 crc kubenswrapper[4885]: I1002 01:49:15.207826 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322825-b7zrt"] Oct 02 01:49:15 crc kubenswrapper[4885]: I1002 01:49:15.215873 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-7nhrb"] Oct 02 01:49:15 crc kubenswrapper[4885]: I1002 01:49:15.269002 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:49:15 crc kubenswrapper[4885]: E1002 01:49:15.271424 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:49:15.771382787 +0000 UTC m=+144.583130186 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:49:15 crc kubenswrapper[4885]: I1002 01:49:15.273435 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-6bfsg"] Oct 02 01:49:15 crc kubenswrapper[4885]: I1002 01:49:15.275755 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-f6gkd"] Oct 02 01:49:15 crc kubenswrapper[4885]: I1002 01:49:15.283800 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-vs6d8" podStartSLOduration=122.283781009 podStartE2EDuration="2m2.283781009s" podCreationTimestamp="2025-10-02 01:47:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:49:15.280199683 +0000 UTC m=+144.091947082" watchObservedRunningTime="2025-10-02 01:49:15.283781009 +0000 UTC m=+144.095528408" Oct 02 01:49:15 crc kubenswrapper[4885]: I1002 01:49:15.285580 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-n7bs2"] Oct 02 01:49:15 crc kubenswrapper[4885]: I1002 01:49:15.312766 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-4wjj8"] Oct 02 01:49:15 crc kubenswrapper[4885]: I1002 01:49:15.372712 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gdxhh\" (UID: \"35380309-be1e-4d61-a2f2-e40a09ff4814\") " pod="openshift-image-registry/image-registry-697d97f7c8-gdxhh" Oct 02 01:49:15 crc kubenswrapper[4885]: E1002 01:49:15.373174 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 01:49:15.873163245 +0000 UTC m=+144.684910644 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gdxhh" (UID: "35380309-be1e-4d61-a2f2-e40a09ff4814") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:49:15 crc kubenswrapper[4885]: W1002 01:49:15.429386 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod691b626a_d098_4351_9d7b_8bacf7803327.slice/crio-4409e24e8ba11b8af0beb1212e499779bec64f26c57dd507b5c0b33d2d9c8189 WatchSource:0}: Error finding container 4409e24e8ba11b8af0beb1212e499779bec64f26c57dd507b5c0b33d2d9c8189: Status 404 returned error can't find the container with id 4409e24e8ba11b8af0beb1212e499779bec64f26c57dd507b5c0b33d2d9c8189 Oct 02 01:49:15 crc kubenswrapper[4885]: I1002 01:49:15.431765 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-z4rlr"] Oct 02 01:49:15 crc kubenswrapper[4885]: I1002 01:49:15.437428 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-xft2d"] Oct 02 01:49:15 crc kubenswrapper[4885]: I1002 01:49:15.452506 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xjfdv"] Oct 02 01:49:15 crc kubenswrapper[4885]: I1002 01:49:15.483605 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:49:15 crc kubenswrapper[4885]: E1002 01:49:15.483719 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:49:15.983702326 +0000 UTC m=+144.795449725 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:49:15 crc kubenswrapper[4885]: I1002 01:49:15.484130 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gdxhh\" (UID: \"35380309-be1e-4d61-a2f2-e40a09ff4814\") " pod="openshift-image-registry/image-registry-697d97f7c8-gdxhh" Oct 02 01:49:15 crc kubenswrapper[4885]: E1002 01:49:15.484769 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 01:49:15.984754989 +0000 UTC m=+144.796502388 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gdxhh" (UID: "35380309-be1e-4d61-a2f2-e40a09ff4814") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:49:15 crc kubenswrapper[4885]: I1002 01:49:15.497521 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-k8r8q"] Oct 02 01:49:15 crc kubenswrapper[4885]: I1002 01:49:15.538075 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-c94sm"] Oct 02 01:49:15 crc kubenswrapper[4885]: I1002 01:49:15.544036 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-7hk2v"] Oct 02 01:49:15 crc kubenswrapper[4885]: W1002 01:49:15.571961 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd2e53137_c638_4ec1_8177_00f07a5fa604.slice/crio-0c2010355138b55df0270f0bcc52cf8fc6993299a631dd614ebd3347764725c7 WatchSource:0}: Error finding container 0c2010355138b55df0270f0bcc52cf8fc6993299a631dd614ebd3347764725c7: Status 404 returned error can't find the container with id 0c2010355138b55df0270f0bcc52cf8fc6993299a631dd614ebd3347764725c7 Oct 02 01:49:15 crc kubenswrapper[4885]: W1002 01:49:15.581497 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8f81f79b_3fa8_430a_a21b_df5fd127e485.slice/crio-478128a57f7f5969302fb76de4af86220f4afb75c7e0c8b07ad494cf8faf85ba WatchSource:0}: Error finding container 478128a57f7f5969302fb76de4af86220f4afb75c7e0c8b07ad494cf8faf85ba: Status 404 returned error can't find the container with id 478128a57f7f5969302fb76de4af86220f4afb75c7e0c8b07ad494cf8faf85ba Oct 02 01:49:15 crc kubenswrapper[4885]: I1002 01:49:15.584616 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:49:15 crc kubenswrapper[4885]: E1002 01:49:15.584766 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:49:16.084746509 +0000 UTC m=+144.896493908 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:49:15 crc kubenswrapper[4885]: I1002 01:49:15.584887 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gdxhh\" (UID: \"35380309-be1e-4d61-a2f2-e40a09ff4814\") " pod="openshift-image-registry/image-registry-697d97f7c8-gdxhh" Oct 02 01:49:15 crc kubenswrapper[4885]: E1002 01:49:15.585188 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 01:49:16.085181053 +0000 UTC m=+144.896928452 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gdxhh" (UID: "35380309-be1e-4d61-a2f2-e40a09ff4814") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:49:15 crc kubenswrapper[4885]: I1002 01:49:15.685783 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:49:15 crc kubenswrapper[4885]: E1002 01:49:15.686152 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:49:16.186119803 +0000 UTC m=+144.997867202 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:49:15 crc kubenswrapper[4885]: I1002 01:49:15.686254 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gdxhh\" (UID: \"35380309-be1e-4d61-a2f2-e40a09ff4814\") " pod="openshift-image-registry/image-registry-697d97f7c8-gdxhh" Oct 02 01:49:15 crc kubenswrapper[4885]: E1002 01:49:15.686906 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 01:49:16.186895308 +0000 UTC m=+144.998642707 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gdxhh" (UID: "35380309-be1e-4d61-a2f2-e40a09ff4814") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:49:15 crc kubenswrapper[4885]: I1002 01:49:15.788562 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:49:15 crc kubenswrapper[4885]: E1002 01:49:15.788976 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:49:16.288959804 +0000 UTC m=+145.100707203 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:49:15 crc kubenswrapper[4885]: I1002 01:49:15.854609 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-fxhqt" podStartSLOduration=122.854590871 podStartE2EDuration="2m2.854590871s" podCreationTimestamp="2025-10-02 01:47:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:49:15.854449786 +0000 UTC m=+144.666197185" watchObservedRunningTime="2025-10-02 01:49:15.854590871 +0000 UTC m=+144.666338270" Oct 02 01:49:15 crc kubenswrapper[4885]: I1002 01:49:15.891069 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gdxhh\" (UID: \"35380309-be1e-4d61-a2f2-e40a09ff4814\") " pod="openshift-image-registry/image-registry-697d97f7c8-gdxhh" Oct 02 01:49:15 crc kubenswrapper[4885]: E1002 01:49:15.891793 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 01:49:16.391765825 +0000 UTC m=+145.203513224 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gdxhh" (UID: "35380309-be1e-4d61-a2f2-e40a09ff4814") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:49:15 crc kubenswrapper[4885]: I1002 01:49:15.996069 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:49:15 crc kubenswrapper[4885]: E1002 01:49:15.996334 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:49:16.496308741 +0000 UTC m=+145.308056140 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:49:15 crc kubenswrapper[4885]: I1002 01:49:15.996434 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gdxhh\" (UID: \"35380309-be1e-4d61-a2f2-e40a09ff4814\") " pod="openshift-image-registry/image-registry-697d97f7c8-gdxhh" Oct 02 01:49:15 crc kubenswrapper[4885]: E1002 01:49:15.996775 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 01:49:16.496766977 +0000 UTC m=+145.308514376 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gdxhh" (UID: "35380309-be1e-4d61-a2f2-e40a09ff4814") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:49:16 crc kubenswrapper[4885]: I1002 01:49:16.070348 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-5lbxn" event={"ID":"408ea460-05a8-4c98-a372-8eb9b3e43334","Type":"ContainerStarted","Data":"58aa93ca34f0474632a4bca7455b5f31017cd99f84f4ffd8dbceac961569f99d"} Oct 02 01:49:16 crc kubenswrapper[4885]: I1002 01:49:16.070387 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-5lbxn" event={"ID":"408ea460-05a8-4c98-a372-8eb9b3e43334","Type":"ContainerStarted","Data":"f783f3ed480e2c358e0828c62cb84190d54861bee873d790068fb05e3b2d80e7"} Oct 02 01:49:16 crc kubenswrapper[4885]: I1002 01:49:16.071130 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-n7bs2" event={"ID":"691b626a-d098-4351-9d7b-8bacf7803327","Type":"ContainerStarted","Data":"4409e24e8ba11b8af0beb1212e499779bec64f26c57dd507b5c0b33d2d9c8189"} Oct 02 01:49:16 crc kubenswrapper[4885]: I1002 01:49:16.073163 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5vkgw" event={"ID":"3de23ab3-e037-4163-b744-81ecbb340b13","Type":"ContainerStarted","Data":"00112826e428b58648e608288d5b4b2e6ceb7ce801fcbdf762bcba3e26202446"} Oct 02 01:49:16 crc kubenswrapper[4885]: I1002 01:49:16.079030 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-f6gkd" event={"ID":"c5b65af5-6c3f-42db-88a4-b5a73090a390","Type":"ContainerStarted","Data":"e2135f828fc0f737af0e01b6fc1811fd8adcd7d7d6f953265da661eb678e1acb"} Oct 02 01:49:16 crc kubenswrapper[4885]: I1002 01:49:16.079071 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-f6gkd" event={"ID":"c5b65af5-6c3f-42db-88a4-b5a73090a390","Type":"ContainerStarted","Data":"2655dc06878ea1a5e593f9016ac03d824593d6aa6fd143f69f3e4ea5b8c2e05e"} Oct 02 01:49:16 crc kubenswrapper[4885]: I1002 01:49:16.083603 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8d6k4" podStartSLOduration=123.083587469 podStartE2EDuration="2m3.083587469s" podCreationTimestamp="2025-10-02 01:47:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:49:16.083279029 +0000 UTC m=+144.895026428" watchObservedRunningTime="2025-10-02 01:49:16.083587469 +0000 UTC m=+144.895334868" Oct 02 01:49:16 crc kubenswrapper[4885]: I1002 01:49:16.089422 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-zqxvm" event={"ID":"028a80a9-9747-4856-a2e8-f721abd82683","Type":"ContainerStarted","Data":"7c81629052ad2fefe9beba60388bd558577e9747c95c67473e9ba728537ef867"} Oct 02 01:49:16 crc kubenswrapper[4885]: I1002 01:49:16.097152 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:49:16 crc kubenswrapper[4885]: E1002 01:49:16.097445 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:49:16.597322554 +0000 UTC m=+145.409069953 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:49:16 crc kubenswrapper[4885]: I1002 01:49:16.097849 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gdxhh\" (UID: \"35380309-be1e-4d61-a2f2-e40a09ff4814\") " pod="openshift-image-registry/image-registry-697d97f7c8-gdxhh" Oct 02 01:49:16 crc kubenswrapper[4885]: E1002 01:49:16.098494 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 01:49:16.598481572 +0000 UTC m=+145.410228971 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gdxhh" (UID: "35380309-be1e-4d61-a2f2-e40a09ff4814") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:49:16 crc kubenswrapper[4885]: I1002 01:49:16.106754 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cm9ql" event={"ID":"4a781c56-3979-49df-a7f1-340c44bcc9b5","Type":"ContainerStarted","Data":"156a9e55e4042a17c29490f036c5e101731e52413884ae11882b2b058c79ecc3"} Oct 02 01:49:16 crc kubenswrapper[4885]: I1002 01:49:16.107619 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cm9ql" Oct 02 01:49:16 crc kubenswrapper[4885]: I1002 01:49:16.109795 4885 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-cm9ql container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.22:8443/healthz\": dial tcp 10.217.0.22:8443: connect: connection refused" start-of-body= Oct 02 01:49:16 crc kubenswrapper[4885]: I1002 01:49:16.109844 4885 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cm9ql" podUID="4a781c56-3979-49df-a7f1-340c44bcc9b5" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.22:8443/healthz\": dial tcp 10.217.0.22:8443: connect: connection refused" Oct 02 01:49:16 crc kubenswrapper[4885]: I1002 01:49:16.112092 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-98cfh" event={"ID":"54328f81-6f98-4576-b16e-4a1ce2bb481b","Type":"ContainerStarted","Data":"319bb3fea3b6b8e0661ccd03a0b232447691f211dd1579fc0855e7572520dbf6"} Oct 02 01:49:16 crc kubenswrapper[4885]: I1002 01:49:16.112130 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-98cfh" event={"ID":"54328f81-6f98-4576-b16e-4a1ce2bb481b","Type":"ContainerStarted","Data":"8a7f95025feedf3940aa2aeb35e53068630ed50baec5a9404528be960f642076"} Oct 02 01:49:16 crc kubenswrapper[4885]: I1002 01:49:16.118837 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dpqkj" event={"ID":"cc6bb484-0230-46d9-9507-b641d2a5f330","Type":"ContainerStarted","Data":"1d7b658f571105f780a373017394764193b385edd349fbfee379337fd305fabc"} Oct 02 01:49:16 crc kubenswrapper[4885]: I1002 01:49:16.118876 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dpqkj" event={"ID":"cc6bb484-0230-46d9-9507-b641d2a5f330","Type":"ContainerStarted","Data":"cb8eba95551e50c39cb0192e82cfd172e895cdd5694728fb16375ed4a094a3cd"} Oct 02 01:49:16 crc kubenswrapper[4885]: I1002 01:49:16.122322 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-7nhrb" event={"ID":"1b7f5b51-ddb5-4889-9d85-6a4940b8908e","Type":"ContainerStarted","Data":"6cd8ff27654067d1ebedd8e37e5a059998d3d327e2a837c84153f7a9bf3074c2"} Oct 02 01:49:16 crc kubenswrapper[4885]: I1002 01:49:16.127031 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-kckm8" event={"ID":"a86af5ef-01e3-47e4-b324-208537e5ec71","Type":"ContainerStarted","Data":"d39fd1cf35e39882b5385bdcc10cce5b20cc3842cc876b09766cfecd0dde162f"} Oct 02 01:49:16 crc kubenswrapper[4885]: I1002 01:49:16.127059 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-kckm8" event={"ID":"a86af5ef-01e3-47e4-b324-208537e5ec71","Type":"ContainerStarted","Data":"56f1c5035b5dae4325cc2f40ec856cf63cd4d17ebf0505ccdb38b814b3f7224b"} Oct 02 01:49:16 crc kubenswrapper[4885]: I1002 01:49:16.132054 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xjfdv" event={"ID":"1d4a35f8-1702-438b-ad5c-e48eacc16666","Type":"ContainerStarted","Data":"a32d2286e002c554f325997569d8eec1d73b3942479033114ecdad50922d790f"} Oct 02 01:49:16 crc kubenswrapper[4885]: I1002 01:49:16.143157 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-kgxn6" event={"ID":"a2bf9cd6-f368-4eb0-ac04-cd818c43c469","Type":"ContainerStarted","Data":"23a030f0b7f436d769d148b81763af9d86d2f93247dee78d0ab37825ea4be6aa"} Oct 02 01:49:16 crc kubenswrapper[4885]: I1002 01:49:16.143300 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-kgxn6" event={"ID":"a2bf9cd6-f368-4eb0-ac04-cd818c43c469","Type":"ContainerStarted","Data":"2dc91bb2cdbc93199b2ff8b910ed87132b49593a6d92d5363227b27a16c78345"} Oct 02 01:49:16 crc kubenswrapper[4885]: I1002 01:49:16.144128 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-kgxn6" Oct 02 01:49:16 crc kubenswrapper[4885]: I1002 01:49:16.145457 4885 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-kgxn6 container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.39:8443/healthz\": dial tcp 10.217.0.39:8443: connect: connection refused" start-of-body= Oct 02 01:49:16 crc kubenswrapper[4885]: I1002 01:49:16.145504 4885 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-kgxn6" podUID="a2bf9cd6-f368-4eb0-ac04-cd818c43c469" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.39:8443/healthz\": dial tcp 10.217.0.39:8443: connect: connection refused" Oct 02 01:49:16 crc kubenswrapper[4885]: I1002 01:49:16.155913 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-c94sm" event={"ID":"d2e53137-c638-4ec1-8177-00f07a5fa604","Type":"ContainerStarted","Data":"0c2010355138b55df0270f0bcc52cf8fc6993299a631dd614ebd3347764725c7"} Oct 02 01:49:16 crc kubenswrapper[4885]: I1002 01:49:16.163670 4885 generic.go:334] "Generic (PLEG): container finished" podID="88149eca-e1e7-4dd4-8cdf-170db9460b30" containerID="88a4c87e843825f6df2d1c5cfe9d0de5b82db7b56518be89d0a07ea3fb94ead6" exitCode=0 Oct 02 01:49:16 crc kubenswrapper[4885]: I1002 01:49:16.163861 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hkwl9" event={"ID":"88149eca-e1e7-4dd4-8cdf-170db9460b30","Type":"ContainerDied","Data":"88a4c87e843825f6df2d1c5cfe9d0de5b82db7b56518be89d0a07ea3fb94ead6"} Oct 02 01:49:16 crc kubenswrapper[4885]: I1002 01:49:16.165776 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-6bfsg" event={"ID":"3b3cd833-9da7-47ca-979d-464cde398589","Type":"ContainerStarted","Data":"6774249badfb4400d912a0f3126069fe34896cc5f09474f07bd05f111ce78a10"} Oct 02 01:49:16 crc kubenswrapper[4885]: I1002 01:49:16.168868 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-swrb9" event={"ID":"472cd80c-0597-40b4-babb-89a3fce346b0","Type":"ContainerStarted","Data":"220ff2365c9400ba61f1ab4b4d5dc350be5b5eae1c654e0d06a32b132d48fde5"} Oct 02 01:49:16 crc kubenswrapper[4885]: I1002 01:49:16.172163 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-qztpv" event={"ID":"559ee4a2-19e7-4c8a-a9a2-75e559f5517a","Type":"ContainerStarted","Data":"2ba6cba2cb352429ee9d722e42f292eaf5a544ced5b8ad176857959b770147fe"} Oct 02 01:49:16 crc kubenswrapper[4885]: I1002 01:49:16.172202 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-qztpv" event={"ID":"559ee4a2-19e7-4c8a-a9a2-75e559f5517a","Type":"ContainerStarted","Data":"890c7191b2c81e8b04d3af9fe634f2e672e056370632b815a565a4906fcdbd56"} Oct 02 01:49:16 crc kubenswrapper[4885]: I1002 01:49:16.185560 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29322825-b7zrt" event={"ID":"4bc961f3-ae85-4c19-8a57-be0af8e7bd10","Type":"ContainerStarted","Data":"93d41334a68e7f750f8bfdc89a8e5ed4352a32f7c03e51cc62f2ca09320c50cc"} Oct 02 01:49:16 crc kubenswrapper[4885]: I1002 01:49:16.190768 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-66jkw" event={"ID":"941951e3-26bd-4bb1-a5ae-a7f90813666a","Type":"ContainerStarted","Data":"d1013af72b2a0485d1d9a743f706ce6d8bcd869cd7d6b60b3fa37aca70c5f926"} Oct 02 01:49:16 crc kubenswrapper[4885]: I1002 01:49:16.195492 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-7hk2v" event={"ID":"8f81f79b-3fa8-430a-a21b-df5fd127e485","Type":"ContainerStarted","Data":"478128a57f7f5969302fb76de4af86220f4afb75c7e0c8b07ad494cf8faf85ba"} Oct 02 01:49:16 crc kubenswrapper[4885]: I1002 01:49:16.199121 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:49:16 crc kubenswrapper[4885]: E1002 01:49:16.200323 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:49:16.70030761 +0000 UTC m=+145.512055009 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:49:16 crc kubenswrapper[4885]: I1002 01:49:16.206541 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-8j5cm" event={"ID":"c3a7de1e-09f7-41eb-8bf2-3737eee4ae74","Type":"ContainerStarted","Data":"c0de2c022c59979f7181002f960153b078ebe43e470a523974c82d873fbc68e4"} Oct 02 01:49:16 crc kubenswrapper[4885]: I1002 01:49:16.223748 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-k8r8q" event={"ID":"e8f0ddce-4638-4e1c-b27b-00b991807f23","Type":"ContainerStarted","Data":"2770b66d07f25ff0b23a64d4e52ec39b44365fa88ed6accbb7871940aca28574"} Oct 02 01:49:16 crc kubenswrapper[4885]: I1002 01:49:16.235070 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-xft2d" event={"ID":"858faa4e-3646-4a78-b9a8-0fdb8b2a54df","Type":"ContainerStarted","Data":"500dc46aa9a3ac1e15ce6885704f36b4ce9b8a51829d9ac5a22b5b01b4b880c7"} Oct 02 01:49:16 crc kubenswrapper[4885]: I1002 01:49:16.271686 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-5kn8g" event={"ID":"fd852137-cff1-4fae-a921-42be5c511db8","Type":"ContainerStarted","Data":"86c446556fd6f6d3faf644bb8af0d3260d188d2daaef2541095a3b14f687695c"} Oct 02 01:49:16 crc kubenswrapper[4885]: I1002 01:49:16.272574 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-5kn8g" Oct 02 01:49:16 crc kubenswrapper[4885]: I1002 01:49:16.273379 4885 patch_prober.go:28] interesting pod/downloads-7954f5f757-5kn8g container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Oct 02 01:49:16 crc kubenswrapper[4885]: I1002 01:49:16.273420 4885 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-5kn8g" podUID="fd852137-cff1-4fae-a921-42be5c511db8" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Oct 02 01:49:16 crc kubenswrapper[4885]: I1002 01:49:16.274455 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6w2pm" event={"ID":"e925f407-5010-4a6f-bcb5-b555f2f7f8b1","Type":"ContainerStarted","Data":"ae72a6c69e000a9360818cf25f629466632be61f9c8f4eaca7339a205b4a63bc"} Oct 02 01:49:16 crc kubenswrapper[4885]: I1002 01:49:16.274478 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6w2pm" event={"ID":"e925f407-5010-4a6f-bcb5-b555f2f7f8b1","Type":"ContainerStarted","Data":"086ffd80e618ce508b8edc91e3f486ba3304404061902bf49a2e86c67ba68c40"} Oct 02 01:49:16 crc kubenswrapper[4885]: I1002 01:49:16.275746 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-4wjj8" event={"ID":"0b1aface-d1e4-4d8b-a0e6-08f9032c3554","Type":"ContainerStarted","Data":"d2b827e6f4749c44c2906cc6a933647dcf1733f2e430565189f663942180ca5d"} Oct 02 01:49:16 crc kubenswrapper[4885]: I1002 01:49:16.301252 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-5wldz" event={"ID":"76fb6dce-01dd-4aef-8213-e03be7a5e055","Type":"ContainerStarted","Data":"32410a0a9d191e6a6a6ce98a556c849691ae1bee272a03f77d113e7ffbffddea"} Oct 02 01:49:16 crc kubenswrapper[4885]: I1002 01:49:16.303869 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gdxhh\" (UID: \"35380309-be1e-4d61-a2f2-e40a09ff4814\") " pod="openshift-image-registry/image-registry-697d97f7c8-gdxhh" Oct 02 01:49:16 crc kubenswrapper[4885]: E1002 01:49:16.305476 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 01:49:16.805464077 +0000 UTC m=+145.617211476 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gdxhh" (UID: "35380309-be1e-4d61-a2f2-e40a09ff4814") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:49:16 crc kubenswrapper[4885]: I1002 01:49:16.318137 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-zxz72" event={"ID":"f0d19318-e77c-40d4-89b6-da97f7b1f91a","Type":"ContainerStarted","Data":"fabb8a901f80ccb91e00fd0a7a53a80573cfaabfc487562d330d69fde9da59c2"} Oct 02 01:49:16 crc kubenswrapper[4885]: I1002 01:49:16.339735 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-gg9lz" event={"ID":"243fb6e4-1bf5-4c4a-b471-87f50d5e147d","Type":"ContainerStarted","Data":"f1ddeb2715668a88919d60d2d2a62726c0b4c8afaefbccfd39ca978ae8f2f793"} Oct 02 01:49:16 crc kubenswrapper[4885]: I1002 01:49:16.340292 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-gg9lz" Oct 02 01:49:16 crc kubenswrapper[4885]: I1002 01:49:16.343350 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-g4vj4" podStartSLOduration=122.343338633 podStartE2EDuration="2m2.343338633s" podCreationTimestamp="2025-10-02 01:47:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:49:16.307191543 +0000 UTC m=+145.118938942" watchObservedRunningTime="2025-10-02 01:49:16.343338633 +0000 UTC m=+145.155086032" Oct 02 01:49:16 crc kubenswrapper[4885]: I1002 01:49:16.364111 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-smtqm" event={"ID":"5feb1fb5-73be-4ce5-a6d5-d426e2eada2e","Type":"ContainerStarted","Data":"fb36127f81944cb6f1259b1b97f0badc620061e9c5abc2762b50b524a4936917"} Oct 02 01:49:16 crc kubenswrapper[4885]: I1002 01:49:16.368757 4885 generic.go:334] "Generic (PLEG): container finished" podID="f0062618-a397-45c0-a4f5-df3ebe244228" containerID="dad315d51afacdae15dde1c23f262f072de3e54b7555b9271a62c1fdb1bfa649" exitCode=0 Oct 02 01:49:16 crc kubenswrapper[4885]: I1002 01:49:16.368831 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-hg5rx" event={"ID":"f0062618-a397-45c0-a4f5-df3ebe244228","Type":"ContainerDied","Data":"dad315d51afacdae15dde1c23f262f072de3e54b7555b9271a62c1fdb1bfa649"} Oct 02 01:49:16 crc kubenswrapper[4885]: I1002 01:49:16.383187 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-vbm65" event={"ID":"f257ddf8-bc9c-4670-b8c3-943829b80433","Type":"ContainerStarted","Data":"671054026bfb5f86aaea376eb3bfce71b9d3533dfe7a37e4622844f8fcfc57f0"} Oct 02 01:49:16 crc kubenswrapper[4885]: I1002 01:49:16.395721 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-xd4hd" event={"ID":"96d8ec00-6314-4955-97bd-2d00f28b6f19","Type":"ContainerStarted","Data":"e09aa8edd3c81e21be17b702c18a621c2cc35865c3964672d65ba04f472353bc"} Oct 02 01:49:16 crc kubenswrapper[4885]: I1002 01:49:16.413696 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:49:16 crc kubenswrapper[4885]: E1002 01:49:16.415008 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:49:16.914987105 +0000 UTC m=+145.726734504 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:49:16 crc kubenswrapper[4885]: I1002 01:49:16.454244 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8nb8p" event={"ID":"2c17c274-7853-4717-a68e-9fa6467ee199","Type":"ContainerStarted","Data":"881f1d86dc819b4741edef919174dba87b68ae531297e1f5bd9c8eaa848de4aa"} Oct 02 01:49:16 crc kubenswrapper[4885]: I1002 01:49:16.454498 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8nb8p" event={"ID":"2c17c274-7853-4717-a68e-9fa6467ee199","Type":"ContainerStarted","Data":"4121f5b0c48d8f76bd0d41a6da7ff91ff520f0000d7bd1deab5d692c893e8c4b"} Oct 02 01:49:16 crc kubenswrapper[4885]: I1002 01:49:16.464469 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-qgdp5" event={"ID":"607e06a5-d8aa-403e-8976-e48c794bb041","Type":"ContainerStarted","Data":"4e51d66013bf5c14130571bf582c0e9067fce7dbe6658f00e10e27f2986de189"} Oct 02 01:49:16 crc kubenswrapper[4885]: I1002 01:49:16.464518 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-qgdp5" event={"ID":"607e06a5-d8aa-403e-8976-e48c794bb041","Type":"ContainerStarted","Data":"a981cb1662fe9319d368499b4485365400c550c14583618bc4ba1827274b04b0"} Oct 02 01:49:16 crc kubenswrapper[4885]: I1002 01:49:16.475140 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-z4rlr" event={"ID":"aa89bc07-6d38-40ba-8e90-9ecae726246d","Type":"ContainerStarted","Data":"99bdb8b94a6ada7de44e3b7bef3ef8e258caabc772eefeaa6ac77cb8b466dcb4"} Oct 02 01:49:16 crc kubenswrapper[4885]: I1002 01:49:16.481003 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-g4vj4" Oct 02 01:49:16 crc kubenswrapper[4885]: I1002 01:49:16.482278 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-mc59z" Oct 02 01:49:16 crc kubenswrapper[4885]: I1002 01:49:16.494032 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-vs6d8" Oct 02 01:49:16 crc kubenswrapper[4885]: I1002 01:49:16.499693 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-zxz72" Oct 02 01:49:16 crc kubenswrapper[4885]: I1002 01:49:16.504173 4885 patch_prober.go:28] interesting pod/router-default-5444994796-zxz72 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 01:49:16 crc kubenswrapper[4885]: [-]has-synced failed: reason withheld Oct 02 01:49:16 crc kubenswrapper[4885]: [+]process-running ok Oct 02 01:49:16 crc kubenswrapper[4885]: healthz check failed Oct 02 01:49:16 crc kubenswrapper[4885]: I1002 01:49:16.504414 4885 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-zxz72" podUID="f0d19318-e77c-40d4-89b6-da97f7b1f91a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 01:49:16 crc kubenswrapper[4885]: I1002 01:49:16.515211 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gdxhh\" (UID: \"35380309-be1e-4d61-a2f2-e40a09ff4814\") " pod="openshift-image-registry/image-registry-697d97f7c8-gdxhh" Oct 02 01:49:16 crc kubenswrapper[4885]: E1002 01:49:16.517711 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 01:49:17.017696302 +0000 UTC m=+145.829443691 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gdxhh" (UID: "35380309-be1e-4d61-a2f2-e40a09ff4814") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:49:16 crc kubenswrapper[4885]: I1002 01:49:16.543963 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-8j5cm" podStartSLOduration=123.543941182 podStartE2EDuration="2m3.543941182s" podCreationTimestamp="2025-10-02 01:47:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:49:16.54079017 +0000 UTC m=+145.352537579" watchObservedRunningTime="2025-10-02 01:49:16.543941182 +0000 UTC m=+145.355688581" Oct 02 01:49:16 crc kubenswrapper[4885]: I1002 01:49:16.589314 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dpqkj" podStartSLOduration=123.589300302 podStartE2EDuration="2m3.589300302s" podCreationTimestamp="2025-10-02 01:47:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:49:16.58863531 +0000 UTC m=+145.400382709" watchObservedRunningTime="2025-10-02 01:49:16.589300302 +0000 UTC m=+145.401047691" Oct 02 01:49:16 crc kubenswrapper[4885]: I1002 01:49:16.620883 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:49:16 crc kubenswrapper[4885]: E1002 01:49:16.621223 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:49:17.121204615 +0000 UTC m=+145.932952014 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:49:16 crc kubenswrapper[4885]: I1002 01:49:16.655572 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-smtqm" podStartSLOduration=123.655547858 podStartE2EDuration="2m3.655547858s" podCreationTimestamp="2025-10-02 01:47:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:49:16.64820132 +0000 UTC m=+145.459948719" watchObservedRunningTime="2025-10-02 01:49:16.655547858 +0000 UTC m=+145.467295257" Oct 02 01:49:16 crc kubenswrapper[4885]: I1002 01:49:16.722398 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gdxhh\" (UID: \"35380309-be1e-4d61-a2f2-e40a09ff4814\") " pod="openshift-image-registry/image-registry-697d97f7c8-gdxhh" Oct 02 01:49:16 crc kubenswrapper[4885]: E1002 01:49:16.722682 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 01:49:17.222669142 +0000 UTC m=+146.034416541 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gdxhh" (UID: "35380309-be1e-4d61-a2f2-e40a09ff4814") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:49:16 crc kubenswrapper[4885]: I1002 01:49:16.726065 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5vkgw" podStartSLOduration=123.726051492 podStartE2EDuration="2m3.726051492s" podCreationTimestamp="2025-10-02 01:47:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:49:16.724835463 +0000 UTC m=+145.536582862" watchObservedRunningTime="2025-10-02 01:49:16.726051492 +0000 UTC m=+145.537798891" Oct 02 01:49:16 crc kubenswrapper[4885]: I1002 01:49:16.736356 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-zxz72" podStartSLOduration=123.736333155 podStartE2EDuration="2m3.736333155s" podCreationTimestamp="2025-10-02 01:47:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:49:16.673960264 +0000 UTC m=+145.485707663" watchObservedRunningTime="2025-10-02 01:49:16.736333155 +0000 UTC m=+145.548080554" Oct 02 01:49:16 crc kubenswrapper[4885]: I1002 01:49:16.755564 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-qztpv" podStartSLOduration=5.755546897 podStartE2EDuration="5.755546897s" podCreationTimestamp="2025-10-02 01:49:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:49:16.753222912 +0000 UTC m=+145.564970311" watchObservedRunningTime="2025-10-02 01:49:16.755546897 +0000 UTC m=+145.567294296" Oct 02 01:49:16 crc kubenswrapper[4885]: I1002 01:49:16.823009 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:49:16 crc kubenswrapper[4885]: I1002 01:49:16.823372 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-zqxvm" podStartSLOduration=123.823354914 podStartE2EDuration="2m3.823354914s" podCreationTimestamp="2025-10-02 01:47:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:49:16.786993276 +0000 UTC m=+145.598740675" watchObservedRunningTime="2025-10-02 01:49:16.823354914 +0000 UTC m=+145.635102313" Oct 02 01:49:16 crc kubenswrapper[4885]: E1002 01:49:16.823668 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:49:17.323650614 +0000 UTC m=+146.135398013 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:49:16 crc kubenswrapper[4885]: I1002 01:49:16.823672 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-vbm65" podStartSLOduration=123.823667974 podStartE2EDuration="2m3.823667974s" podCreationTimestamp="2025-10-02 01:47:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:49:16.822859228 +0000 UTC m=+145.634606627" watchObservedRunningTime="2025-10-02 01:49:16.823667974 +0000 UTC m=+145.635415373" Oct 02 01:49:16 crc kubenswrapper[4885]: I1002 01:49:16.891114 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-fxhqt" Oct 02 01:49:16 crc kubenswrapper[4885]: I1002 01:49:16.909865 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cm9ql" podStartSLOduration=123.909851556 podStartE2EDuration="2m3.909851556s" podCreationTimestamp="2025-10-02 01:47:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:49:16.884007058 +0000 UTC m=+145.695754447" watchObservedRunningTime="2025-10-02 01:49:16.909851556 +0000 UTC m=+145.721598955" Oct 02 01:49:16 crc kubenswrapper[4885]: I1002 01:49:16.910299 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-5kn8g" podStartSLOduration=123.91029483 podStartE2EDuration="2m3.91029483s" podCreationTimestamp="2025-10-02 01:47:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:49:16.907868742 +0000 UTC m=+145.719616141" watchObservedRunningTime="2025-10-02 01:49:16.91029483 +0000 UTC m=+145.722042229" Oct 02 01:49:16 crc kubenswrapper[4885]: I1002 01:49:16.925837 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gdxhh\" (UID: \"35380309-be1e-4d61-a2f2-e40a09ff4814\") " pod="openshift-image-registry/image-registry-697d97f7c8-gdxhh" Oct 02 01:49:16 crc kubenswrapper[4885]: E1002 01:49:16.926333 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 01:49:17.42632197 +0000 UTC m=+146.238069369 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gdxhh" (UID: "35380309-be1e-4d61-a2f2-e40a09ff4814") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:49:16 crc kubenswrapper[4885]: I1002 01:49:16.943678 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-kgxn6" podStartSLOduration=123.943664232 podStartE2EDuration="2m3.943664232s" podCreationTimestamp="2025-10-02 01:47:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:49:16.94054686 +0000 UTC m=+145.752294259" watchObservedRunningTime="2025-10-02 01:49:16.943664232 +0000 UTC m=+145.755411631" Oct 02 01:49:17 crc kubenswrapper[4885]: I1002 01:49:17.033777 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:49:17 crc kubenswrapper[4885]: E1002 01:49:17.033966 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:49:17.533940366 +0000 UTC m=+146.345687765 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:49:17 crc kubenswrapper[4885]: I1002 01:49:17.034081 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gdxhh\" (UID: \"35380309-be1e-4d61-a2f2-e40a09ff4814\") " pod="openshift-image-registry/image-registry-697d97f7c8-gdxhh" Oct 02 01:49:17 crc kubenswrapper[4885]: E1002 01:49:17.034556 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 01:49:17.534547596 +0000 UTC m=+146.346294995 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gdxhh" (UID: "35380309-be1e-4d61-a2f2-e40a09ff4814") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:49:17 crc kubenswrapper[4885]: I1002 01:49:17.035589 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-gg9lz" podStartSLOduration=124.035571609 podStartE2EDuration="2m4.035571609s" podCreationTimestamp="2025-10-02 01:47:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:49:16.999223251 +0000 UTC m=+145.810970650" watchObservedRunningTime="2025-10-02 01:49:17.035571609 +0000 UTC m=+145.847319008" Oct 02 01:49:17 crc kubenswrapper[4885]: I1002 01:49:17.036106 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-kckm8" podStartSLOduration=124.036100306 podStartE2EDuration="2m4.036100306s" podCreationTimestamp="2025-10-02 01:47:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:49:17.03281778 +0000 UTC m=+145.844565179" watchObservedRunningTime="2025-10-02 01:49:17.036100306 +0000 UTC m=+145.847847705" Oct 02 01:49:17 crc kubenswrapper[4885]: I1002 01:49:17.135452 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:49:17 crc kubenswrapper[4885]: E1002 01:49:17.135613 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:49:17.635586329 +0000 UTC m=+146.447333738 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:49:17 crc kubenswrapper[4885]: I1002 01:49:17.135778 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gdxhh\" (UID: \"35380309-be1e-4d61-a2f2-e40a09ff4814\") " pod="openshift-image-registry/image-registry-697d97f7c8-gdxhh" Oct 02 01:49:17 crc kubenswrapper[4885]: E1002 01:49:17.136160 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 01:49:17.636148007 +0000 UTC m=+146.447895406 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gdxhh" (UID: "35380309-be1e-4d61-a2f2-e40a09ff4814") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:49:17 crc kubenswrapper[4885]: I1002 01:49:17.217747 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-98cfh" podStartSLOduration=124.217731239 podStartE2EDuration="2m4.217731239s" podCreationTimestamp="2025-10-02 01:47:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:49:17.164781455 +0000 UTC m=+145.976528854" watchObservedRunningTime="2025-10-02 01:49:17.217731239 +0000 UTC m=+146.029478639" Oct 02 01:49:17 crc kubenswrapper[4885]: I1002 01:49:17.237445 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:49:17 crc kubenswrapper[4885]: E1002 01:49:17.239671 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:49:17.73964801 +0000 UTC m=+146.551395409 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:49:17 crc kubenswrapper[4885]: I1002 01:49:17.327572 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-z4rlr" podStartSLOduration=124.327557938 podStartE2EDuration="2m4.327557938s" podCreationTimestamp="2025-10-02 01:47:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:49:17.325178211 +0000 UTC m=+146.136925610" watchObservedRunningTime="2025-10-02 01:49:17.327557938 +0000 UTC m=+146.139305337" Oct 02 01:49:17 crc kubenswrapper[4885]: I1002 01:49:17.346162 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gdxhh\" (UID: \"35380309-be1e-4d61-a2f2-e40a09ff4814\") " pod="openshift-image-registry/image-registry-697d97f7c8-gdxhh" Oct 02 01:49:17 crc kubenswrapper[4885]: E1002 01:49:17.346566 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 01:49:17.846552653 +0000 UTC m=+146.658300052 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gdxhh" (UID: "35380309-be1e-4d61-a2f2-e40a09ff4814") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:49:17 crc kubenswrapper[4885]: I1002 01:49:17.447755 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:49:17 crc kubenswrapper[4885]: E1002 01:49:17.448253 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:49:17.948226257 +0000 UTC m=+146.759973656 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:49:17 crc kubenswrapper[4885]: I1002 01:49:17.448602 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gdxhh\" (UID: \"35380309-be1e-4d61-a2f2-e40a09ff4814\") " pod="openshift-image-registry/image-registry-697d97f7c8-gdxhh" Oct 02 01:49:17 crc kubenswrapper[4885]: E1002 01:49:17.449089 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 01:49:17.949081785 +0000 UTC m=+146.760829184 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gdxhh" (UID: "35380309-be1e-4d61-a2f2-e40a09ff4814") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:49:17 crc kubenswrapper[4885]: I1002 01:49:17.478647 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-c94sm" event={"ID":"d2e53137-c638-4ec1-8177-00f07a5fa604","Type":"ContainerStarted","Data":"5617d6cbeed2fcf8045b3385255d153722f946fe0447013b6996eba9173d814d"} Oct 02 01:49:17 crc kubenswrapper[4885]: I1002 01:49:17.482377 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-66jkw" event={"ID":"941951e3-26bd-4bb1-a5ae-a7f90813666a","Type":"ContainerStarted","Data":"3c3c0a1ee30825fc1be2be0417771abdc88736aacb775a74087219402044e4b1"} Oct 02 01:49:17 crc kubenswrapper[4885]: I1002 01:49:17.485037 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-swrb9" event={"ID":"472cd80c-0597-40b4-babb-89a3fce346b0","Type":"ContainerStarted","Data":"8aad2643fc55c4dafd561f78858afaa5e841d1ebc788d79a34a3a13fb3a8f550"} Oct 02 01:49:17 crc kubenswrapper[4885]: I1002 01:49:17.487194 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-k8r8q" event={"ID":"e8f0ddce-4638-4e1c-b27b-00b991807f23","Type":"ContainerStarted","Data":"2f0813ec3bd391476fb4dfdacec6eea675ce5772dde1ced1cea6737b52f3f5ec"} Oct 02 01:49:17 crc kubenswrapper[4885]: I1002 01:49:17.488475 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-xd4hd" event={"ID":"96d8ec00-6314-4955-97bd-2d00f28b6f19","Type":"ContainerStarted","Data":"adb366576214e03b732ec6a438c853ccd2f8fc7b4b15128e704faa0d475794fa"} Oct 02 01:49:17 crc kubenswrapper[4885]: I1002 01:49:17.491535 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8nb8p" event={"ID":"2c17c274-7853-4717-a68e-9fa6467ee199","Type":"ContainerStarted","Data":"51930ed0de9c677fc2ba72de580262a59a769788bd0d5e5a146498fc4ad004ea"} Oct 02 01:49:17 crc kubenswrapper[4885]: I1002 01:49:17.493486 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-qgdp5" event={"ID":"607e06a5-d8aa-403e-8976-e48c794bb041","Type":"ContainerStarted","Data":"b0ec0901bfc0c365a1bbe316126392e91921773fb6a0c21677b69048cb76e603"} Oct 02 01:49:17 crc kubenswrapper[4885]: I1002 01:49:17.495445 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-xft2d" event={"ID":"858faa4e-3646-4a78-b9a8-0fdb8b2a54df","Type":"ContainerStarted","Data":"c92d53c243bc63686c64c25fed7c45c181ebed08098871f2723c7a8754e45b22"} Oct 02 01:49:17 crc kubenswrapper[4885]: I1002 01:49:17.497386 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-vbm65" event={"ID":"f257ddf8-bc9c-4670-b8c3-943829b80433","Type":"ContainerStarted","Data":"d08bfac903cc0621e88a53404ff62a4771f276d6e0230a213b1e7b86b5e9188b"} Oct 02 01:49:17 crc kubenswrapper[4885]: I1002 01:49:17.506777 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-c94sm" podStartSLOduration=123.506762833 podStartE2EDuration="2m3.506762833s" podCreationTimestamp="2025-10-02 01:47:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:49:17.498253208 +0000 UTC m=+146.310000607" watchObservedRunningTime="2025-10-02 01:49:17.506762833 +0000 UTC m=+146.318510232" Oct 02 01:49:17 crc kubenswrapper[4885]: I1002 01:49:17.509433 4885 patch_prober.go:28] interesting pod/router-default-5444994796-zxz72 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 01:49:17 crc kubenswrapper[4885]: [-]has-synced failed: reason withheld Oct 02 01:49:17 crc kubenswrapper[4885]: [+]process-running ok Oct 02 01:49:17 crc kubenswrapper[4885]: healthz check failed Oct 02 01:49:17 crc kubenswrapper[4885]: I1002 01:49:17.509487 4885 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-zxz72" podUID="f0d19318-e77c-40d4-89b6-da97f7b1f91a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 01:49:17 crc kubenswrapper[4885]: I1002 01:49:17.510503 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-6bfsg" event={"ID":"3b3cd833-9da7-47ca-979d-464cde398589","Type":"ContainerStarted","Data":"d90654115b3084915a50bfc0e7ac68b65faa8ef1d299e5c20a6b67b232124225"} Oct 02 01:49:17 crc kubenswrapper[4885]: I1002 01:49:17.510538 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-6bfsg" event={"ID":"3b3cd833-9da7-47ca-979d-464cde398589","Type":"ContainerStarted","Data":"971f1b4b51160db50ce8eadf10ae8c92e968aad6ea25d5e9c168f49eade3b60e"} Oct 02 01:49:17 crc kubenswrapper[4885]: I1002 01:49:17.511116 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-6bfsg" Oct 02 01:49:17 crc kubenswrapper[4885]: I1002 01:49:17.517036 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-4wjj8" event={"ID":"0b1aface-d1e4-4d8b-a0e6-08f9032c3554","Type":"ContainerStarted","Data":"908ce1b0ee8ea3b567b8f426e499df6cd53877841b1890df97239948c4c7fdd1"} Oct 02 01:49:17 crc kubenswrapper[4885]: I1002 01:49:17.517941 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-4wjj8" Oct 02 01:49:17 crc kubenswrapper[4885]: I1002 01:49:17.519148 4885 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-4wjj8 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.24:5443/healthz\": dial tcp 10.217.0.24:5443: connect: connection refused" start-of-body= Oct 02 01:49:17 crc kubenswrapper[4885]: I1002 01:49:17.519187 4885 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-4wjj8" podUID="0b1aface-d1e4-4d8b-a0e6-08f9032c3554" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.24:5443/healthz\": dial tcp 10.217.0.24:5443: connect: connection refused" Oct 02 01:49:17 crc kubenswrapper[4885]: I1002 01:49:17.525242 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-5wldz" event={"ID":"76fb6dce-01dd-4aef-8213-e03be7a5e055","Type":"ContainerStarted","Data":"c86a55978615a66f0c69c8c7c1a0c0677f10c2d016ec3bfb9b3ae7185767c97a"} Oct 02 01:49:17 crc kubenswrapper[4885]: I1002 01:49:17.529559 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-5lbxn" event={"ID":"408ea460-05a8-4c98-a372-8eb9b3e43334","Type":"ContainerStarted","Data":"d1c43db1219c1481b10bd77276b63f3609f345508c98218850885652d930a885"} Oct 02 01:49:17 crc kubenswrapper[4885]: I1002 01:49:17.537978 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6w2pm" event={"ID":"e925f407-5010-4a6f-bcb5-b555f2f7f8b1","Type":"ContainerStarted","Data":"7226038cc155d9191e3aa4f717ee96c58d762e3094d7c2c2129b945cd10b3597"} Oct 02 01:49:17 crc kubenswrapper[4885]: I1002 01:49:17.549930 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:49:17 crc kubenswrapper[4885]: E1002 01:49:17.550171 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:49:18.050148629 +0000 UTC m=+146.861896028 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:49:17 crc kubenswrapper[4885]: I1002 01:49:17.550659 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gdxhh\" (UID: \"35380309-be1e-4d61-a2f2-e40a09ff4814\") " pod="openshift-image-registry/image-registry-697d97f7c8-gdxhh" Oct 02 01:49:17 crc kubenswrapper[4885]: E1002 01:49:17.552471 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 01:49:18.052462813 +0000 UTC m=+146.864210202 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gdxhh" (UID: "35380309-be1e-4d61-a2f2-e40a09ff4814") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:49:17 crc kubenswrapper[4885]: I1002 01:49:17.554199 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29322825-b7zrt" event={"ID":"4bc961f3-ae85-4c19-8a57-be0af8e7bd10","Type":"ContainerStarted","Data":"6930508575fab66820a624c8626849925a65067064512e41a2af37ab86368c07"} Oct 02 01:49:17 crc kubenswrapper[4885]: I1002 01:49:17.559851 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8nb8p" podStartSLOduration=124.559835913 podStartE2EDuration="2m4.559835913s" podCreationTimestamp="2025-10-02 01:47:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:49:17.537418526 +0000 UTC m=+146.349165925" watchObservedRunningTime="2025-10-02 01:49:17.559835913 +0000 UTC m=+146.371583312" Oct 02 01:49:17 crc kubenswrapper[4885]: I1002 01:49:17.559964 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-66jkw" podStartSLOduration=124.559959967 podStartE2EDuration="2m4.559959967s" podCreationTimestamp="2025-10-02 01:47:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:49:17.559669137 +0000 UTC m=+146.371416536" watchObservedRunningTime="2025-10-02 01:49:17.559959967 +0000 UTC m=+146.371707366" Oct 02 01:49:17 crc kubenswrapper[4885]: I1002 01:49:17.577025 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-f6gkd" event={"ID":"c5b65af5-6c3f-42db-88a4-b5a73090a390","Type":"ContainerStarted","Data":"b23699523b931b853380060b94f2d289aedc60df5f0e804053390f6b5d2a1342"} Oct 02 01:49:17 crc kubenswrapper[4885]: I1002 01:49:17.586570 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-z4rlr" event={"ID":"aa89bc07-6d38-40ba-8e90-9ecae726246d","Type":"ContainerStarted","Data":"f04387030d901d22dd8ad8245b4b97815ac740667f7a7642348a4bfb750b45f7"} Oct 02 01:49:17 crc kubenswrapper[4885]: I1002 01:49:17.587465 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-z4rlr" Oct 02 01:49:17 crc kubenswrapper[4885]: I1002 01:49:17.590422 4885 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-z4rlr container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.23:8080/healthz\": dial tcp 10.217.0.23:8080: connect: connection refused" start-of-body= Oct 02 01:49:17 crc kubenswrapper[4885]: I1002 01:49:17.590473 4885 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-z4rlr" podUID="aa89bc07-6d38-40ba-8e90-9ecae726246d" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.23:8080/healthz\": dial tcp 10.217.0.23:8080: connect: connection refused" Oct 02 01:49:17 crc kubenswrapper[4885]: I1002 01:49:17.595363 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-xd4hd" podStartSLOduration=124.595350033 podStartE2EDuration="2m4.595350033s" podCreationTimestamp="2025-10-02 01:47:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:49:17.593639058 +0000 UTC m=+146.405386457" watchObservedRunningTime="2025-10-02 01:49:17.595350033 +0000 UTC m=+146.407097432" Oct 02 01:49:17 crc kubenswrapper[4885]: I1002 01:49:17.605896 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-7nhrb" event={"ID":"1b7f5b51-ddb5-4889-9d85-6a4940b8908e","Type":"ContainerStarted","Data":"44ff27087fba0c84105c14124a832e80526d064def1dce4eca0319d44bc209d4"} Oct 02 01:49:17 crc kubenswrapper[4885]: I1002 01:49:17.608099 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xjfdv" event={"ID":"1d4a35f8-1702-438b-ad5c-e48eacc16666","Type":"ContainerStarted","Data":"bee958a0934dce7b0e1a9b7d37b554df2434b754d346a9132a878566a12ba13c"} Oct 02 01:49:17 crc kubenswrapper[4885]: I1002 01:49:17.608116 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xjfdv" event={"ID":"1d4a35f8-1702-438b-ad5c-e48eacc16666","Type":"ContainerStarted","Data":"d0e6b77ef64bc80fee34e1c751e7f05d416abd3d447df0bbadfe5b5c1b9aeff2"} Oct 02 01:49:17 crc kubenswrapper[4885]: I1002 01:49:17.608471 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xjfdv" Oct 02 01:49:17 crc kubenswrapper[4885]: I1002 01:49:17.610629 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hkwl9" event={"ID":"88149eca-e1e7-4dd4-8cdf-170db9460b30","Type":"ContainerStarted","Data":"ec2359f3eaf3bc5ba1df44d77df747e456983a21d11c2ae16d6c9378df7ded65"} Oct 02 01:49:17 crc kubenswrapper[4885]: I1002 01:49:17.641857 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-k8r8q" podStartSLOduration=124.641843909 podStartE2EDuration="2m4.641843909s" podCreationTimestamp="2025-10-02 01:47:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:49:17.641249981 +0000 UTC m=+146.452997380" watchObservedRunningTime="2025-10-02 01:49:17.641843909 +0000 UTC m=+146.453591308" Oct 02 01:49:17 crc kubenswrapper[4885]: I1002 01:49:17.642823 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-hg5rx" event={"ID":"f0062618-a397-45c0-a4f5-df3ebe244228","Type":"ContainerStarted","Data":"8939ad814c7859ba8f6e7dc8099c445e39f33a91fd6d9146b4c4b02420c5ee91"} Oct 02 01:49:17 crc kubenswrapper[4885]: I1002 01:49:17.645064 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-7hk2v" event={"ID":"8f81f79b-3fa8-430a-a21b-df5fd127e485","Type":"ContainerStarted","Data":"b760c546cf7bedfb13051f4e19ccfc3f0983b7cad892e280402bf94df7200070"} Oct 02 01:49:17 crc kubenswrapper[4885]: I1002 01:49:17.651207 4885 patch_prober.go:28] interesting pod/downloads-7954f5f757-5kn8g container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Oct 02 01:49:17 crc kubenswrapper[4885]: I1002 01:49:17.651268 4885 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-5kn8g" podUID="fd852137-cff1-4fae-a921-42be5c511db8" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Oct 02 01:49:17 crc kubenswrapper[4885]: I1002 01:49:17.651978 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:49:17 crc kubenswrapper[4885]: E1002 01:49:17.652079 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:49:18.15206447 +0000 UTC m=+146.963811869 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:49:17 crc kubenswrapper[4885]: I1002 01:49:17.652539 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gdxhh\" (UID: \"35380309-be1e-4d61-a2f2-e40a09ff4814\") " pod="openshift-image-registry/image-registry-697d97f7c8-gdxhh" Oct 02 01:49:17 crc kubenswrapper[4885]: I1002 01:49:17.655239 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cm9ql" Oct 02 01:49:17 crc kubenswrapper[4885]: E1002 01:49:17.658795 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 01:49:18.158779518 +0000 UTC m=+146.970527007 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gdxhh" (UID: "35380309-be1e-4d61-a2f2-e40a09ff4814") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:49:17 crc kubenswrapper[4885]: I1002 01:49:17.668409 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-qgdp5" podStartSLOduration=124.668391069 podStartE2EDuration="2m4.668391069s" podCreationTimestamp="2025-10-02 01:47:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:49:17.659821792 +0000 UTC m=+146.471569191" watchObservedRunningTime="2025-10-02 01:49:17.668391069 +0000 UTC m=+146.480138468" Oct 02 01:49:17 crc kubenswrapper[4885]: I1002 01:49:17.700728 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-xft2d" podStartSLOduration=6.700712626 podStartE2EDuration="6.700712626s" podCreationTimestamp="2025-10-02 01:49:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:49:17.684608284 +0000 UTC m=+146.496355683" watchObservedRunningTime="2025-10-02 01:49:17.700712626 +0000 UTC m=+146.512460025" Oct 02 01:49:17 crc kubenswrapper[4885]: I1002 01:49:17.706198 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-kgxn6" Oct 02 01:49:17 crc kubenswrapper[4885]: I1002 01:49:17.726514 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-swrb9" podStartSLOduration=124.726496462 podStartE2EDuration="2m4.726496462s" podCreationTimestamp="2025-10-02 01:47:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:49:17.726186231 +0000 UTC m=+146.537933630" watchObservedRunningTime="2025-10-02 01:49:17.726496462 +0000 UTC m=+146.538243861" Oct 02 01:49:17 crc kubenswrapper[4885]: I1002 01:49:17.759247 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:49:17 crc kubenswrapper[4885]: E1002 01:49:17.769299 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:49:18.269236427 +0000 UTC m=+147.080983826 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:49:17 crc kubenswrapper[4885]: I1002 01:49:17.804564 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-6bfsg" podStartSLOduration=6.80454425 podStartE2EDuration="6.80454425s" podCreationTimestamp="2025-10-02 01:49:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:49:17.789989368 +0000 UTC m=+146.601736767" watchObservedRunningTime="2025-10-02 01:49:17.80454425 +0000 UTC m=+146.616291649" Oct 02 01:49:17 crc kubenswrapper[4885]: I1002 01:49:17.810650 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29322825-b7zrt" podStartSLOduration=124.810636767 podStartE2EDuration="2m4.810636767s" podCreationTimestamp="2025-10-02 01:47:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:49:17.80886723 +0000 UTC m=+146.620614629" watchObservedRunningTime="2025-10-02 01:49:17.810636767 +0000 UTC m=+146.622384156" Oct 02 01:49:17 crc kubenswrapper[4885]: I1002 01:49:17.865728 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gdxhh\" (UID: \"35380309-be1e-4d61-a2f2-e40a09ff4814\") " pod="openshift-image-registry/image-registry-697d97f7c8-gdxhh" Oct 02 01:49:17 crc kubenswrapper[4885]: E1002 01:49:17.866056 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 01:49:18.366043963 +0000 UTC m=+147.177791362 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gdxhh" (UID: "35380309-be1e-4d61-a2f2-e40a09ff4814") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:49:17 crc kubenswrapper[4885]: I1002 01:49:17.878034 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-7nhrb" podStartSLOduration=124.87801867 podStartE2EDuration="2m4.87801867s" podCreationTimestamp="2025-10-02 01:47:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:49:17.877362869 +0000 UTC m=+146.689110268" watchObservedRunningTime="2025-10-02 01:49:17.87801867 +0000 UTC m=+146.689766069" Oct 02 01:49:17 crc kubenswrapper[4885]: I1002 01:49:17.900182 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-4wjj8" podStartSLOduration=123.900164858 podStartE2EDuration="2m3.900164858s" podCreationTimestamp="2025-10-02 01:47:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:49:17.897916575 +0000 UTC m=+146.709663974" watchObservedRunningTime="2025-10-02 01:49:17.900164858 +0000 UTC m=+146.711912257" Oct 02 01:49:17 crc kubenswrapper[4885]: I1002 01:49:17.966790 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:49:17 crc kubenswrapper[4885]: E1002 01:49:17.967098 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:49:18.467081835 +0000 UTC m=+147.278829224 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:49:18 crc kubenswrapper[4885]: I1002 01:49:18.038518 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-hg5rx" podStartSLOduration=125.038500679 podStartE2EDuration="2m5.038500679s" podCreationTimestamp="2025-10-02 01:47:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:49:17.981625247 +0000 UTC m=+146.793372646" watchObservedRunningTime="2025-10-02 01:49:18.038500679 +0000 UTC m=+146.850248078" Oct 02 01:49:18 crc kubenswrapper[4885]: I1002 01:49:18.068049 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gdxhh\" (UID: \"35380309-be1e-4d61-a2f2-e40a09ff4814\") " pod="openshift-image-registry/image-registry-697d97f7c8-gdxhh" Oct 02 01:49:18 crc kubenswrapper[4885]: E1002 01:49:18.068401 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 01:49:18.568384907 +0000 UTC m=+147.380132306 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gdxhh" (UID: "35380309-be1e-4d61-a2f2-e40a09ff4814") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:49:18 crc kubenswrapper[4885]: I1002 01:49:18.110778 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-5wldz" podStartSLOduration=125.11075977 podStartE2EDuration="2m5.11075977s" podCreationTimestamp="2025-10-02 01:47:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:49:18.044450232 +0000 UTC m=+146.856197631" watchObservedRunningTime="2025-10-02 01:49:18.11075977 +0000 UTC m=+146.922507179" Oct 02 01:49:18 crc kubenswrapper[4885]: I1002 01:49:18.111496 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hkwl9" podStartSLOduration=125.111491924 podStartE2EDuration="2m5.111491924s" podCreationTimestamp="2025-10-02 01:47:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:49:18.109747907 +0000 UTC m=+146.921495306" watchObservedRunningTime="2025-10-02 01:49:18.111491924 +0000 UTC m=+146.923239323" Oct 02 01:49:18 crc kubenswrapper[4885]: I1002 01:49:18.169480 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:49:18 crc kubenswrapper[4885]: E1002 01:49:18.169671 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:49:18.669630357 +0000 UTC m=+147.481377756 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:49:18 crc kubenswrapper[4885]: I1002 01:49:18.169780 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gdxhh\" (UID: \"35380309-be1e-4d61-a2f2-e40a09ff4814\") " pod="openshift-image-registry/image-registry-697d97f7c8-gdxhh" Oct 02 01:49:18 crc kubenswrapper[4885]: E1002 01:49:18.170059 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 01:49:18.6700509 +0000 UTC m=+147.481798299 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gdxhh" (UID: "35380309-be1e-4d61-a2f2-e40a09ff4814") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:49:18 crc kubenswrapper[4885]: I1002 01:49:18.181127 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-5lbxn" podStartSLOduration=125.181107669 podStartE2EDuration="2m5.181107669s" podCreationTimestamp="2025-10-02 01:47:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:49:18.180277042 +0000 UTC m=+146.992024441" watchObservedRunningTime="2025-10-02 01:49:18.181107669 +0000 UTC m=+146.992855068" Oct 02 01:49:18 crc kubenswrapper[4885]: I1002 01:49:18.229162 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6w2pm" podStartSLOduration=125.229147555 podStartE2EDuration="2m5.229147555s" podCreationTimestamp="2025-10-02 01:47:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:49:18.228134393 +0000 UTC m=+147.039881792" watchObservedRunningTime="2025-10-02 01:49:18.229147555 +0000 UTC m=+147.040894954" Oct 02 01:49:18 crc kubenswrapper[4885]: I1002 01:49:18.270608 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:49:18 crc kubenswrapper[4885]: E1002 01:49:18.270949 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:49:18.770932119 +0000 UTC m=+147.582679518 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:49:18 crc kubenswrapper[4885]: I1002 01:49:18.311099 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-f6gkd" podStartSLOduration=125.311085659 podStartE2EDuration="2m5.311085659s" podCreationTimestamp="2025-10-02 01:47:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:49:18.308859987 +0000 UTC m=+147.120607386" watchObservedRunningTime="2025-10-02 01:49:18.311085659 +0000 UTC m=+147.122833058" Oct 02 01:49:18 crc kubenswrapper[4885]: I1002 01:49:18.372118 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gdxhh\" (UID: \"35380309-be1e-4d61-a2f2-e40a09ff4814\") " pod="openshift-image-registry/image-registry-697d97f7c8-gdxhh" Oct 02 01:49:18 crc kubenswrapper[4885]: E1002 01:49:18.372440 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 01:49:18.872430147 +0000 UTC m=+147.684177546 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gdxhh" (UID: "35380309-be1e-4d61-a2f2-e40a09ff4814") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:49:18 crc kubenswrapper[4885]: I1002 01:49:18.381753 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xjfdv" podStartSLOduration=125.381742279 podStartE2EDuration="2m5.381742279s" podCreationTimestamp="2025-10-02 01:47:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:49:18.35492682 +0000 UTC m=+147.166674209" watchObservedRunningTime="2025-10-02 01:49:18.381742279 +0000 UTC m=+147.193489678" Oct 02 01:49:18 crc kubenswrapper[4885]: I1002 01:49:18.473738 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:49:18 crc kubenswrapper[4885]: E1002 01:49:18.473897 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:49:18.973871653 +0000 UTC m=+147.785619052 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:49:18 crc kubenswrapper[4885]: I1002 01:49:18.474146 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gdxhh\" (UID: \"35380309-be1e-4d61-a2f2-e40a09ff4814\") " pod="openshift-image-registry/image-registry-697d97f7c8-gdxhh" Oct 02 01:49:18 crc kubenswrapper[4885]: E1002 01:49:18.474485 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 01:49:18.974472703 +0000 UTC m=+147.786220102 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gdxhh" (UID: "35380309-be1e-4d61-a2f2-e40a09ff4814") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:49:18 crc kubenswrapper[4885]: I1002 01:49:18.504975 4885 patch_prober.go:28] interesting pod/router-default-5444994796-zxz72 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 01:49:18 crc kubenswrapper[4885]: [-]has-synced failed: reason withheld Oct 02 01:49:18 crc kubenswrapper[4885]: [+]process-running ok Oct 02 01:49:18 crc kubenswrapper[4885]: healthz check failed Oct 02 01:49:18 crc kubenswrapper[4885]: I1002 01:49:18.505020 4885 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-zxz72" podUID="f0d19318-e77c-40d4-89b6-da97f7b1f91a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 01:49:18 crc kubenswrapper[4885]: I1002 01:49:18.574649 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:49:18 crc kubenswrapper[4885]: E1002 01:49:18.574808 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:49:19.074780982 +0000 UTC m=+147.886528381 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:49:18 crc kubenswrapper[4885]: I1002 01:49:18.574879 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gdxhh\" (UID: \"35380309-be1e-4d61-a2f2-e40a09ff4814\") " pod="openshift-image-registry/image-registry-697d97f7c8-gdxhh" Oct 02 01:49:18 crc kubenswrapper[4885]: E1002 01:49:18.575187 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 01:49:19.075175145 +0000 UTC m=+147.886922544 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gdxhh" (UID: "35380309-be1e-4d61-a2f2-e40a09ff4814") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:49:18 crc kubenswrapper[4885]: I1002 01:49:18.651145 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-hg5rx" event={"ID":"f0062618-a397-45c0-a4f5-df3ebe244228","Type":"ContainerStarted","Data":"6299bd87cda83fa1fbe014c34a2b770443ef85d98fe884c56ad7e816d5118f93"} Oct 02 01:49:18 crc kubenswrapper[4885]: I1002 01:49:18.653126 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-n7bs2" event={"ID":"691b626a-d098-4351-9d7b-8bacf7803327","Type":"ContainerStarted","Data":"f1c70fd5fba7d686e1db6d3ef72b14d1d5fcd54313d094672cb65d66e6379ff2"} Oct 02 01:49:18 crc kubenswrapper[4885]: I1002 01:49:18.657841 4885 patch_prober.go:28] interesting pod/downloads-7954f5f757-5kn8g container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Oct 02 01:49:18 crc kubenswrapper[4885]: I1002 01:49:18.657870 4885 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-z4rlr container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.23:8080/healthz\": dial tcp 10.217.0.23:8080: connect: connection refused" start-of-body= Oct 02 01:49:18 crc kubenswrapper[4885]: I1002 01:49:18.657894 4885 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-5kn8g" podUID="fd852137-cff1-4fae-a921-42be5c511db8" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Oct 02 01:49:18 crc kubenswrapper[4885]: I1002 01:49:18.657911 4885 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-z4rlr" podUID="aa89bc07-6d38-40ba-8e90-9ecae726246d" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.23:8080/healthz\": dial tcp 10.217.0.23:8080: connect: connection refused" Oct 02 01:49:18 crc kubenswrapper[4885]: I1002 01:49:18.675631 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:49:18 crc kubenswrapper[4885]: E1002 01:49:18.675815 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:49:19.175790705 +0000 UTC m=+147.987538104 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:49:18 crc kubenswrapper[4885]: I1002 01:49:18.675940 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gdxhh\" (UID: \"35380309-be1e-4d61-a2f2-e40a09ff4814\") " pod="openshift-image-registry/image-registry-697d97f7c8-gdxhh" Oct 02 01:49:18 crc kubenswrapper[4885]: E1002 01:49:18.676182 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 01:49:19.176171827 +0000 UTC m=+147.987919226 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gdxhh" (UID: "35380309-be1e-4d61-a2f2-e40a09ff4814") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:49:18 crc kubenswrapper[4885]: I1002 01:49:18.682843 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-hg5rx" Oct 02 01:49:18 crc kubenswrapper[4885]: I1002 01:49:18.683066 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-hg5rx" Oct 02 01:49:18 crc kubenswrapper[4885]: I1002 01:49:18.683997 4885 patch_prober.go:28] interesting pod/apiserver-76f77b778f-hg5rx container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="Get \"https://10.217.0.9:8443/livez\": dial tcp 10.217.0.9:8443: connect: connection refused" start-of-body= Oct 02 01:49:18 crc kubenswrapper[4885]: I1002 01:49:18.684030 4885 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-hg5rx" podUID="f0062618-a397-45c0-a4f5-df3ebe244228" containerName="openshift-apiserver" probeResult="failure" output="Get \"https://10.217.0.9:8443/livez\": dial tcp 10.217.0.9:8443: connect: connection refused" Oct 02 01:49:18 crc kubenswrapper[4885]: I1002 01:49:18.776444 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:49:18 crc kubenswrapper[4885]: E1002 01:49:18.778101 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:49:19.278085988 +0000 UTC m=+148.089833387 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:49:18 crc kubenswrapper[4885]: I1002 01:49:18.879059 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gdxhh\" (UID: \"35380309-be1e-4d61-a2f2-e40a09ff4814\") " pod="openshift-image-registry/image-registry-697d97f7c8-gdxhh" Oct 02 01:49:18 crc kubenswrapper[4885]: E1002 01:49:18.879408 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 01:49:19.37939221 +0000 UTC m=+148.191139609 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gdxhh" (UID: "35380309-be1e-4d61-a2f2-e40a09ff4814") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:49:18 crc kubenswrapper[4885]: I1002 01:49:18.980709 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:49:18 crc kubenswrapper[4885]: E1002 01:49:18.980853 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:49:19.480823166 +0000 UTC m=+148.292570565 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:49:18 crc kubenswrapper[4885]: I1002 01:49:18.981095 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gdxhh\" (UID: \"35380309-be1e-4d61-a2f2-e40a09ff4814\") " pod="openshift-image-registry/image-registry-697d97f7c8-gdxhh" Oct 02 01:49:18 crc kubenswrapper[4885]: E1002 01:49:18.981424 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 01:49:19.481416866 +0000 UTC m=+148.293164265 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gdxhh" (UID: "35380309-be1e-4d61-a2f2-e40a09ff4814") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:49:19 crc kubenswrapper[4885]: I1002 01:49:19.038748 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hkwl9" Oct 02 01:49:19 crc kubenswrapper[4885]: I1002 01:49:19.038813 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hkwl9" Oct 02 01:49:19 crc kubenswrapper[4885]: I1002 01:49:19.081925 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:49:19 crc kubenswrapper[4885]: E1002 01:49:19.082085 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:49:19.582063666 +0000 UTC m=+148.393811065 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:49:19 crc kubenswrapper[4885]: I1002 01:49:19.082136 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gdxhh\" (UID: \"35380309-be1e-4d61-a2f2-e40a09ff4814\") " pod="openshift-image-registry/image-registry-697d97f7c8-gdxhh" Oct 02 01:49:19 crc kubenswrapper[4885]: E1002 01:49:19.082423 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 01:49:19.582412008 +0000 UTC m=+148.394159407 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gdxhh" (UID: "35380309-be1e-4d61-a2f2-e40a09ff4814") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:49:19 crc kubenswrapper[4885]: I1002 01:49:19.183833 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:49:19 crc kubenswrapper[4885]: E1002 01:49:19.184019 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:49:19.683993498 +0000 UTC m=+148.495740897 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:49:19 crc kubenswrapper[4885]: I1002 01:49:19.184122 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gdxhh\" (UID: \"35380309-be1e-4d61-a2f2-e40a09ff4814\") " pod="openshift-image-registry/image-registry-697d97f7c8-gdxhh" Oct 02 01:49:19 crc kubenswrapper[4885]: E1002 01:49:19.184435 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 01:49:19.684419131 +0000 UTC m=+148.496166530 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gdxhh" (UID: "35380309-be1e-4d61-a2f2-e40a09ff4814") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:49:19 crc kubenswrapper[4885]: I1002 01:49:19.221426 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-gg9lz" Oct 02 01:49:19 crc kubenswrapper[4885]: I1002 01:49:19.285563 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:49:19 crc kubenswrapper[4885]: E1002 01:49:19.285745 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:49:19.785717794 +0000 UTC m=+148.597465193 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:49:19 crc kubenswrapper[4885]: I1002 01:49:19.285899 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gdxhh\" (UID: \"35380309-be1e-4d61-a2f2-e40a09ff4814\") " pod="openshift-image-registry/image-registry-697d97f7c8-gdxhh" Oct 02 01:49:19 crc kubenswrapper[4885]: E1002 01:49:19.286181 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 01:49:19.786172468 +0000 UTC m=+148.597919867 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gdxhh" (UID: "35380309-be1e-4d61-a2f2-e40a09ff4814") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:49:19 crc kubenswrapper[4885]: I1002 01:49:19.298217 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-7hk2v" podStartSLOduration=125.298201677 podStartE2EDuration="2m5.298201677s" podCreationTimestamp="2025-10-02 01:47:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:49:18.395987281 +0000 UTC m=+147.207734680" watchObservedRunningTime="2025-10-02 01:49:19.298201677 +0000 UTC m=+148.109949076" Oct 02 01:49:19 crc kubenswrapper[4885]: I1002 01:49:19.387181 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:49:19 crc kubenswrapper[4885]: E1002 01:49:19.387412 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:49:19.887381567 +0000 UTC m=+148.699128966 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:49:19 crc kubenswrapper[4885]: I1002 01:49:19.387502 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gdxhh\" (UID: \"35380309-be1e-4d61-a2f2-e40a09ff4814\") " pod="openshift-image-registry/image-registry-697d97f7c8-gdxhh" Oct 02 01:49:19 crc kubenswrapper[4885]: E1002 01:49:19.387823 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 01:49:19.887813951 +0000 UTC m=+148.699561350 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gdxhh" (UID: "35380309-be1e-4d61-a2f2-e40a09ff4814") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:49:19 crc kubenswrapper[4885]: I1002 01:49:19.470956 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-x44t6"] Oct 02 01:49:19 crc kubenswrapper[4885]: I1002 01:49:19.471796 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-x44t6" Oct 02 01:49:19 crc kubenswrapper[4885]: I1002 01:49:19.474147 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 02 01:49:19 crc kubenswrapper[4885]: I1002 01:49:19.487369 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-x44t6"] Oct 02 01:49:19 crc kubenswrapper[4885]: I1002 01:49:19.488390 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:49:19 crc kubenswrapper[4885]: E1002 01:49:19.488558 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:49:19.988531664 +0000 UTC m=+148.800279063 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:49:19 crc kubenswrapper[4885]: I1002 01:49:19.488689 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gdxhh\" (UID: \"35380309-be1e-4d61-a2f2-e40a09ff4814\") " pod="openshift-image-registry/image-registry-697d97f7c8-gdxhh" Oct 02 01:49:19 crc kubenswrapper[4885]: E1002 01:49:19.489011 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 01:49:19.988998279 +0000 UTC m=+148.800745668 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gdxhh" (UID: "35380309-be1e-4d61-a2f2-e40a09ff4814") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:49:19 crc kubenswrapper[4885]: I1002 01:49:19.506425 4885 patch_prober.go:28] interesting pod/router-default-5444994796-zxz72 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 01:49:19 crc kubenswrapper[4885]: [-]has-synced failed: reason withheld Oct 02 01:49:19 crc kubenswrapper[4885]: [+]process-running ok Oct 02 01:49:19 crc kubenswrapper[4885]: healthz check failed Oct 02 01:49:19 crc kubenswrapper[4885]: I1002 01:49:19.506487 4885 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-zxz72" podUID="f0d19318-e77c-40d4-89b6-da97f7b1f91a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 01:49:19 crc kubenswrapper[4885]: I1002 01:49:19.589803 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:49:19 crc kubenswrapper[4885]: I1002 01:49:19.589984 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/961bc3c9-b6be-4ece-98f1-971a583f9db8-utilities\") pod \"community-operators-x44t6\" (UID: \"961bc3c9-b6be-4ece-98f1-971a583f9db8\") " pod="openshift-marketplace/community-operators-x44t6" Oct 02 01:49:19 crc kubenswrapper[4885]: I1002 01:49:19.590030 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-svczt\" (UniqueName: \"kubernetes.io/projected/961bc3c9-b6be-4ece-98f1-971a583f9db8-kube-api-access-svczt\") pod \"community-operators-x44t6\" (UID: \"961bc3c9-b6be-4ece-98f1-971a583f9db8\") " pod="openshift-marketplace/community-operators-x44t6" Oct 02 01:49:19 crc kubenswrapper[4885]: I1002 01:49:19.590065 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/961bc3c9-b6be-4ece-98f1-971a583f9db8-catalog-content\") pod \"community-operators-x44t6\" (UID: \"961bc3c9-b6be-4ece-98f1-971a583f9db8\") " pod="openshift-marketplace/community-operators-x44t6" Oct 02 01:49:19 crc kubenswrapper[4885]: E1002 01:49:19.590170 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:49:20.090152576 +0000 UTC m=+148.901899975 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:49:19 crc kubenswrapper[4885]: I1002 01:49:19.652664 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-vb2j4"] Oct 02 01:49:19 crc kubenswrapper[4885]: I1002 01:49:19.653639 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vb2j4" Oct 02 01:49:19 crc kubenswrapper[4885]: I1002 01:49:19.655605 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 02 01:49:19 crc kubenswrapper[4885]: I1002 01:49:19.657796 4885 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-4wjj8 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.24:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 02 01:49:19 crc kubenswrapper[4885]: I1002 01:49:19.657845 4885 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-4wjj8" podUID="0b1aface-d1e4-4d8b-a0e6-08f9032c3554" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.24:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 02 01:49:19 crc kubenswrapper[4885]: I1002 01:49:19.667115 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-n7bs2" event={"ID":"691b626a-d098-4351-9d7b-8bacf7803327","Type":"ContainerStarted","Data":"944dbeb5b5d9a09cafc5037018cdd210e8ea14da73615736a85468ed9dcf6635"} Oct 02 01:49:19 crc kubenswrapper[4885]: I1002 01:49:19.667187 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-n7bs2" event={"ID":"691b626a-d098-4351-9d7b-8bacf7803327","Type":"ContainerStarted","Data":"b724b14783415fa8bd221222fa53a5a6523de74f1464bdb378d283b2e2cf6111"} Oct 02 01:49:19 crc kubenswrapper[4885]: I1002 01:49:19.667793 4885 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-z4rlr container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.23:8080/healthz\": dial tcp 10.217.0.23:8080: connect: connection refused" start-of-body= Oct 02 01:49:19 crc kubenswrapper[4885]: I1002 01:49:19.667838 4885 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-z4rlr" podUID="aa89bc07-6d38-40ba-8e90-9ecae726246d" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.23:8080/healthz\": dial tcp 10.217.0.23:8080: connect: connection refused" Oct 02 01:49:19 crc kubenswrapper[4885]: I1002 01:49:19.675032 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vb2j4"] Oct 02 01:49:19 crc kubenswrapper[4885]: I1002 01:49:19.691252 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-svczt\" (UniqueName: \"kubernetes.io/projected/961bc3c9-b6be-4ece-98f1-971a583f9db8-kube-api-access-svczt\") pod \"community-operators-x44t6\" (UID: \"961bc3c9-b6be-4ece-98f1-971a583f9db8\") " pod="openshift-marketplace/community-operators-x44t6" Oct 02 01:49:19 crc kubenswrapper[4885]: I1002 01:49:19.691302 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gdxhh\" (UID: \"35380309-be1e-4d61-a2f2-e40a09ff4814\") " pod="openshift-image-registry/image-registry-697d97f7c8-gdxhh" Oct 02 01:49:19 crc kubenswrapper[4885]: I1002 01:49:19.691337 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/961bc3c9-b6be-4ece-98f1-971a583f9db8-catalog-content\") pod \"community-operators-x44t6\" (UID: \"961bc3c9-b6be-4ece-98f1-971a583f9db8\") " pod="openshift-marketplace/community-operators-x44t6" Oct 02 01:49:19 crc kubenswrapper[4885]: I1002 01:49:19.691382 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a28f881c-dd5e-4bf8-82ad-2c27b98dcceb-utilities\") pod \"certified-operators-vb2j4\" (UID: \"a28f881c-dd5e-4bf8-82ad-2c27b98dcceb\") " pod="openshift-marketplace/certified-operators-vb2j4" Oct 02 01:49:19 crc kubenswrapper[4885]: I1002 01:49:19.691409 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a28f881c-dd5e-4bf8-82ad-2c27b98dcceb-catalog-content\") pod \"certified-operators-vb2j4\" (UID: \"a28f881c-dd5e-4bf8-82ad-2c27b98dcceb\") " pod="openshift-marketplace/certified-operators-vb2j4" Oct 02 01:49:19 crc kubenswrapper[4885]: I1002 01:49:19.691441 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/961bc3c9-b6be-4ece-98f1-971a583f9db8-utilities\") pod \"community-operators-x44t6\" (UID: \"961bc3c9-b6be-4ece-98f1-971a583f9db8\") " pod="openshift-marketplace/community-operators-x44t6" Oct 02 01:49:19 crc kubenswrapper[4885]: I1002 01:49:19.691478 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cdnq6\" (UniqueName: \"kubernetes.io/projected/a28f881c-dd5e-4bf8-82ad-2c27b98dcceb-kube-api-access-cdnq6\") pod \"certified-operators-vb2j4\" (UID: \"a28f881c-dd5e-4bf8-82ad-2c27b98dcceb\") " pod="openshift-marketplace/certified-operators-vb2j4" Oct 02 01:49:19 crc kubenswrapper[4885]: E1002 01:49:19.691964 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 01:49:20.191953194 +0000 UTC m=+149.003700593 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gdxhh" (UID: "35380309-be1e-4d61-a2f2-e40a09ff4814") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:49:19 crc kubenswrapper[4885]: I1002 01:49:19.692653 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/961bc3c9-b6be-4ece-98f1-971a583f9db8-catalog-content\") pod \"community-operators-x44t6\" (UID: \"961bc3c9-b6be-4ece-98f1-971a583f9db8\") " pod="openshift-marketplace/community-operators-x44t6" Oct 02 01:49:19 crc kubenswrapper[4885]: I1002 01:49:19.692875 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/961bc3c9-b6be-4ece-98f1-971a583f9db8-utilities\") pod \"community-operators-x44t6\" (UID: \"961bc3c9-b6be-4ece-98f1-971a583f9db8\") " pod="openshift-marketplace/community-operators-x44t6" Oct 02 01:49:19 crc kubenswrapper[4885]: I1002 01:49:19.722396 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-4wjj8" Oct 02 01:49:19 crc kubenswrapper[4885]: I1002 01:49:19.730197 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-svczt\" (UniqueName: \"kubernetes.io/projected/961bc3c9-b6be-4ece-98f1-971a583f9db8-kube-api-access-svczt\") pod \"community-operators-x44t6\" (UID: \"961bc3c9-b6be-4ece-98f1-971a583f9db8\") " pod="openshift-marketplace/community-operators-x44t6" Oct 02 01:49:19 crc kubenswrapper[4885]: I1002 01:49:19.782476 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-x44t6" Oct 02 01:49:19 crc kubenswrapper[4885]: I1002 01:49:19.816062 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:49:19 crc kubenswrapper[4885]: I1002 01:49:19.816368 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a28f881c-dd5e-4bf8-82ad-2c27b98dcceb-utilities\") pod \"certified-operators-vb2j4\" (UID: \"a28f881c-dd5e-4bf8-82ad-2c27b98dcceb\") " pod="openshift-marketplace/certified-operators-vb2j4" Oct 02 01:49:19 crc kubenswrapper[4885]: I1002 01:49:19.816415 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a28f881c-dd5e-4bf8-82ad-2c27b98dcceb-catalog-content\") pod \"certified-operators-vb2j4\" (UID: \"a28f881c-dd5e-4bf8-82ad-2c27b98dcceb\") " pod="openshift-marketplace/certified-operators-vb2j4" Oct 02 01:49:19 crc kubenswrapper[4885]: I1002 01:49:19.816585 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cdnq6\" (UniqueName: \"kubernetes.io/projected/a28f881c-dd5e-4bf8-82ad-2c27b98dcceb-kube-api-access-cdnq6\") pod \"certified-operators-vb2j4\" (UID: \"a28f881c-dd5e-4bf8-82ad-2c27b98dcceb\") " pod="openshift-marketplace/certified-operators-vb2j4" Oct 02 01:49:19 crc kubenswrapper[4885]: E1002 01:49:19.817318 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:49:20.317301925 +0000 UTC m=+149.129049324 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:49:19 crc kubenswrapper[4885]: I1002 01:49:19.829620 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a28f881c-dd5e-4bf8-82ad-2c27b98dcceb-utilities\") pod \"certified-operators-vb2j4\" (UID: \"a28f881c-dd5e-4bf8-82ad-2c27b98dcceb\") " pod="openshift-marketplace/certified-operators-vb2j4" Oct 02 01:49:19 crc kubenswrapper[4885]: I1002 01:49:19.833488 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a28f881c-dd5e-4bf8-82ad-2c27b98dcceb-catalog-content\") pod \"certified-operators-vb2j4\" (UID: \"a28f881c-dd5e-4bf8-82ad-2c27b98dcceb\") " pod="openshift-marketplace/certified-operators-vb2j4" Oct 02 01:49:19 crc kubenswrapper[4885]: I1002 01:49:19.860251 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-vgkg6"] Oct 02 01:49:19 crc kubenswrapper[4885]: I1002 01:49:19.861161 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vgkg6" Oct 02 01:49:19 crc kubenswrapper[4885]: I1002 01:49:19.867041 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cdnq6\" (UniqueName: \"kubernetes.io/projected/a28f881c-dd5e-4bf8-82ad-2c27b98dcceb-kube-api-access-cdnq6\") pod \"certified-operators-vb2j4\" (UID: \"a28f881c-dd5e-4bf8-82ad-2c27b98dcceb\") " pod="openshift-marketplace/certified-operators-vb2j4" Oct 02 01:49:19 crc kubenswrapper[4885]: I1002 01:49:19.875580 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vgkg6"] Oct 02 01:49:19 crc kubenswrapper[4885]: I1002 01:49:19.917978 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gdxhh\" (UID: \"35380309-be1e-4d61-a2f2-e40a09ff4814\") " pod="openshift-image-registry/image-registry-697d97f7c8-gdxhh" Oct 02 01:49:19 crc kubenswrapper[4885]: E1002 01:49:19.918242 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 01:49:20.418229544 +0000 UTC m=+149.229976943 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gdxhh" (UID: "35380309-be1e-4d61-a2f2-e40a09ff4814") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:49:19 crc kubenswrapper[4885]: I1002 01:49:19.967548 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vb2j4" Oct 02 01:49:20 crc kubenswrapper[4885]: I1002 01:49:20.018667 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:49:20 crc kubenswrapper[4885]: I1002 01:49:20.018830 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f94th\" (UniqueName: \"kubernetes.io/projected/64a7eacf-552b-4a87-91db-4f6f72ee26b1-kube-api-access-f94th\") pod \"community-operators-vgkg6\" (UID: \"64a7eacf-552b-4a87-91db-4f6f72ee26b1\") " pod="openshift-marketplace/community-operators-vgkg6" Oct 02 01:49:20 crc kubenswrapper[4885]: I1002 01:49:20.018862 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/64a7eacf-552b-4a87-91db-4f6f72ee26b1-catalog-content\") pod \"community-operators-vgkg6\" (UID: \"64a7eacf-552b-4a87-91db-4f6f72ee26b1\") " pod="openshift-marketplace/community-operators-vgkg6" Oct 02 01:49:20 crc kubenswrapper[4885]: I1002 01:49:20.018893 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/64a7eacf-552b-4a87-91db-4f6f72ee26b1-utilities\") pod \"community-operators-vgkg6\" (UID: \"64a7eacf-552b-4a87-91db-4f6f72ee26b1\") " pod="openshift-marketplace/community-operators-vgkg6" Oct 02 01:49:20 crc kubenswrapper[4885]: I1002 01:49:20.018960 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:49:20 crc kubenswrapper[4885]: E1002 01:49:20.020517 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:49:20.520490007 +0000 UTC m=+149.332237406 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:49:20 crc kubenswrapper[4885]: I1002 01:49:20.026805 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:49:20 crc kubenswrapper[4885]: I1002 01:49:20.046710 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-bj65p"] Oct 02 01:49:20 crc kubenswrapper[4885]: I1002 01:49:20.050904 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bj65p" Oct 02 01:49:20 crc kubenswrapper[4885]: I1002 01:49:20.064486 4885 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Oct 02 01:49:20 crc kubenswrapper[4885]: I1002 01:49:20.119784 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f94th\" (UniqueName: \"kubernetes.io/projected/64a7eacf-552b-4a87-91db-4f6f72ee26b1-kube-api-access-f94th\") pod \"community-operators-vgkg6\" (UID: \"64a7eacf-552b-4a87-91db-4f6f72ee26b1\") " pod="openshift-marketplace/community-operators-vgkg6" Oct 02 01:49:20 crc kubenswrapper[4885]: I1002 01:49:20.119828 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:49:20 crc kubenswrapper[4885]: I1002 01:49:20.119846 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/64a7eacf-552b-4a87-91db-4f6f72ee26b1-catalog-content\") pod \"community-operators-vgkg6\" (UID: \"64a7eacf-552b-4a87-91db-4f6f72ee26b1\") " pod="openshift-marketplace/community-operators-vgkg6" Oct 02 01:49:20 crc kubenswrapper[4885]: I1002 01:49:20.119871 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d3cc54c-759c-43a7-b88c-4c5079b42e45-catalog-content\") pod \"certified-operators-bj65p\" (UID: \"7d3cc54c-759c-43a7-b88c-4c5079b42e45\") " pod="openshift-marketplace/certified-operators-bj65p" Oct 02 01:49:20 crc kubenswrapper[4885]: I1002 01:49:20.119890 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/64a7eacf-552b-4a87-91db-4f6f72ee26b1-utilities\") pod \"community-operators-vgkg6\" (UID: \"64a7eacf-552b-4a87-91db-4f6f72ee26b1\") " pod="openshift-marketplace/community-operators-vgkg6" Oct 02 01:49:20 crc kubenswrapper[4885]: I1002 01:49:20.119910 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gdxhh\" (UID: \"35380309-be1e-4d61-a2f2-e40a09ff4814\") " pod="openshift-image-registry/image-registry-697d97f7c8-gdxhh" Oct 02 01:49:20 crc kubenswrapper[4885]: I1002 01:49:20.119938 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:49:20 crc kubenswrapper[4885]: I1002 01:49:20.119953 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d3cc54c-759c-43a7-b88c-4c5079b42e45-utilities\") pod \"certified-operators-bj65p\" (UID: \"7d3cc54c-759c-43a7-b88c-4c5079b42e45\") " pod="openshift-marketplace/certified-operators-bj65p" Oct 02 01:49:20 crc kubenswrapper[4885]: I1002 01:49:20.119971 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:49:20 crc kubenswrapper[4885]: I1002 01:49:20.120000 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hjnzp\" (UniqueName: \"kubernetes.io/projected/7d3cc54c-759c-43a7-b88c-4c5079b42e45-kube-api-access-hjnzp\") pod \"certified-operators-bj65p\" (UID: \"7d3cc54c-759c-43a7-b88c-4c5079b42e45\") " pod="openshift-marketplace/certified-operators-bj65p" Oct 02 01:49:20 crc kubenswrapper[4885]: E1002 01:49:20.121135 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 01:49:20.621115307 +0000 UTC m=+149.432862706 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gdxhh" (UID: "35380309-be1e-4d61-a2f2-e40a09ff4814") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:49:20 crc kubenswrapper[4885]: I1002 01:49:20.121635 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/64a7eacf-552b-4a87-91db-4f6f72ee26b1-catalog-content\") pod \"community-operators-vgkg6\" (UID: \"64a7eacf-552b-4a87-91db-4f6f72ee26b1\") " pod="openshift-marketplace/community-operators-vgkg6" Oct 02 01:49:20 crc kubenswrapper[4885]: I1002 01:49:20.121840 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/64a7eacf-552b-4a87-91db-4f6f72ee26b1-utilities\") pod \"community-operators-vgkg6\" (UID: \"64a7eacf-552b-4a87-91db-4f6f72ee26b1\") " pod="openshift-marketplace/community-operators-vgkg6" Oct 02 01:49:20 crc kubenswrapper[4885]: I1002 01:49:20.122438 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bj65p"] Oct 02 01:49:20 crc kubenswrapper[4885]: I1002 01:49:20.124852 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:49:20 crc kubenswrapper[4885]: I1002 01:49:20.125800 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:49:20 crc kubenswrapper[4885]: I1002 01:49:20.128377 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:49:20 crc kubenswrapper[4885]: I1002 01:49:20.149217 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hkwl9" Oct 02 01:49:20 crc kubenswrapper[4885]: I1002 01:49:20.153069 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f94th\" (UniqueName: \"kubernetes.io/projected/64a7eacf-552b-4a87-91db-4f6f72ee26b1-kube-api-access-f94th\") pod \"community-operators-vgkg6\" (UID: \"64a7eacf-552b-4a87-91db-4f6f72ee26b1\") " pod="openshift-marketplace/community-operators-vgkg6" Oct 02 01:49:20 crc kubenswrapper[4885]: I1002 01:49:20.187045 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vgkg6" Oct 02 01:49:20 crc kubenswrapper[4885]: I1002 01:49:20.221376 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:49:20 crc kubenswrapper[4885]: I1002 01:49:20.221530 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hjnzp\" (UniqueName: \"kubernetes.io/projected/7d3cc54c-759c-43a7-b88c-4c5079b42e45-kube-api-access-hjnzp\") pod \"certified-operators-bj65p\" (UID: \"7d3cc54c-759c-43a7-b88c-4c5079b42e45\") " pod="openshift-marketplace/certified-operators-bj65p" Oct 02 01:49:20 crc kubenswrapper[4885]: E1002 01:49:20.222652 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:49:20.722632746 +0000 UTC m=+149.534380145 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:49:20 crc kubenswrapper[4885]: I1002 01:49:20.222711 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d3cc54c-759c-43a7-b88c-4c5079b42e45-catalog-content\") pod \"certified-operators-bj65p\" (UID: \"7d3cc54c-759c-43a7-b88c-4c5079b42e45\") " pod="openshift-marketplace/certified-operators-bj65p" Oct 02 01:49:20 crc kubenswrapper[4885]: I1002 01:49:20.222739 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gdxhh\" (UID: \"35380309-be1e-4d61-a2f2-e40a09ff4814\") " pod="openshift-image-registry/image-registry-697d97f7c8-gdxhh" Oct 02 01:49:20 crc kubenswrapper[4885]: I1002 01:49:20.222773 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d3cc54c-759c-43a7-b88c-4c5079b42e45-utilities\") pod \"certified-operators-bj65p\" (UID: \"7d3cc54c-759c-43a7-b88c-4c5079b42e45\") " pod="openshift-marketplace/certified-operators-bj65p" Oct 02 01:49:20 crc kubenswrapper[4885]: E1002 01:49:20.224174 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 01:49:20.724166645 +0000 UTC m=+149.535914044 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gdxhh" (UID: "35380309-be1e-4d61-a2f2-e40a09ff4814") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:49:20 crc kubenswrapper[4885]: I1002 01:49:20.224556 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d3cc54c-759c-43a7-b88c-4c5079b42e45-utilities\") pod \"certified-operators-bj65p\" (UID: \"7d3cc54c-759c-43a7-b88c-4c5079b42e45\") " pod="openshift-marketplace/certified-operators-bj65p" Oct 02 01:49:20 crc kubenswrapper[4885]: I1002 01:49:20.224961 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d3cc54c-759c-43a7-b88c-4c5079b42e45-catalog-content\") pod \"certified-operators-bj65p\" (UID: \"7d3cc54c-759c-43a7-b88c-4c5079b42e45\") " pod="openshift-marketplace/certified-operators-bj65p" Oct 02 01:49:20 crc kubenswrapper[4885]: I1002 01:49:20.238639 4885 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-10-02T01:49:20.064516793Z","Handler":null,"Name":""} Oct 02 01:49:20 crc kubenswrapper[4885]: I1002 01:49:20.240988 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hjnzp\" (UniqueName: \"kubernetes.io/projected/7d3cc54c-759c-43a7-b88c-4c5079b42e45-kube-api-access-hjnzp\") pod \"certified-operators-bj65p\" (UID: \"7d3cc54c-759c-43a7-b88c-4c5079b42e45\") " pod="openshift-marketplace/certified-operators-bj65p" Oct 02 01:49:20 crc kubenswrapper[4885]: I1002 01:49:20.254009 4885 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Oct 02 01:49:20 crc kubenswrapper[4885]: I1002 01:49:20.254044 4885 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Oct 02 01:49:20 crc kubenswrapper[4885]: I1002 01:49:20.255767 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 02 01:49:20 crc kubenswrapper[4885]: I1002 01:49:20.256436 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 02 01:49:20 crc kubenswrapper[4885]: I1002 01:49:20.257547 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 02 01:49:20 crc kubenswrapper[4885]: I1002 01:49:20.261621 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Oct 02 01:49:20 crc kubenswrapper[4885]: I1002 01:49:20.261770 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Oct 02 01:49:20 crc kubenswrapper[4885]: I1002 01:49:20.305537 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vb2j4"] Oct 02 01:49:20 crc kubenswrapper[4885]: I1002 01:49:20.323842 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:49:20 crc kubenswrapper[4885]: I1002 01:49:20.324310 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/72a00566-8802-4940-9ed2-852e226dc735-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"72a00566-8802-4940-9ed2-852e226dc735\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 02 01:49:20 crc kubenswrapper[4885]: I1002 01:49:20.324394 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/72a00566-8802-4940-9ed2-852e226dc735-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"72a00566-8802-4940-9ed2-852e226dc735\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 02 01:49:20 crc kubenswrapper[4885]: I1002 01:49:20.327878 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 02 01:49:20 crc kubenswrapper[4885]: I1002 01:49:20.334080 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-x44t6"] Oct 02 01:49:20 crc kubenswrapper[4885]: I1002 01:49:20.377434 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:49:20 crc kubenswrapper[4885]: I1002 01:49:20.395581 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:49:20 crc kubenswrapper[4885]: I1002 01:49:20.408167 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:49:20 crc kubenswrapper[4885]: I1002 01:49:20.417806 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bj65p" Oct 02 01:49:20 crc kubenswrapper[4885]: I1002 01:49:20.425928 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/72a00566-8802-4940-9ed2-852e226dc735-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"72a00566-8802-4940-9ed2-852e226dc735\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 02 01:49:20 crc kubenswrapper[4885]: I1002 01:49:20.425970 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gdxhh\" (UID: \"35380309-be1e-4d61-a2f2-e40a09ff4814\") " pod="openshift-image-registry/image-registry-697d97f7c8-gdxhh" Oct 02 01:49:20 crc kubenswrapper[4885]: I1002 01:49:20.425989 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/72a00566-8802-4940-9ed2-852e226dc735-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"72a00566-8802-4940-9ed2-852e226dc735\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 02 01:49:20 crc kubenswrapper[4885]: I1002 01:49:20.426052 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/72a00566-8802-4940-9ed2-852e226dc735-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"72a00566-8802-4940-9ed2-852e226dc735\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 02 01:49:20 crc kubenswrapper[4885]: I1002 01:49:20.440394 4885 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 02 01:49:20 crc kubenswrapper[4885]: I1002 01:49:20.440440 4885 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gdxhh\" (UID: \"35380309-be1e-4d61-a2f2-e40a09ff4814\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-gdxhh" Oct 02 01:49:20 crc kubenswrapper[4885]: I1002 01:49:20.450601 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/72a00566-8802-4940-9ed2-852e226dc735-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"72a00566-8802-4940-9ed2-852e226dc735\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 02 01:49:20 crc kubenswrapper[4885]: I1002 01:49:20.454070 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vgkg6"] Oct 02 01:49:20 crc kubenswrapper[4885]: W1002 01:49:20.470282 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod64a7eacf_552b_4a87_91db_4f6f72ee26b1.slice/crio-9d0cdc00b0c36d81cf64f9787b38b25b56156fac9ffba5fc555fcaadacdf7333 WatchSource:0}: Error finding container 9d0cdc00b0c36d81cf64f9787b38b25b56156fac9ffba5fc555fcaadacdf7333: Status 404 returned error can't find the container with id 9d0cdc00b0c36d81cf64f9787b38b25b56156fac9ffba5fc555fcaadacdf7333 Oct 02 01:49:20 crc kubenswrapper[4885]: I1002 01:49:20.481839 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gdxhh\" (UID: \"35380309-be1e-4d61-a2f2-e40a09ff4814\") " pod="openshift-image-registry/image-registry-697d97f7c8-gdxhh" Oct 02 01:49:20 crc kubenswrapper[4885]: I1002 01:49:20.515811 4885 patch_prober.go:28] interesting pod/router-default-5444994796-zxz72 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 01:49:20 crc kubenswrapper[4885]: [-]has-synced failed: reason withheld Oct 02 01:49:20 crc kubenswrapper[4885]: [+]process-running ok Oct 02 01:49:20 crc kubenswrapper[4885]: healthz check failed Oct 02 01:49:20 crc kubenswrapper[4885]: I1002 01:49:20.515858 4885 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-zxz72" podUID="f0d19318-e77c-40d4-89b6-da97f7b1f91a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 01:49:20 crc kubenswrapper[4885]: I1002 01:49:20.577775 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 02 01:49:20 crc kubenswrapper[4885]: I1002 01:49:20.707403 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-gdxhh" Oct 02 01:49:20 crc kubenswrapper[4885]: I1002 01:49:20.749254 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vgkg6" event={"ID":"64a7eacf-552b-4a87-91db-4f6f72ee26b1","Type":"ContainerStarted","Data":"9d0cdc00b0c36d81cf64f9787b38b25b56156fac9ffba5fc555fcaadacdf7333"} Oct 02 01:49:20 crc kubenswrapper[4885]: I1002 01:49:20.750540 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vb2j4" event={"ID":"a28f881c-dd5e-4bf8-82ad-2c27b98dcceb","Type":"ContainerStarted","Data":"2efd1cee25f69426399bbace390e0354e4fe87fc61ac719c7e5bd32126074664"} Oct 02 01:49:20 crc kubenswrapper[4885]: I1002 01:49:20.750558 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vb2j4" event={"ID":"a28f881c-dd5e-4bf8-82ad-2c27b98dcceb","Type":"ContainerStarted","Data":"168bb5104de006a02899dc8905b192d43d958565115989616801f8349aa9e56f"} Oct 02 01:49:20 crc kubenswrapper[4885]: I1002 01:49:20.751744 4885 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 01:49:20 crc kubenswrapper[4885]: I1002 01:49:20.765901 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x44t6" event={"ID":"961bc3c9-b6be-4ece-98f1-971a583f9db8","Type":"ContainerStarted","Data":"24036c008472e8e838156422e286e72d19432b6e007bf9d20ad0da5750f5f8d9"} Oct 02 01:49:20 crc kubenswrapper[4885]: I1002 01:49:20.792955 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-n7bs2" event={"ID":"691b626a-d098-4351-9d7b-8bacf7803327","Type":"ContainerStarted","Data":"e0833a6a189aaa2e36ace3f1e4c4222a0a1b1803dab566d36c5608aaa21d5bd9"} Oct 02 01:49:20 crc kubenswrapper[4885]: I1002 01:49:20.816995 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hkwl9" Oct 02 01:49:20 crc kubenswrapper[4885]: I1002 01:49:20.837017 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-n7bs2" podStartSLOduration=9.836998208 podStartE2EDuration="9.836998208s" podCreationTimestamp="2025-10-02 01:49:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:49:20.834319472 +0000 UTC m=+149.646066871" watchObservedRunningTime="2025-10-02 01:49:20.836998208 +0000 UTC m=+149.648745607" Oct 02 01:49:20 crc kubenswrapper[4885]: I1002 01:49:20.856242 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bj65p"] Oct 02 01:49:21 crc kubenswrapper[4885]: W1002 01:49:21.129808 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-28fec75f84881cfb1b91c1ca1f8ba0e8a9064a4bc38c04f5c99e820e56cd9195 WatchSource:0}: Error finding container 28fec75f84881cfb1b91c1ca1f8ba0e8a9064a4bc38c04f5c99e820e56cd9195: Status 404 returned error can't find the container with id 28fec75f84881cfb1b91c1ca1f8ba0e8a9064a4bc38c04f5c99e820e56cd9195 Oct 02 01:49:21 crc kubenswrapper[4885]: W1002 01:49:21.145413 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b6479f0_333b_4a96_9adf_2099afdc2447.slice/crio-ca39b15513c0ca5ad4ad4ec33529f30b12cbd289df7cfa580d9814fb52d20ee3 WatchSource:0}: Error finding container ca39b15513c0ca5ad4ad4ec33529f30b12cbd289df7cfa580d9814fb52d20ee3: Status 404 returned error can't find the container with id ca39b15513c0ca5ad4ad4ec33529f30b12cbd289df7cfa580d9814fb52d20ee3 Oct 02 01:49:21 crc kubenswrapper[4885]: I1002 01:49:21.145959 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-gdxhh"] Oct 02 01:49:21 crc kubenswrapper[4885]: I1002 01:49:21.261927 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 02 01:49:21 crc kubenswrapper[4885]: I1002 01:49:21.503512 4885 patch_prober.go:28] interesting pod/router-default-5444994796-zxz72 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 01:49:21 crc kubenswrapper[4885]: [-]has-synced failed: reason withheld Oct 02 01:49:21 crc kubenswrapper[4885]: [+]process-running ok Oct 02 01:49:21 crc kubenswrapper[4885]: healthz check failed Oct 02 01:49:21 crc kubenswrapper[4885]: I1002 01:49:21.503840 4885 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-zxz72" podUID="f0d19318-e77c-40d4-89b6-da97f7b1f91a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 01:49:21 crc kubenswrapper[4885]: I1002 01:49:21.668920 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-bz9hg"] Oct 02 01:49:21 crc kubenswrapper[4885]: I1002 01:49:21.669851 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bz9hg" Oct 02 01:49:21 crc kubenswrapper[4885]: I1002 01:49:21.671573 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 02 01:49:21 crc kubenswrapper[4885]: I1002 01:49:21.692613 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bz9hg"] Oct 02 01:49:21 crc kubenswrapper[4885]: I1002 01:49:21.754599 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jqcx5\" (UniqueName: \"kubernetes.io/projected/aecca2a3-40b0-4666-a86d-d67ecb580e99-kube-api-access-jqcx5\") pod \"redhat-marketplace-bz9hg\" (UID: \"aecca2a3-40b0-4666-a86d-d67ecb580e99\") " pod="openshift-marketplace/redhat-marketplace-bz9hg" Oct 02 01:49:21 crc kubenswrapper[4885]: I1002 01:49:21.754876 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aecca2a3-40b0-4666-a86d-d67ecb580e99-utilities\") pod \"redhat-marketplace-bz9hg\" (UID: \"aecca2a3-40b0-4666-a86d-d67ecb580e99\") " pod="openshift-marketplace/redhat-marketplace-bz9hg" Oct 02 01:49:21 crc kubenswrapper[4885]: I1002 01:49:21.754982 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aecca2a3-40b0-4666-a86d-d67ecb580e99-catalog-content\") pod \"redhat-marketplace-bz9hg\" (UID: \"aecca2a3-40b0-4666-a86d-d67ecb580e99\") " pod="openshift-marketplace/redhat-marketplace-bz9hg" Oct 02 01:49:21 crc kubenswrapper[4885]: I1002 01:49:21.800777 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"72a00566-8802-4940-9ed2-852e226dc735","Type":"ContainerStarted","Data":"8f7344570a2e047f77a5bd7137edf13910c4cf9c41784d9287bbed0025faf93a"} Oct 02 01:49:21 crc kubenswrapper[4885]: I1002 01:49:21.800819 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"72a00566-8802-4940-9ed2-852e226dc735","Type":"ContainerStarted","Data":"bea8cbcb7ba8fa989c5e153275a2725a6455d3d57e8a26df202346fa79533c94"} Oct 02 01:49:21 crc kubenswrapper[4885]: I1002 01:49:21.804287 4885 generic.go:334] "Generic (PLEG): container finished" podID="961bc3c9-b6be-4ece-98f1-971a583f9db8" containerID="90606a90b0518a25cced263a6b69e97903effb88b507ade600dace77b974890e" exitCode=0 Oct 02 01:49:21 crc kubenswrapper[4885]: I1002 01:49:21.804354 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x44t6" event={"ID":"961bc3c9-b6be-4ece-98f1-971a583f9db8","Type":"ContainerDied","Data":"90606a90b0518a25cced263a6b69e97903effb88b507ade600dace77b974890e"} Oct 02 01:49:21 crc kubenswrapper[4885]: I1002 01:49:21.805939 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"19d18f0c2a8242af967b337680422ab5b0d31c634219584e9dcc8573c6679633"} Oct 02 01:49:21 crc kubenswrapper[4885]: I1002 01:49:21.805963 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"9f9d8a032530067190d5d2202296676267020660c6c4f593994743d7de71f060"} Oct 02 01:49:21 crc kubenswrapper[4885]: I1002 01:49:21.810924 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-gdxhh" event={"ID":"35380309-be1e-4d61-a2f2-e40a09ff4814","Type":"ContainerStarted","Data":"bc1b9e337176117271850d975cd2912508089832ab08adbeed14222591ad3db4"} Oct 02 01:49:21 crc kubenswrapper[4885]: I1002 01:49:21.810948 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-gdxhh" event={"ID":"35380309-be1e-4d61-a2f2-e40a09ff4814","Type":"ContainerStarted","Data":"3daaff641431a9adaf830a1a9cecd02cac7be521fe571b5b42f7b7dc957b2865"} Oct 02 01:49:21 crc kubenswrapper[4885]: I1002 01:49:21.811155 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-gdxhh" Oct 02 01:49:21 crc kubenswrapper[4885]: I1002 01:49:21.812759 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"7651d06c89cd8cb7a41768dbee34dec91462294ed9aad09368b9ac6ecd2edd3d"} Oct 02 01:49:21 crc kubenswrapper[4885]: I1002 01:49:21.812833 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"28fec75f84881cfb1b91c1ca1f8ba0e8a9064a4bc38c04f5c99e820e56cd9195"} Oct 02 01:49:21 crc kubenswrapper[4885]: I1002 01:49:21.815340 4885 generic.go:334] "Generic (PLEG): container finished" podID="64a7eacf-552b-4a87-91db-4f6f72ee26b1" containerID="a6a30ee6e64ab13e1b7f07dd3d2a9163a36b9d194dcf05693ef9ec4f453688b7" exitCode=0 Oct 02 01:49:21 crc kubenswrapper[4885]: I1002 01:49:21.815434 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vgkg6" event={"ID":"64a7eacf-552b-4a87-91db-4f6f72ee26b1","Type":"ContainerDied","Data":"a6a30ee6e64ab13e1b7f07dd3d2a9163a36b9d194dcf05693ef9ec4f453688b7"} Oct 02 01:49:21 crc kubenswrapper[4885]: I1002 01:49:21.818920 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=1.8189014380000001 podStartE2EDuration="1.818901438s" podCreationTimestamp="2025-10-02 01:49:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:49:21.818033639 +0000 UTC m=+150.629781038" watchObservedRunningTime="2025-10-02 01:49:21.818901438 +0000 UTC m=+150.630648867" Oct 02 01:49:21 crc kubenswrapper[4885]: I1002 01:49:21.823981 4885 generic.go:334] "Generic (PLEG): container finished" podID="a28f881c-dd5e-4bf8-82ad-2c27b98dcceb" containerID="2efd1cee25f69426399bbace390e0354e4fe87fc61ac719c7e5bd32126074664" exitCode=0 Oct 02 01:49:21 crc kubenswrapper[4885]: I1002 01:49:21.824067 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vb2j4" event={"ID":"a28f881c-dd5e-4bf8-82ad-2c27b98dcceb","Type":"ContainerDied","Data":"2efd1cee25f69426399bbace390e0354e4fe87fc61ac719c7e5bd32126074664"} Oct 02 01:49:21 crc kubenswrapper[4885]: I1002 01:49:21.826432 4885 generic.go:334] "Generic (PLEG): container finished" podID="7d3cc54c-759c-43a7-b88c-4c5079b42e45" containerID="7487c5f977be380c0eebe25d7111c9a0ab055e691c511a081b77dedca2881541" exitCode=0 Oct 02 01:49:21 crc kubenswrapper[4885]: I1002 01:49:21.826512 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bj65p" event={"ID":"7d3cc54c-759c-43a7-b88c-4c5079b42e45","Type":"ContainerDied","Data":"7487c5f977be380c0eebe25d7111c9a0ab055e691c511a081b77dedca2881541"} Oct 02 01:49:21 crc kubenswrapper[4885]: I1002 01:49:21.826550 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bj65p" event={"ID":"7d3cc54c-759c-43a7-b88c-4c5079b42e45","Type":"ContainerStarted","Data":"7f12910935271cd413d9748a6c302ee159ee903fc91afb8fa8b128b8b7352739"} Oct 02 01:49:21 crc kubenswrapper[4885]: I1002 01:49:21.828905 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"59a6cc4a344934f898a74baf4b077bbc39696ccb9b6d277e0c9807d37a659db5"} Oct 02 01:49:21 crc kubenswrapper[4885]: I1002 01:49:21.828952 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"ca39b15513c0ca5ad4ad4ec33529f30b12cbd289df7cfa580d9814fb52d20ee3"} Oct 02 01:49:21 crc kubenswrapper[4885]: I1002 01:49:21.856431 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jqcx5\" (UniqueName: \"kubernetes.io/projected/aecca2a3-40b0-4666-a86d-d67ecb580e99-kube-api-access-jqcx5\") pod \"redhat-marketplace-bz9hg\" (UID: \"aecca2a3-40b0-4666-a86d-d67ecb580e99\") " pod="openshift-marketplace/redhat-marketplace-bz9hg" Oct 02 01:49:21 crc kubenswrapper[4885]: I1002 01:49:21.856571 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aecca2a3-40b0-4666-a86d-d67ecb580e99-utilities\") pod \"redhat-marketplace-bz9hg\" (UID: \"aecca2a3-40b0-4666-a86d-d67ecb580e99\") " pod="openshift-marketplace/redhat-marketplace-bz9hg" Oct 02 01:49:21 crc kubenswrapper[4885]: I1002 01:49:21.856622 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aecca2a3-40b0-4666-a86d-d67ecb580e99-catalog-content\") pod \"redhat-marketplace-bz9hg\" (UID: \"aecca2a3-40b0-4666-a86d-d67ecb580e99\") " pod="openshift-marketplace/redhat-marketplace-bz9hg" Oct 02 01:49:21 crc kubenswrapper[4885]: I1002 01:49:21.859726 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aecca2a3-40b0-4666-a86d-d67ecb580e99-catalog-content\") pod \"redhat-marketplace-bz9hg\" (UID: \"aecca2a3-40b0-4666-a86d-d67ecb580e99\") " pod="openshift-marketplace/redhat-marketplace-bz9hg" Oct 02 01:49:21 crc kubenswrapper[4885]: I1002 01:49:21.860190 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aecca2a3-40b0-4666-a86d-d67ecb580e99-utilities\") pod \"redhat-marketplace-bz9hg\" (UID: \"aecca2a3-40b0-4666-a86d-d67ecb580e99\") " pod="openshift-marketplace/redhat-marketplace-bz9hg" Oct 02 01:49:21 crc kubenswrapper[4885]: I1002 01:49:21.870486 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-gdxhh" podStartSLOduration=128.870469138 podStartE2EDuration="2m8.870469138s" podCreationTimestamp="2025-10-02 01:47:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:49:21.847054139 +0000 UTC m=+150.658801578" watchObservedRunningTime="2025-10-02 01:49:21.870469138 +0000 UTC m=+150.682216537" Oct 02 01:49:21 crc kubenswrapper[4885]: I1002 01:49:21.877959 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jqcx5\" (UniqueName: \"kubernetes.io/projected/aecca2a3-40b0-4666-a86d-d67ecb580e99-kube-api-access-jqcx5\") pod \"redhat-marketplace-bz9hg\" (UID: \"aecca2a3-40b0-4666-a86d-d67ecb580e99\") " pod="openshift-marketplace/redhat-marketplace-bz9hg" Oct 02 01:49:21 crc kubenswrapper[4885]: I1002 01:49:21.992531 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bz9hg" Oct 02 01:49:22 crc kubenswrapper[4885]: I1002 01:49:22.040305 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-4p78s"] Oct 02 01:49:22 crc kubenswrapper[4885]: I1002 01:49:22.051883 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4p78s" Oct 02 01:49:22 crc kubenswrapper[4885]: I1002 01:49:22.071420 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Oct 02 01:49:22 crc kubenswrapper[4885]: I1002 01:49:22.072221 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4p78s"] Oct 02 01:49:22 crc kubenswrapper[4885]: I1002 01:49:22.179574 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/adaf0543-043d-48b8-829f-928df3bf6e90-catalog-content\") pod \"redhat-marketplace-4p78s\" (UID: \"adaf0543-043d-48b8-829f-928df3bf6e90\") " pod="openshift-marketplace/redhat-marketplace-4p78s" Oct 02 01:49:22 crc kubenswrapper[4885]: I1002 01:49:22.180358 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s4hm5\" (UniqueName: \"kubernetes.io/projected/adaf0543-043d-48b8-829f-928df3bf6e90-kube-api-access-s4hm5\") pod \"redhat-marketplace-4p78s\" (UID: \"adaf0543-043d-48b8-829f-928df3bf6e90\") " pod="openshift-marketplace/redhat-marketplace-4p78s" Oct 02 01:49:22 crc kubenswrapper[4885]: I1002 01:49:22.180390 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/adaf0543-043d-48b8-829f-928df3bf6e90-utilities\") pod \"redhat-marketplace-4p78s\" (UID: \"adaf0543-043d-48b8-829f-928df3bf6e90\") " pod="openshift-marketplace/redhat-marketplace-4p78s" Oct 02 01:49:22 crc kubenswrapper[4885]: I1002 01:49:22.281760 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s4hm5\" (UniqueName: \"kubernetes.io/projected/adaf0543-043d-48b8-829f-928df3bf6e90-kube-api-access-s4hm5\") pod \"redhat-marketplace-4p78s\" (UID: \"adaf0543-043d-48b8-829f-928df3bf6e90\") " pod="openshift-marketplace/redhat-marketplace-4p78s" Oct 02 01:49:22 crc kubenswrapper[4885]: I1002 01:49:22.281800 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/adaf0543-043d-48b8-829f-928df3bf6e90-utilities\") pod \"redhat-marketplace-4p78s\" (UID: \"adaf0543-043d-48b8-829f-928df3bf6e90\") " pod="openshift-marketplace/redhat-marketplace-4p78s" Oct 02 01:49:22 crc kubenswrapper[4885]: I1002 01:49:22.281875 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/adaf0543-043d-48b8-829f-928df3bf6e90-catalog-content\") pod \"redhat-marketplace-4p78s\" (UID: \"adaf0543-043d-48b8-829f-928df3bf6e90\") " pod="openshift-marketplace/redhat-marketplace-4p78s" Oct 02 01:49:22 crc kubenswrapper[4885]: I1002 01:49:22.282417 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/adaf0543-043d-48b8-829f-928df3bf6e90-catalog-content\") pod \"redhat-marketplace-4p78s\" (UID: \"adaf0543-043d-48b8-829f-928df3bf6e90\") " pod="openshift-marketplace/redhat-marketplace-4p78s" Oct 02 01:49:22 crc kubenswrapper[4885]: I1002 01:49:22.282864 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/adaf0543-043d-48b8-829f-928df3bf6e90-utilities\") pod \"redhat-marketplace-4p78s\" (UID: \"adaf0543-043d-48b8-829f-928df3bf6e90\") " pod="openshift-marketplace/redhat-marketplace-4p78s" Oct 02 01:49:22 crc kubenswrapper[4885]: I1002 01:49:22.297191 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s4hm5\" (UniqueName: \"kubernetes.io/projected/adaf0543-043d-48b8-829f-928df3bf6e90-kube-api-access-s4hm5\") pod \"redhat-marketplace-4p78s\" (UID: \"adaf0543-043d-48b8-829f-928df3bf6e90\") " pod="openshift-marketplace/redhat-marketplace-4p78s" Oct 02 01:49:22 crc kubenswrapper[4885]: I1002 01:49:22.388403 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4p78s" Oct 02 01:49:22 crc kubenswrapper[4885]: I1002 01:49:22.456108 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bz9hg"] Oct 02 01:49:22 crc kubenswrapper[4885]: I1002 01:49:22.503824 4885 patch_prober.go:28] interesting pod/router-default-5444994796-zxz72 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 01:49:22 crc kubenswrapper[4885]: [-]has-synced failed: reason withheld Oct 02 01:49:22 crc kubenswrapper[4885]: [+]process-running ok Oct 02 01:49:22 crc kubenswrapper[4885]: healthz check failed Oct 02 01:49:22 crc kubenswrapper[4885]: I1002 01:49:22.503895 4885 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-zxz72" podUID="f0d19318-e77c-40d4-89b6-da97f7b1f91a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 01:49:22 crc kubenswrapper[4885]: I1002 01:49:22.584711 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4p78s"] Oct 02 01:49:22 crc kubenswrapper[4885]: W1002 01:49:22.602512 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podadaf0543_043d_48b8_829f_928df3bf6e90.slice/crio-5ea24c226c89380df68904a299e46b054f2178d9e00a0b3dc5637146f62742a9 WatchSource:0}: Error finding container 5ea24c226c89380df68904a299e46b054f2178d9e00a0b3dc5637146f62742a9: Status 404 returned error can't find the container with id 5ea24c226c89380df68904a299e46b054f2178d9e00a0b3dc5637146f62742a9 Oct 02 01:49:22 crc kubenswrapper[4885]: I1002 01:49:22.632623 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-bjxpx"] Oct 02 01:49:22 crc kubenswrapper[4885]: I1002 01:49:22.633564 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bjxpx" Oct 02 01:49:22 crc kubenswrapper[4885]: I1002 01:49:22.639953 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 02 01:49:22 crc kubenswrapper[4885]: I1002 01:49:22.644158 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bjxpx"] Oct 02 01:49:22 crc kubenswrapper[4885]: I1002 01:49:22.790206 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-smmgf\" (UniqueName: \"kubernetes.io/projected/18a2b456-262c-43fd-97c6-bfaf2d658463-kube-api-access-smmgf\") pod \"redhat-operators-bjxpx\" (UID: \"18a2b456-262c-43fd-97c6-bfaf2d658463\") " pod="openshift-marketplace/redhat-operators-bjxpx" Oct 02 01:49:22 crc kubenswrapper[4885]: I1002 01:49:22.790292 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18a2b456-262c-43fd-97c6-bfaf2d658463-catalog-content\") pod \"redhat-operators-bjxpx\" (UID: \"18a2b456-262c-43fd-97c6-bfaf2d658463\") " pod="openshift-marketplace/redhat-operators-bjxpx" Oct 02 01:49:22 crc kubenswrapper[4885]: I1002 01:49:22.790427 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18a2b456-262c-43fd-97c6-bfaf2d658463-utilities\") pod \"redhat-operators-bjxpx\" (UID: \"18a2b456-262c-43fd-97c6-bfaf2d658463\") " pod="openshift-marketplace/redhat-operators-bjxpx" Oct 02 01:49:22 crc kubenswrapper[4885]: I1002 01:49:22.841960 4885 generic.go:334] "Generic (PLEG): container finished" podID="aecca2a3-40b0-4666-a86d-d67ecb580e99" containerID="f2bc855d60ae2f644fbcb6ce6f352db3d78ca46e877869ee645e59ef859d6482" exitCode=0 Oct 02 01:49:22 crc kubenswrapper[4885]: I1002 01:49:22.842048 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bz9hg" event={"ID":"aecca2a3-40b0-4666-a86d-d67ecb580e99","Type":"ContainerDied","Data":"f2bc855d60ae2f644fbcb6ce6f352db3d78ca46e877869ee645e59ef859d6482"} Oct 02 01:49:22 crc kubenswrapper[4885]: I1002 01:49:22.842077 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bz9hg" event={"ID":"aecca2a3-40b0-4666-a86d-d67ecb580e99","Type":"ContainerStarted","Data":"782ea6d3943d5c7111e54f12edf8a09d145a4689c9943241adbb6da0379ff934"} Oct 02 01:49:22 crc kubenswrapper[4885]: I1002 01:49:22.843770 4885 generic.go:334] "Generic (PLEG): container finished" podID="72a00566-8802-4940-9ed2-852e226dc735" containerID="8f7344570a2e047f77a5bd7137edf13910c4cf9c41784d9287bbed0025faf93a" exitCode=0 Oct 02 01:49:22 crc kubenswrapper[4885]: I1002 01:49:22.843816 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"72a00566-8802-4940-9ed2-852e226dc735","Type":"ContainerDied","Data":"8f7344570a2e047f77a5bd7137edf13910c4cf9c41784d9287bbed0025faf93a"} Oct 02 01:49:22 crc kubenswrapper[4885]: I1002 01:49:22.846308 4885 generic.go:334] "Generic (PLEG): container finished" podID="4bc961f3-ae85-4c19-8a57-be0af8e7bd10" containerID="6930508575fab66820a624c8626849925a65067064512e41a2af37ab86368c07" exitCode=0 Oct 02 01:49:22 crc kubenswrapper[4885]: I1002 01:49:22.846385 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29322825-b7zrt" event={"ID":"4bc961f3-ae85-4c19-8a57-be0af8e7bd10","Type":"ContainerDied","Data":"6930508575fab66820a624c8626849925a65067064512e41a2af37ab86368c07"} Oct 02 01:49:22 crc kubenswrapper[4885]: I1002 01:49:22.847971 4885 generic.go:334] "Generic (PLEG): container finished" podID="adaf0543-043d-48b8-829f-928df3bf6e90" containerID="6759585f89e100d2eec13552ca691f04cac3cd733d99323bed065d6cb55b0b70" exitCode=0 Oct 02 01:49:22 crc kubenswrapper[4885]: I1002 01:49:22.848054 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4p78s" event={"ID":"adaf0543-043d-48b8-829f-928df3bf6e90","Type":"ContainerDied","Data":"6759585f89e100d2eec13552ca691f04cac3cd733d99323bed065d6cb55b0b70"} Oct 02 01:49:22 crc kubenswrapper[4885]: I1002 01:49:22.848081 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4p78s" event={"ID":"adaf0543-043d-48b8-829f-928df3bf6e90","Type":"ContainerStarted","Data":"5ea24c226c89380df68904a299e46b054f2178d9e00a0b3dc5637146f62742a9"} Oct 02 01:49:22 crc kubenswrapper[4885]: I1002 01:49:22.894381 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18a2b456-262c-43fd-97c6-bfaf2d658463-utilities\") pod \"redhat-operators-bjxpx\" (UID: \"18a2b456-262c-43fd-97c6-bfaf2d658463\") " pod="openshift-marketplace/redhat-operators-bjxpx" Oct 02 01:49:22 crc kubenswrapper[4885]: I1002 01:49:22.894503 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-smmgf\" (UniqueName: \"kubernetes.io/projected/18a2b456-262c-43fd-97c6-bfaf2d658463-kube-api-access-smmgf\") pod \"redhat-operators-bjxpx\" (UID: \"18a2b456-262c-43fd-97c6-bfaf2d658463\") " pod="openshift-marketplace/redhat-operators-bjxpx" Oct 02 01:49:22 crc kubenswrapper[4885]: I1002 01:49:22.894543 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18a2b456-262c-43fd-97c6-bfaf2d658463-catalog-content\") pod \"redhat-operators-bjxpx\" (UID: \"18a2b456-262c-43fd-97c6-bfaf2d658463\") " pod="openshift-marketplace/redhat-operators-bjxpx" Oct 02 01:49:22 crc kubenswrapper[4885]: I1002 01:49:22.894980 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18a2b456-262c-43fd-97c6-bfaf2d658463-catalog-content\") pod \"redhat-operators-bjxpx\" (UID: \"18a2b456-262c-43fd-97c6-bfaf2d658463\") " pod="openshift-marketplace/redhat-operators-bjxpx" Oct 02 01:49:22 crc kubenswrapper[4885]: I1002 01:49:22.895372 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18a2b456-262c-43fd-97c6-bfaf2d658463-utilities\") pod \"redhat-operators-bjxpx\" (UID: \"18a2b456-262c-43fd-97c6-bfaf2d658463\") " pod="openshift-marketplace/redhat-operators-bjxpx" Oct 02 01:49:22 crc kubenswrapper[4885]: I1002 01:49:22.914948 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-smmgf\" (UniqueName: \"kubernetes.io/projected/18a2b456-262c-43fd-97c6-bfaf2d658463-kube-api-access-smmgf\") pod \"redhat-operators-bjxpx\" (UID: \"18a2b456-262c-43fd-97c6-bfaf2d658463\") " pod="openshift-marketplace/redhat-operators-bjxpx" Oct 02 01:49:22 crc kubenswrapper[4885]: I1002 01:49:22.978589 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bjxpx" Oct 02 01:49:23 crc kubenswrapper[4885]: I1002 01:49:23.029835 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-xq6kh"] Oct 02 01:49:23 crc kubenswrapper[4885]: I1002 01:49:23.031035 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xq6kh" Oct 02 01:49:23 crc kubenswrapper[4885]: I1002 01:49:23.037628 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xq6kh"] Oct 02 01:49:23 crc kubenswrapper[4885]: I1002 01:49:23.205598 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-24dnt\" (UniqueName: \"kubernetes.io/projected/21d974e0-8d62-4daf-b383-5a6823df7b8e-kube-api-access-24dnt\") pod \"redhat-operators-xq6kh\" (UID: \"21d974e0-8d62-4daf-b383-5a6823df7b8e\") " pod="openshift-marketplace/redhat-operators-xq6kh" Oct 02 01:49:23 crc kubenswrapper[4885]: I1002 01:49:23.205897 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/21d974e0-8d62-4daf-b383-5a6823df7b8e-catalog-content\") pod \"redhat-operators-xq6kh\" (UID: \"21d974e0-8d62-4daf-b383-5a6823df7b8e\") " pod="openshift-marketplace/redhat-operators-xq6kh" Oct 02 01:49:23 crc kubenswrapper[4885]: I1002 01:49:23.205936 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/21d974e0-8d62-4daf-b383-5a6823df7b8e-utilities\") pod \"redhat-operators-xq6kh\" (UID: \"21d974e0-8d62-4daf-b383-5a6823df7b8e\") " pod="openshift-marketplace/redhat-operators-xq6kh" Oct 02 01:49:23 crc kubenswrapper[4885]: I1002 01:49:23.306960 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-24dnt\" (UniqueName: \"kubernetes.io/projected/21d974e0-8d62-4daf-b383-5a6823df7b8e-kube-api-access-24dnt\") pod \"redhat-operators-xq6kh\" (UID: \"21d974e0-8d62-4daf-b383-5a6823df7b8e\") " pod="openshift-marketplace/redhat-operators-xq6kh" Oct 02 01:49:23 crc kubenswrapper[4885]: I1002 01:49:23.307017 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/21d974e0-8d62-4daf-b383-5a6823df7b8e-catalog-content\") pod \"redhat-operators-xq6kh\" (UID: \"21d974e0-8d62-4daf-b383-5a6823df7b8e\") " pod="openshift-marketplace/redhat-operators-xq6kh" Oct 02 01:49:23 crc kubenswrapper[4885]: I1002 01:49:23.307036 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/21d974e0-8d62-4daf-b383-5a6823df7b8e-utilities\") pod \"redhat-operators-xq6kh\" (UID: \"21d974e0-8d62-4daf-b383-5a6823df7b8e\") " pod="openshift-marketplace/redhat-operators-xq6kh" Oct 02 01:49:23 crc kubenswrapper[4885]: I1002 01:49:23.307510 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/21d974e0-8d62-4daf-b383-5a6823df7b8e-utilities\") pod \"redhat-operators-xq6kh\" (UID: \"21d974e0-8d62-4daf-b383-5a6823df7b8e\") " pod="openshift-marketplace/redhat-operators-xq6kh" Oct 02 01:49:23 crc kubenswrapper[4885]: I1002 01:49:23.307720 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/21d974e0-8d62-4daf-b383-5a6823df7b8e-catalog-content\") pod \"redhat-operators-xq6kh\" (UID: \"21d974e0-8d62-4daf-b383-5a6823df7b8e\") " pod="openshift-marketplace/redhat-operators-xq6kh" Oct 02 01:49:23 crc kubenswrapper[4885]: I1002 01:49:23.322605 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-24dnt\" (UniqueName: \"kubernetes.io/projected/21d974e0-8d62-4daf-b383-5a6823df7b8e-kube-api-access-24dnt\") pod \"redhat-operators-xq6kh\" (UID: \"21d974e0-8d62-4daf-b383-5a6823df7b8e\") " pod="openshift-marketplace/redhat-operators-xq6kh" Oct 02 01:49:23 crc kubenswrapper[4885]: I1002 01:49:23.349877 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xq6kh" Oct 02 01:49:23 crc kubenswrapper[4885]: I1002 01:49:23.458003 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bjxpx"] Oct 02 01:49:23 crc kubenswrapper[4885]: W1002 01:49:23.473813 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod18a2b456_262c_43fd_97c6_bfaf2d658463.slice/crio-e78614ca8f60a5b44f38c8025dd93bb1c8d5d7b8f35750263ca0a38471bcc53e WatchSource:0}: Error finding container e78614ca8f60a5b44f38c8025dd93bb1c8d5d7b8f35750263ca0a38471bcc53e: Status 404 returned error can't find the container with id e78614ca8f60a5b44f38c8025dd93bb1c8d5d7b8f35750263ca0a38471bcc53e Oct 02 01:49:23 crc kubenswrapper[4885]: I1002 01:49:23.503400 4885 patch_prober.go:28] interesting pod/router-default-5444994796-zxz72 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 01:49:23 crc kubenswrapper[4885]: [-]has-synced failed: reason withheld Oct 02 01:49:23 crc kubenswrapper[4885]: [+]process-running ok Oct 02 01:49:23 crc kubenswrapper[4885]: healthz check failed Oct 02 01:49:23 crc kubenswrapper[4885]: I1002 01:49:23.503449 4885 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-zxz72" podUID="f0d19318-e77c-40d4-89b6-da97f7b1f91a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 01:49:23 crc kubenswrapper[4885]: I1002 01:49:23.621128 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xq6kh"] Oct 02 01:49:23 crc kubenswrapper[4885]: W1002 01:49:23.628407 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod21d974e0_8d62_4daf_b383_5a6823df7b8e.slice/crio-78033f0e604969f00abf669646a13ff7757a6a01337a3d9270fe18522c086da0 WatchSource:0}: Error finding container 78033f0e604969f00abf669646a13ff7757a6a01337a3d9270fe18522c086da0: Status 404 returned error can't find the container with id 78033f0e604969f00abf669646a13ff7757a6a01337a3d9270fe18522c086da0 Oct 02 01:49:23 crc kubenswrapper[4885]: I1002 01:49:23.689018 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-hg5rx" Oct 02 01:49:23 crc kubenswrapper[4885]: I1002 01:49:23.695193 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-hg5rx" Oct 02 01:49:23 crc kubenswrapper[4885]: I1002 01:49:23.712186 4885 patch_prober.go:28] interesting pod/downloads-7954f5f757-5kn8g container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Oct 02 01:49:23 crc kubenswrapper[4885]: I1002 01:49:23.712217 4885 patch_prober.go:28] interesting pod/downloads-7954f5f757-5kn8g container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Oct 02 01:49:23 crc kubenswrapper[4885]: I1002 01:49:23.712234 4885 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-5kn8g" podUID="fd852137-cff1-4fae-a921-42be5c511db8" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Oct 02 01:49:23 crc kubenswrapper[4885]: I1002 01:49:23.712268 4885 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-5kn8g" podUID="fd852137-cff1-4fae-a921-42be5c511db8" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Oct 02 01:49:23 crc kubenswrapper[4885]: I1002 01:49:23.863024 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xq6kh" event={"ID":"21d974e0-8d62-4daf-b383-5a6823df7b8e","Type":"ContainerStarted","Data":"7e5d8fe53556ae09ff185262c2de844873acae065f72e9740d554d72ff3968e1"} Oct 02 01:49:23 crc kubenswrapper[4885]: I1002 01:49:23.863062 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xq6kh" event={"ID":"21d974e0-8d62-4daf-b383-5a6823df7b8e","Type":"ContainerStarted","Data":"78033f0e604969f00abf669646a13ff7757a6a01337a3d9270fe18522c086da0"} Oct 02 01:49:23 crc kubenswrapper[4885]: I1002 01:49:23.866218 4885 generic.go:334] "Generic (PLEG): container finished" podID="18a2b456-262c-43fd-97c6-bfaf2d658463" containerID="7e2a6b3be7e1a960b7e630282ad27b151752fa01cd3c4af7d6808b5f6e302500" exitCode=0 Oct 02 01:49:23 crc kubenswrapper[4885]: I1002 01:49:23.866774 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bjxpx" event={"ID":"18a2b456-262c-43fd-97c6-bfaf2d658463","Type":"ContainerDied","Data":"7e2a6b3be7e1a960b7e630282ad27b151752fa01cd3c4af7d6808b5f6e302500"} Oct 02 01:49:23 crc kubenswrapper[4885]: I1002 01:49:23.866794 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bjxpx" event={"ID":"18a2b456-262c-43fd-97c6-bfaf2d658463","Type":"ContainerStarted","Data":"e78614ca8f60a5b44f38c8025dd93bb1c8d5d7b8f35750263ca0a38471bcc53e"} Oct 02 01:49:24 crc kubenswrapper[4885]: I1002 01:49:24.129391 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322825-b7zrt" Oct 02 01:49:24 crc kubenswrapper[4885]: I1002 01:49:24.234643 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4bc961f3-ae85-4c19-8a57-be0af8e7bd10-config-volume\") pod \"4bc961f3-ae85-4c19-8a57-be0af8e7bd10\" (UID: \"4bc961f3-ae85-4c19-8a57-be0af8e7bd10\") " Oct 02 01:49:24 crc kubenswrapper[4885]: I1002 01:49:24.234724 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4bc961f3-ae85-4c19-8a57-be0af8e7bd10-secret-volume\") pod \"4bc961f3-ae85-4c19-8a57-be0af8e7bd10\" (UID: \"4bc961f3-ae85-4c19-8a57-be0af8e7bd10\") " Oct 02 01:49:24 crc kubenswrapper[4885]: I1002 01:49:24.234799 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7rqt5\" (UniqueName: \"kubernetes.io/projected/4bc961f3-ae85-4c19-8a57-be0af8e7bd10-kube-api-access-7rqt5\") pod \"4bc961f3-ae85-4c19-8a57-be0af8e7bd10\" (UID: \"4bc961f3-ae85-4c19-8a57-be0af8e7bd10\") " Oct 02 01:49:24 crc kubenswrapper[4885]: I1002 01:49:24.236328 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bc961f3-ae85-4c19-8a57-be0af8e7bd10-config-volume" (OuterVolumeSpecName: "config-volume") pod "4bc961f3-ae85-4c19-8a57-be0af8e7bd10" (UID: "4bc961f3-ae85-4c19-8a57-be0af8e7bd10"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:49:24 crc kubenswrapper[4885]: I1002 01:49:24.242355 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bc961f3-ae85-4c19-8a57-be0af8e7bd10-kube-api-access-7rqt5" (OuterVolumeSpecName: "kube-api-access-7rqt5") pod "4bc961f3-ae85-4c19-8a57-be0af8e7bd10" (UID: "4bc961f3-ae85-4c19-8a57-be0af8e7bd10"). InnerVolumeSpecName "kube-api-access-7rqt5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:49:24 crc kubenswrapper[4885]: I1002 01:49:24.253583 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4bc961f3-ae85-4c19-8a57-be0af8e7bd10-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "4bc961f3-ae85-4c19-8a57-be0af8e7bd10" (UID: "4bc961f3-ae85-4c19-8a57-be0af8e7bd10"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:49:24 crc kubenswrapper[4885]: I1002 01:49:24.272472 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 02 01:49:24 crc kubenswrapper[4885]: I1002 01:49:24.336504 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7rqt5\" (UniqueName: \"kubernetes.io/projected/4bc961f3-ae85-4c19-8a57-be0af8e7bd10-kube-api-access-7rqt5\") on node \"crc\" DevicePath \"\"" Oct 02 01:49:24 crc kubenswrapper[4885]: I1002 01:49:24.336534 4885 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4bc961f3-ae85-4c19-8a57-be0af8e7bd10-config-volume\") on node \"crc\" DevicePath \"\"" Oct 02 01:49:24 crc kubenswrapper[4885]: I1002 01:49:24.336547 4885 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4bc961f3-ae85-4c19-8a57-be0af8e7bd10-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 02 01:49:24 crc kubenswrapper[4885]: I1002 01:49:24.360245 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-kckm8" Oct 02 01:49:24 crc kubenswrapper[4885]: I1002 01:49:24.360966 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-kckm8" Oct 02 01:49:24 crc kubenswrapper[4885]: I1002 01:49:24.363149 4885 patch_prober.go:28] interesting pod/console-f9d7485db-kckm8 container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.21:8443/health\": dial tcp 10.217.0.21:8443: connect: connection refused" start-of-body= Oct 02 01:49:24 crc kubenswrapper[4885]: I1002 01:49:24.363190 4885 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-kckm8" podUID="a86af5ef-01e3-47e4-b324-208537e5ec71" containerName="console" probeResult="failure" output="Get \"https://10.217.0.21:8443/health\": dial tcp 10.217.0.21:8443: connect: connection refused" Oct 02 01:49:24 crc kubenswrapper[4885]: I1002 01:49:24.437253 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/72a00566-8802-4940-9ed2-852e226dc735-kubelet-dir\") pod \"72a00566-8802-4940-9ed2-852e226dc735\" (UID: \"72a00566-8802-4940-9ed2-852e226dc735\") " Oct 02 01:49:24 crc kubenswrapper[4885]: I1002 01:49:24.437404 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/72a00566-8802-4940-9ed2-852e226dc735-kube-api-access\") pod \"72a00566-8802-4940-9ed2-852e226dc735\" (UID: \"72a00566-8802-4940-9ed2-852e226dc735\") " Oct 02 01:49:24 crc kubenswrapper[4885]: I1002 01:49:24.440485 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/72a00566-8802-4940-9ed2-852e226dc735-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "72a00566-8802-4940-9ed2-852e226dc735" (UID: "72a00566-8802-4940-9ed2-852e226dc735"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 01:49:24 crc kubenswrapper[4885]: I1002 01:49:24.444339 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/72a00566-8802-4940-9ed2-852e226dc735-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "72a00566-8802-4940-9ed2-852e226dc735" (UID: "72a00566-8802-4940-9ed2-852e226dc735"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:49:24 crc kubenswrapper[4885]: I1002 01:49:24.500538 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-zxz72" Oct 02 01:49:24 crc kubenswrapper[4885]: I1002 01:49:24.503097 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-zxz72" Oct 02 01:49:24 crc kubenswrapper[4885]: I1002 01:49:24.535244 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-z4rlr" Oct 02 01:49:24 crc kubenswrapper[4885]: I1002 01:49:24.539817 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/72a00566-8802-4940-9ed2-852e226dc735-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 02 01:49:24 crc kubenswrapper[4885]: I1002 01:49:24.539848 4885 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/72a00566-8802-4940-9ed2-852e226dc735-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 02 01:49:24 crc kubenswrapper[4885]: I1002 01:49:24.880600 4885 generic.go:334] "Generic (PLEG): container finished" podID="21d974e0-8d62-4daf-b383-5a6823df7b8e" containerID="7e5d8fe53556ae09ff185262c2de844873acae065f72e9740d554d72ff3968e1" exitCode=0 Oct 02 01:49:24 crc kubenswrapper[4885]: I1002 01:49:24.880657 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xq6kh" event={"ID":"21d974e0-8d62-4daf-b383-5a6823df7b8e","Type":"ContainerDied","Data":"7e5d8fe53556ae09ff185262c2de844873acae065f72e9740d554d72ff3968e1"} Oct 02 01:49:24 crc kubenswrapper[4885]: I1002 01:49:24.901207 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"72a00566-8802-4940-9ed2-852e226dc735","Type":"ContainerDied","Data":"bea8cbcb7ba8fa989c5e153275a2725a6455d3d57e8a26df202346fa79533c94"} Oct 02 01:49:24 crc kubenswrapper[4885]: I1002 01:49:24.901281 4885 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bea8cbcb7ba8fa989c5e153275a2725a6455d3d57e8a26df202346fa79533c94" Oct 02 01:49:24 crc kubenswrapper[4885]: I1002 01:49:24.901745 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 02 01:49:24 crc kubenswrapper[4885]: I1002 01:49:24.905140 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322825-b7zrt" Oct 02 01:49:24 crc kubenswrapper[4885]: I1002 01:49:24.913968 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29322825-b7zrt" event={"ID":"4bc961f3-ae85-4c19-8a57-be0af8e7bd10","Type":"ContainerDied","Data":"93d41334a68e7f750f8bfdc89a8e5ed4352a32f7c03e51cc62f2ca09320c50cc"} Oct 02 01:49:24 crc kubenswrapper[4885]: I1002 01:49:24.914016 4885 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="93d41334a68e7f750f8bfdc89a8e5ed4352a32f7c03e51cc62f2ca09320c50cc" Oct 02 01:49:24 crc kubenswrapper[4885]: I1002 01:49:24.916320 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-zxz72" Oct 02 01:49:25 crc kubenswrapper[4885]: I1002 01:49:25.403027 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 02 01:49:25 crc kubenswrapper[4885]: E1002 01:49:25.403477 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72a00566-8802-4940-9ed2-852e226dc735" containerName="pruner" Oct 02 01:49:25 crc kubenswrapper[4885]: I1002 01:49:25.403488 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="72a00566-8802-4940-9ed2-852e226dc735" containerName="pruner" Oct 02 01:49:25 crc kubenswrapper[4885]: E1002 01:49:25.403496 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4bc961f3-ae85-4c19-8a57-be0af8e7bd10" containerName="collect-profiles" Oct 02 01:49:25 crc kubenswrapper[4885]: I1002 01:49:25.403502 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="4bc961f3-ae85-4c19-8a57-be0af8e7bd10" containerName="collect-profiles" Oct 02 01:49:25 crc kubenswrapper[4885]: I1002 01:49:25.403601 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="72a00566-8802-4940-9ed2-852e226dc735" containerName="pruner" Oct 02 01:49:25 crc kubenswrapper[4885]: I1002 01:49:25.403611 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="4bc961f3-ae85-4c19-8a57-be0af8e7bd10" containerName="collect-profiles" Oct 02 01:49:25 crc kubenswrapper[4885]: I1002 01:49:25.403918 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 02 01:49:25 crc kubenswrapper[4885]: I1002 01:49:25.406108 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Oct 02 01:49:25 crc kubenswrapper[4885]: I1002 01:49:25.406238 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Oct 02 01:49:25 crc kubenswrapper[4885]: I1002 01:49:25.415509 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 02 01:49:25 crc kubenswrapper[4885]: I1002 01:49:25.559567 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/11b7ddd4-611b-41ee-9e12-b436939be852-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"11b7ddd4-611b-41ee-9e12-b436939be852\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 02 01:49:25 crc kubenswrapper[4885]: I1002 01:49:25.559602 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/11b7ddd4-611b-41ee-9e12-b436939be852-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"11b7ddd4-611b-41ee-9e12-b436939be852\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 02 01:49:25 crc kubenswrapper[4885]: I1002 01:49:25.661154 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/11b7ddd4-611b-41ee-9e12-b436939be852-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"11b7ddd4-611b-41ee-9e12-b436939be852\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 02 01:49:25 crc kubenswrapper[4885]: I1002 01:49:25.661202 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/11b7ddd4-611b-41ee-9e12-b436939be852-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"11b7ddd4-611b-41ee-9e12-b436939be852\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 02 01:49:25 crc kubenswrapper[4885]: I1002 01:49:25.661369 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/11b7ddd4-611b-41ee-9e12-b436939be852-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"11b7ddd4-611b-41ee-9e12-b436939be852\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 02 01:49:25 crc kubenswrapper[4885]: I1002 01:49:25.682973 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/11b7ddd4-611b-41ee-9e12-b436939be852-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"11b7ddd4-611b-41ee-9e12-b436939be852\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 02 01:49:25 crc kubenswrapper[4885]: I1002 01:49:25.745141 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 02 01:49:26 crc kubenswrapper[4885]: I1002 01:49:26.267384 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 02 01:49:26 crc kubenswrapper[4885]: W1002 01:49:26.272215 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod11b7ddd4_611b_41ee_9e12_b436939be852.slice/crio-1da65753e7adc9ee9b6556248f751dd6093c42e2b46bce8e6445f56934de6452 WatchSource:0}: Error finding container 1da65753e7adc9ee9b6556248f751dd6093c42e2b46bce8e6445f56934de6452: Status 404 returned error can't find the container with id 1da65753e7adc9ee9b6556248f751dd6093c42e2b46bce8e6445f56934de6452 Oct 02 01:49:26 crc kubenswrapper[4885]: I1002 01:49:26.646091 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-6bfsg" Oct 02 01:49:26 crc kubenswrapper[4885]: I1002 01:49:26.925832 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"11b7ddd4-611b-41ee-9e12-b436939be852","Type":"ContainerStarted","Data":"1da65753e7adc9ee9b6556248f751dd6093c42e2b46bce8e6445f56934de6452"} Oct 02 01:49:27 crc kubenswrapper[4885]: I1002 01:49:27.960753 4885 generic.go:334] "Generic (PLEG): container finished" podID="11b7ddd4-611b-41ee-9e12-b436939be852" containerID="9a4444b5cae62a55380ec0990796a0396f7d786bbca848e134fb5a1f3b860208" exitCode=0 Oct 02 01:49:27 crc kubenswrapper[4885]: I1002 01:49:27.960819 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"11b7ddd4-611b-41ee-9e12-b436939be852","Type":"ContainerDied","Data":"9a4444b5cae62a55380ec0990796a0396f7d786bbca848e134fb5a1f3b860208"} Oct 02 01:49:30 crc kubenswrapper[4885]: I1002 01:49:30.408977 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:49:33 crc kubenswrapper[4885]: I1002 01:49:33.111023 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 02 01:49:33 crc kubenswrapper[4885]: I1002 01:49:33.269696 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/11b7ddd4-611b-41ee-9e12-b436939be852-kube-api-access\") pod \"11b7ddd4-611b-41ee-9e12-b436939be852\" (UID: \"11b7ddd4-611b-41ee-9e12-b436939be852\") " Oct 02 01:49:33 crc kubenswrapper[4885]: I1002 01:49:33.269752 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/11b7ddd4-611b-41ee-9e12-b436939be852-kubelet-dir\") pod \"11b7ddd4-611b-41ee-9e12-b436939be852\" (UID: \"11b7ddd4-611b-41ee-9e12-b436939be852\") " Oct 02 01:49:33 crc kubenswrapper[4885]: I1002 01:49:33.269957 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/11b7ddd4-611b-41ee-9e12-b436939be852-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "11b7ddd4-611b-41ee-9e12-b436939be852" (UID: "11b7ddd4-611b-41ee-9e12-b436939be852"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 01:49:33 crc kubenswrapper[4885]: I1002 01:49:33.277101 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/11b7ddd4-611b-41ee-9e12-b436939be852-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "11b7ddd4-611b-41ee-9e12-b436939be852" (UID: "11b7ddd4-611b-41ee-9e12-b436939be852"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:49:33 crc kubenswrapper[4885]: I1002 01:49:33.371624 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/11b7ddd4-611b-41ee-9e12-b436939be852-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 02 01:49:33 crc kubenswrapper[4885]: I1002 01:49:33.372142 4885 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/11b7ddd4-611b-41ee-9e12-b436939be852-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 02 01:49:33 crc kubenswrapper[4885]: I1002 01:49:33.713012 4885 patch_prober.go:28] interesting pod/downloads-7954f5f757-5kn8g container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Oct 02 01:49:33 crc kubenswrapper[4885]: I1002 01:49:33.713086 4885 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-5kn8g" podUID="fd852137-cff1-4fae-a921-42be5c511db8" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Oct 02 01:49:33 crc kubenswrapper[4885]: I1002 01:49:33.713178 4885 patch_prober.go:28] interesting pod/downloads-7954f5f757-5kn8g container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Oct 02 01:49:33 crc kubenswrapper[4885]: I1002 01:49:33.713287 4885 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-5kn8g" podUID="fd852137-cff1-4fae-a921-42be5c511db8" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Oct 02 01:49:34 crc kubenswrapper[4885]: I1002 01:49:34.000204 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"11b7ddd4-611b-41ee-9e12-b436939be852","Type":"ContainerDied","Data":"1da65753e7adc9ee9b6556248f751dd6093c42e2b46bce8e6445f56934de6452"} Oct 02 01:49:34 crc kubenswrapper[4885]: I1002 01:49:34.000243 4885 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1da65753e7adc9ee9b6556248f751dd6093c42e2b46bce8e6445f56934de6452" Oct 02 01:49:34 crc kubenswrapper[4885]: I1002 01:49:34.000334 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 02 01:49:34 crc kubenswrapper[4885]: I1002 01:49:34.364688 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-kckm8" Oct 02 01:49:34 crc kubenswrapper[4885]: I1002 01:49:34.368497 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-kckm8" Oct 02 01:49:36 crc kubenswrapper[4885]: I1002 01:49:36.322321 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/724a3e49-001c-45a4-a896-82c13e0ee0e7-metrics-certs\") pod \"network-metrics-daemon-sv4md\" (UID: \"724a3e49-001c-45a4-a896-82c13e0ee0e7\") " pod="openshift-multus/network-metrics-daemon-sv4md" Oct 02 01:49:36 crc kubenswrapper[4885]: I1002 01:49:36.329814 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/724a3e49-001c-45a4-a896-82c13e0ee0e7-metrics-certs\") pod \"network-metrics-daemon-sv4md\" (UID: \"724a3e49-001c-45a4-a896-82c13e0ee0e7\") " pod="openshift-multus/network-metrics-daemon-sv4md" Oct 02 01:49:36 crc kubenswrapper[4885]: I1002 01:49:36.373327 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sv4md" Oct 02 01:49:40 crc kubenswrapper[4885]: I1002 01:49:40.715747 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-gdxhh" Oct 02 01:49:43 crc kubenswrapper[4885]: I1002 01:49:43.266164 4885 patch_prober.go:28] interesting pod/machine-config-daemon-rttx8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 01:49:43 crc kubenswrapper[4885]: I1002 01:49:43.266230 4885 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 01:49:43 crc kubenswrapper[4885]: I1002 01:49:43.721859 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-5kn8g" Oct 02 01:49:54 crc kubenswrapper[4885]: E1002 01:49:54.322909 4885 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Oct 02 01:49:54 crc kubenswrapper[4885]: E1002 01:49:54.323761 4885 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hjnzp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-bj65p_openshift-marketplace(7d3cc54c-759c-43a7-b88c-4c5079b42e45): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 02 01:49:54 crc kubenswrapper[4885]: E1002 01:49:54.324865 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-bj65p" podUID="7d3cc54c-759c-43a7-b88c-4c5079b42e45" Oct 02 01:49:54 crc kubenswrapper[4885]: I1002 01:49:54.557124 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xjfdv" Oct 02 01:49:55 crc kubenswrapper[4885]: E1002 01:49:55.631472 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-bj65p" podUID="7d3cc54c-759c-43a7-b88c-4c5079b42e45" Oct 02 01:49:55 crc kubenswrapper[4885]: E1002 01:49:55.694689 4885 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Oct 02 01:49:55 crc kubenswrapper[4885]: E1002 01:49:55.694902 4885 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-svczt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-x44t6_openshift-marketplace(961bc3c9-b6be-4ece-98f1-971a583f9db8): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 02 01:49:55 crc kubenswrapper[4885]: E1002 01:49:55.696491 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-x44t6" podUID="961bc3c9-b6be-4ece-98f1-971a583f9db8" Oct 02 01:49:55 crc kubenswrapper[4885]: E1002 01:49:55.757866 4885 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Oct 02 01:49:55 crc kubenswrapper[4885]: E1002 01:49:55.758011 4885 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-f94th,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-vgkg6_openshift-marketplace(64a7eacf-552b-4a87-91db-4f6f72ee26b1): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 02 01:49:55 crc kubenswrapper[4885]: E1002 01:49:55.759221 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-vgkg6" podUID="64a7eacf-552b-4a87-91db-4f6f72ee26b1" Oct 02 01:49:55 crc kubenswrapper[4885]: E1002 01:49:55.767594 4885 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Oct 02 01:49:55 crc kubenswrapper[4885]: E1002 01:49:55.767671 4885 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-cdnq6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-vb2j4_openshift-marketplace(a28f881c-dd5e-4bf8-82ad-2c27b98dcceb): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 02 01:49:55 crc kubenswrapper[4885]: E1002 01:49:55.769761 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-vb2j4" podUID="a28f881c-dd5e-4bf8-82ad-2c27b98dcceb" Oct 02 01:49:56 crc kubenswrapper[4885]: E1002 01:49:56.337869 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-x44t6" podUID="961bc3c9-b6be-4ece-98f1-971a583f9db8" Oct 02 01:49:56 crc kubenswrapper[4885]: E1002 01:49:56.337958 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-vb2j4" podUID="a28f881c-dd5e-4bf8-82ad-2c27b98dcceb" Oct 02 01:49:56 crc kubenswrapper[4885]: E1002 01:49:56.338239 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-vgkg6" podUID="64a7eacf-552b-4a87-91db-4f6f72ee26b1" Oct 02 01:49:56 crc kubenswrapper[4885]: E1002 01:49:56.597861 4885 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Oct 02 01:49:56 crc kubenswrapper[4885]: E1002 01:49:56.598075 4885 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s4hm5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-4p78s_openshift-marketplace(adaf0543-043d-48b8-829f-928df3bf6e90): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 02 01:49:56 crc kubenswrapper[4885]: E1002 01:49:56.599352 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-4p78s" podUID="adaf0543-043d-48b8-829f-928df3bf6e90" Oct 02 01:49:56 crc kubenswrapper[4885]: E1002 01:49:56.702011 4885 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Oct 02 01:49:56 crc kubenswrapper[4885]: E1002 01:49:56.702210 4885 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jqcx5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-bz9hg_openshift-marketplace(aecca2a3-40b0-4666-a86d-d67ecb580e99): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 02 01:49:56 crc kubenswrapper[4885]: E1002 01:49:56.703547 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-bz9hg" podUID="aecca2a3-40b0-4666-a86d-d67ecb580e99" Oct 02 01:50:00 crc kubenswrapper[4885]: E1002 01:50:00.330129 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-bz9hg" podUID="aecca2a3-40b0-4666-a86d-d67ecb580e99" Oct 02 01:50:00 crc kubenswrapper[4885]: E1002 01:50:00.330528 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-4p78s" podUID="adaf0543-043d-48b8-829f-928df3bf6e90" Oct 02 01:50:00 crc kubenswrapper[4885]: I1002 01:50:00.511476 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:50:00 crc kubenswrapper[4885]: I1002 01:50:00.795837 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-sv4md"] Oct 02 01:50:00 crc kubenswrapper[4885]: W1002 01:50:00.804737 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod724a3e49_001c_45a4_a896_82c13e0ee0e7.slice/crio-55420548bed249f3fc653ab0f12c9eb7911453a9c1c4e1929261d72265381cab WatchSource:0}: Error finding container 55420548bed249f3fc653ab0f12c9eb7911453a9c1c4e1929261d72265381cab: Status 404 returned error can't find the container with id 55420548bed249f3fc653ab0f12c9eb7911453a9c1c4e1929261d72265381cab Oct 02 01:50:01 crc kubenswrapper[4885]: E1002 01:50:01.058515 4885 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Oct 02 01:50:01 crc kubenswrapper[4885]: E1002 01:50:01.058880 4885 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-smmgf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-bjxpx_openshift-marketplace(18a2b456-262c-43fd-97c6-bfaf2d658463): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 02 01:50:01 crc kubenswrapper[4885]: E1002 01:50:01.060210 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-bjxpx" podUID="18a2b456-262c-43fd-97c6-bfaf2d658463" Oct 02 01:50:01 crc kubenswrapper[4885]: I1002 01:50:01.160884 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-sv4md" event={"ID":"724a3e49-001c-45a4-a896-82c13e0ee0e7","Type":"ContainerStarted","Data":"55420548bed249f3fc653ab0f12c9eb7911453a9c1c4e1929261d72265381cab"} Oct 02 01:50:01 crc kubenswrapper[4885]: E1002 01:50:01.164513 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-bjxpx" podUID="18a2b456-262c-43fd-97c6-bfaf2d658463" Oct 02 01:50:01 crc kubenswrapper[4885]: E1002 01:50:01.177574 4885 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Oct 02 01:50:01 crc kubenswrapper[4885]: E1002 01:50:01.177788 4885 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-24dnt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-xq6kh_openshift-marketplace(21d974e0-8d62-4daf-b383-5a6823df7b8e): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 02 01:50:01 crc kubenswrapper[4885]: E1002 01:50:01.179655 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-xq6kh" podUID="21d974e0-8d62-4daf-b383-5a6823df7b8e" Oct 02 01:50:02 crc kubenswrapper[4885]: I1002 01:50:02.169363 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-sv4md" event={"ID":"724a3e49-001c-45a4-a896-82c13e0ee0e7","Type":"ContainerStarted","Data":"94732f50484b1c5361aa0cd952af0013d730d1d5d390b1d7d672aaa32ab20732"} Oct 02 01:50:02 crc kubenswrapper[4885]: I1002 01:50:02.169453 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-sv4md" event={"ID":"724a3e49-001c-45a4-a896-82c13e0ee0e7","Type":"ContainerStarted","Data":"49d4db8f2f917fa6747da715aea575d726c929e38b32ae8a6c604b84d55be7e4"} Oct 02 01:50:02 crc kubenswrapper[4885]: E1002 01:50:02.172006 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-xq6kh" podUID="21d974e0-8d62-4daf-b383-5a6823df7b8e" Oct 02 01:50:02 crc kubenswrapper[4885]: I1002 01:50:02.192076 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-sv4md" podStartSLOduration=169.192051296 podStartE2EDuration="2m49.192051296s" podCreationTimestamp="2025-10-02 01:47:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:50:02.189748652 +0000 UTC m=+191.001496081" watchObservedRunningTime="2025-10-02 01:50:02.192051296 +0000 UTC m=+191.003798725" Oct 02 01:50:13 crc kubenswrapper[4885]: I1002 01:50:13.266401 4885 patch_prober.go:28] interesting pod/machine-config-daemon-rttx8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 01:50:13 crc kubenswrapper[4885]: I1002 01:50:13.267237 4885 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 01:50:17 crc kubenswrapper[4885]: I1002 01:50:17.285232 4885 generic.go:334] "Generic (PLEG): container finished" podID="7d3cc54c-759c-43a7-b88c-4c5079b42e45" containerID="2e955f61a17e0c87f71927faece6588958f98246589c3a9407d58dfb4c054422" exitCode=0 Oct 02 01:50:17 crc kubenswrapper[4885]: I1002 01:50:17.285311 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bj65p" event={"ID":"7d3cc54c-759c-43a7-b88c-4c5079b42e45","Type":"ContainerDied","Data":"2e955f61a17e0c87f71927faece6588958f98246589c3a9407d58dfb4c054422"} Oct 02 01:50:17 crc kubenswrapper[4885]: I1002 01:50:17.290543 4885 generic.go:334] "Generic (PLEG): container finished" podID="64a7eacf-552b-4a87-91db-4f6f72ee26b1" containerID="bf75d1be318ecc5be4b9fd1a749a99c9a6d3ff4bd6a55c4bb9af17cc54ef6380" exitCode=0 Oct 02 01:50:17 crc kubenswrapper[4885]: I1002 01:50:17.290635 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vgkg6" event={"ID":"64a7eacf-552b-4a87-91db-4f6f72ee26b1","Type":"ContainerDied","Data":"bf75d1be318ecc5be4b9fd1a749a99c9a6d3ff4bd6a55c4bb9af17cc54ef6380"} Oct 02 01:50:17 crc kubenswrapper[4885]: I1002 01:50:17.294099 4885 generic.go:334] "Generic (PLEG): container finished" podID="a28f881c-dd5e-4bf8-82ad-2c27b98dcceb" containerID="d93f7aa02729b304652b4e31bacbaa2bfd913a75cf17811af175d9bee828a608" exitCode=0 Oct 02 01:50:17 crc kubenswrapper[4885]: I1002 01:50:17.294116 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vb2j4" event={"ID":"a28f881c-dd5e-4bf8-82ad-2c27b98dcceb","Type":"ContainerDied","Data":"d93f7aa02729b304652b4e31bacbaa2bfd913a75cf17811af175d9bee828a608"} Oct 02 01:50:18 crc kubenswrapper[4885]: I1002 01:50:18.300674 4885 generic.go:334] "Generic (PLEG): container finished" podID="adaf0543-043d-48b8-829f-928df3bf6e90" containerID="8dcf03eb4611441fba0f138ac8a4bfae7ed29bc220a6344587e4e94cfee94efd" exitCode=0 Oct 02 01:50:18 crc kubenswrapper[4885]: I1002 01:50:18.300775 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4p78s" event={"ID":"adaf0543-043d-48b8-829f-928df3bf6e90","Type":"ContainerDied","Data":"8dcf03eb4611441fba0f138ac8a4bfae7ed29bc220a6344587e4e94cfee94efd"} Oct 02 01:50:18 crc kubenswrapper[4885]: I1002 01:50:18.303382 4885 generic.go:334] "Generic (PLEG): container finished" podID="aecca2a3-40b0-4666-a86d-d67ecb580e99" containerID="21589b19918287bd89fce80ba0975e43acbe59e01dba132d76542afd723ef0ed" exitCode=0 Oct 02 01:50:18 crc kubenswrapper[4885]: I1002 01:50:18.303453 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bz9hg" event={"ID":"aecca2a3-40b0-4666-a86d-d67ecb580e99","Type":"ContainerDied","Data":"21589b19918287bd89fce80ba0975e43acbe59e01dba132d76542afd723ef0ed"} Oct 02 01:50:18 crc kubenswrapper[4885]: I1002 01:50:18.307005 4885 generic.go:334] "Generic (PLEG): container finished" podID="961bc3c9-b6be-4ece-98f1-971a583f9db8" containerID="25eb9e8df43c4aec5ea67236b9ef11962fafa9abcfcb04b1451581f12d298a85" exitCode=0 Oct 02 01:50:18 crc kubenswrapper[4885]: I1002 01:50:18.307085 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x44t6" event={"ID":"961bc3c9-b6be-4ece-98f1-971a583f9db8","Type":"ContainerDied","Data":"25eb9e8df43c4aec5ea67236b9ef11962fafa9abcfcb04b1451581f12d298a85"} Oct 02 01:50:18 crc kubenswrapper[4885]: I1002 01:50:18.311110 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xq6kh" event={"ID":"21d974e0-8d62-4daf-b383-5a6823df7b8e","Type":"ContainerStarted","Data":"9536fd1813dfd410759d213206f08c30e3bec894a291f9f250baae317d8f06ca"} Oct 02 01:50:18 crc kubenswrapper[4885]: I1002 01:50:18.315823 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bjxpx" event={"ID":"18a2b456-262c-43fd-97c6-bfaf2d658463","Type":"ContainerStarted","Data":"7da19d71cc38343e6a7f281f23d03e3b06ebb566c68d8495c0bfa7179aa213d5"} Oct 02 01:50:19 crc kubenswrapper[4885]: I1002 01:50:19.322760 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x44t6" event={"ID":"961bc3c9-b6be-4ece-98f1-971a583f9db8","Type":"ContainerStarted","Data":"fa91784093ffc8297aea22d51f40aa7fe56e7c4e2ea262606bd94b010b505037"} Oct 02 01:50:19 crc kubenswrapper[4885]: I1002 01:50:19.325709 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bj65p" event={"ID":"7d3cc54c-759c-43a7-b88c-4c5079b42e45","Type":"ContainerStarted","Data":"b3a9fb027da67394537ee2d2c52fc9b9b7f26ae078aa1519c69d227596136a1a"} Oct 02 01:50:19 crc kubenswrapper[4885]: I1002 01:50:19.327790 4885 generic.go:334] "Generic (PLEG): container finished" podID="21d974e0-8d62-4daf-b383-5a6823df7b8e" containerID="9536fd1813dfd410759d213206f08c30e3bec894a291f9f250baae317d8f06ca" exitCode=0 Oct 02 01:50:19 crc kubenswrapper[4885]: I1002 01:50:19.327850 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xq6kh" event={"ID":"21d974e0-8d62-4daf-b383-5a6823df7b8e","Type":"ContainerDied","Data":"9536fd1813dfd410759d213206f08c30e3bec894a291f9f250baae317d8f06ca"} Oct 02 01:50:19 crc kubenswrapper[4885]: I1002 01:50:19.332113 4885 generic.go:334] "Generic (PLEG): container finished" podID="18a2b456-262c-43fd-97c6-bfaf2d658463" containerID="7da19d71cc38343e6a7f281f23d03e3b06ebb566c68d8495c0bfa7179aa213d5" exitCode=0 Oct 02 01:50:19 crc kubenswrapper[4885]: I1002 01:50:19.332167 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bjxpx" event={"ID":"18a2b456-262c-43fd-97c6-bfaf2d658463","Type":"ContainerDied","Data":"7da19d71cc38343e6a7f281f23d03e3b06ebb566c68d8495c0bfa7179aa213d5"} Oct 02 01:50:19 crc kubenswrapper[4885]: I1002 01:50:19.335983 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vgkg6" event={"ID":"64a7eacf-552b-4a87-91db-4f6f72ee26b1","Type":"ContainerStarted","Data":"02d3da012e388fa50bb19ce84d485cb149bf727561d6b60ae1cdf7de944b9a3e"} Oct 02 01:50:19 crc kubenswrapper[4885]: I1002 01:50:19.338168 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vb2j4" event={"ID":"a28f881c-dd5e-4bf8-82ad-2c27b98dcceb","Type":"ContainerStarted","Data":"fe571f9130c9ad468ffd6ca21eef2aa385a83079b41b04f8e414c3e5450e39aa"} Oct 02 01:50:19 crc kubenswrapper[4885]: I1002 01:50:19.340653 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4p78s" event={"ID":"adaf0543-043d-48b8-829f-928df3bf6e90","Type":"ContainerStarted","Data":"9b43b72ac1181cb67f6156477158dacffb97fa65274f1c6447b638a9c7e9aadb"} Oct 02 01:50:19 crc kubenswrapper[4885]: I1002 01:50:19.343427 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bz9hg" event={"ID":"aecca2a3-40b0-4666-a86d-d67ecb580e99","Type":"ContainerStarted","Data":"00335c982112fd4f0ea73d0823b195c015c56ae6305bfcd2e105be3c2c335313"} Oct 02 01:50:19 crc kubenswrapper[4885]: I1002 01:50:19.343993 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-x44t6" podStartSLOduration=3.255621723 podStartE2EDuration="1m0.34398515s" podCreationTimestamp="2025-10-02 01:49:19 +0000 UTC" firstStartedPulling="2025-10-02 01:49:21.806812306 +0000 UTC m=+150.618559705" lastFinishedPulling="2025-10-02 01:50:18.895175733 +0000 UTC m=+207.706923132" observedRunningTime="2025-10-02 01:50:19.33939304 +0000 UTC m=+208.151140439" watchObservedRunningTime="2025-10-02 01:50:19.34398515 +0000 UTC m=+208.155732549" Oct 02 01:50:19 crc kubenswrapper[4885]: I1002 01:50:19.474415 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-bj65p" podStartSLOduration=2.813650707 podStartE2EDuration="59.474400853s" podCreationTimestamp="2025-10-02 01:49:20 +0000 UTC" firstStartedPulling="2025-10-02 01:49:21.828002182 +0000 UTC m=+150.639749581" lastFinishedPulling="2025-10-02 01:50:18.488752298 +0000 UTC m=+207.300499727" observedRunningTime="2025-10-02 01:50:19.473170703 +0000 UTC m=+208.284918112" watchObservedRunningTime="2025-10-02 01:50:19.474400853 +0000 UTC m=+208.286148252" Oct 02 01:50:19 crc kubenswrapper[4885]: I1002 01:50:19.493919 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-vb2j4" podStartSLOduration=2.998017681 podStartE2EDuration="1m0.493905099s" podCreationTimestamp="2025-10-02 01:49:19 +0000 UTC" firstStartedPulling="2025-10-02 01:49:20.75154368 +0000 UTC m=+149.563291079" lastFinishedPulling="2025-10-02 01:50:18.247431058 +0000 UTC m=+207.059178497" observedRunningTime="2025-10-02 01:50:19.49332055 +0000 UTC m=+208.305067949" watchObservedRunningTime="2025-10-02 01:50:19.493905099 +0000 UTC m=+208.305652498" Oct 02 01:50:19 crc kubenswrapper[4885]: I1002 01:50:19.512608 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-vgkg6" podStartSLOduration=3.740814813 podStartE2EDuration="1m0.512592729s" podCreationTimestamp="2025-10-02 01:49:19 +0000 UTC" firstStartedPulling="2025-10-02 01:49:21.82145908 +0000 UTC m=+150.633206479" lastFinishedPulling="2025-10-02 01:50:18.593237006 +0000 UTC m=+207.404984395" observedRunningTime="2025-10-02 01:50:19.510336245 +0000 UTC m=+208.322083644" watchObservedRunningTime="2025-10-02 01:50:19.512592729 +0000 UTC m=+208.324340128" Oct 02 01:50:19 crc kubenswrapper[4885]: I1002 01:50:19.534487 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-bz9hg" podStartSLOduration=2.387601653 podStartE2EDuration="58.534469441s" podCreationTimestamp="2025-10-02 01:49:21 +0000 UTC" firstStartedPulling="2025-10-02 01:49:22.844083109 +0000 UTC m=+151.655830508" lastFinishedPulling="2025-10-02 01:50:18.990950897 +0000 UTC m=+207.802698296" observedRunningTime="2025-10-02 01:50:19.533270543 +0000 UTC m=+208.345017942" watchObservedRunningTime="2025-10-02 01:50:19.534469441 +0000 UTC m=+208.346216840" Oct 02 01:50:19 crc kubenswrapper[4885]: I1002 01:50:19.549211 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-4p78s" podStartSLOduration=1.611027219 podStartE2EDuration="57.549194332s" podCreationTimestamp="2025-10-02 01:49:22 +0000 UTC" firstStartedPulling="2025-10-02 01:49:22.85000284 +0000 UTC m=+151.661750239" lastFinishedPulling="2025-10-02 01:50:18.788169953 +0000 UTC m=+207.599917352" observedRunningTime="2025-10-02 01:50:19.54730322 +0000 UTC m=+208.359050619" watchObservedRunningTime="2025-10-02 01:50:19.549194332 +0000 UTC m=+208.360941731" Oct 02 01:50:19 crc kubenswrapper[4885]: I1002 01:50:19.783233 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-x44t6" Oct 02 01:50:19 crc kubenswrapper[4885]: I1002 01:50:19.783295 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-x44t6" Oct 02 01:50:19 crc kubenswrapper[4885]: I1002 01:50:19.969308 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-vb2j4" Oct 02 01:50:19 crc kubenswrapper[4885]: I1002 01:50:19.969348 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-vb2j4" Oct 02 01:50:20 crc kubenswrapper[4885]: I1002 01:50:20.188618 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-vgkg6" Oct 02 01:50:20 crc kubenswrapper[4885]: I1002 01:50:20.188672 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-vgkg6" Oct 02 01:50:20 crc kubenswrapper[4885]: I1002 01:50:20.350220 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xq6kh" event={"ID":"21d974e0-8d62-4daf-b383-5a6823df7b8e","Type":"ContainerStarted","Data":"9061f900e18e6b5622b77850166496ba5b75f7f46f9570ef6a593e6f66e7a2ed"} Oct 02 01:50:20 crc kubenswrapper[4885]: I1002 01:50:20.352514 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bjxpx" event={"ID":"18a2b456-262c-43fd-97c6-bfaf2d658463","Type":"ContainerStarted","Data":"9248711d02d76c11268438ac426cc0097b451fcdadac403d61ebe565076fff04"} Oct 02 01:50:20 crc kubenswrapper[4885]: I1002 01:50:20.376537 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-xq6kh" podStartSLOduration=2.41037517 podStartE2EDuration="57.376517383s" podCreationTimestamp="2025-10-02 01:49:23 +0000 UTC" firstStartedPulling="2025-10-02 01:49:24.883216498 +0000 UTC m=+153.694963897" lastFinishedPulling="2025-10-02 01:50:19.849358711 +0000 UTC m=+208.661106110" observedRunningTime="2025-10-02 01:50:20.375479739 +0000 UTC m=+209.187227128" watchObservedRunningTime="2025-10-02 01:50:20.376517383 +0000 UTC m=+209.188264782" Oct 02 01:50:20 crc kubenswrapper[4885]: I1002 01:50:20.394116 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-bjxpx" podStartSLOduration=2.54044448 podStartE2EDuration="58.394099207s" podCreationTimestamp="2025-10-02 01:49:22 +0000 UTC" firstStartedPulling="2025-10-02 01:49:23.867290476 +0000 UTC m=+152.679037875" lastFinishedPulling="2025-10-02 01:50:19.720945203 +0000 UTC m=+208.532692602" observedRunningTime="2025-10-02 01:50:20.392827395 +0000 UTC m=+209.204574794" watchObservedRunningTime="2025-10-02 01:50:20.394099207 +0000 UTC m=+209.205846606" Oct 02 01:50:20 crc kubenswrapper[4885]: I1002 01:50:20.419012 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-bj65p" Oct 02 01:50:20 crc kubenswrapper[4885]: I1002 01:50:20.419066 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-bj65p" Oct 02 01:50:20 crc kubenswrapper[4885]: I1002 01:50:20.929562 4885 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-x44t6" podUID="961bc3c9-b6be-4ece-98f1-971a583f9db8" containerName="registry-server" probeResult="failure" output=< Oct 02 01:50:20 crc kubenswrapper[4885]: timeout: failed to connect service ":50051" within 1s Oct 02 01:50:20 crc kubenswrapper[4885]: > Oct 02 01:50:21 crc kubenswrapper[4885]: I1002 01:50:21.013651 4885 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-vb2j4" podUID="a28f881c-dd5e-4bf8-82ad-2c27b98dcceb" containerName="registry-server" probeResult="failure" output=< Oct 02 01:50:21 crc kubenswrapper[4885]: timeout: failed to connect service ":50051" within 1s Oct 02 01:50:21 crc kubenswrapper[4885]: > Oct 02 01:50:21 crc kubenswrapper[4885]: I1002 01:50:21.226236 4885 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-vgkg6" podUID="64a7eacf-552b-4a87-91db-4f6f72ee26b1" containerName="registry-server" probeResult="failure" output=< Oct 02 01:50:21 crc kubenswrapper[4885]: timeout: failed to connect service ":50051" within 1s Oct 02 01:50:21 crc kubenswrapper[4885]: > Oct 02 01:50:21 crc kubenswrapper[4885]: I1002 01:50:21.455076 4885 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-bj65p" podUID="7d3cc54c-759c-43a7-b88c-4c5079b42e45" containerName="registry-server" probeResult="failure" output=< Oct 02 01:50:21 crc kubenswrapper[4885]: timeout: failed to connect service ":50051" within 1s Oct 02 01:50:21 crc kubenswrapper[4885]: > Oct 02 01:50:21 crc kubenswrapper[4885]: I1002 01:50:21.994508 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-bz9hg" Oct 02 01:50:21 crc kubenswrapper[4885]: I1002 01:50:21.994588 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-bz9hg" Oct 02 01:50:22 crc kubenswrapper[4885]: I1002 01:50:22.069514 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-bz9hg" Oct 02 01:50:22 crc kubenswrapper[4885]: I1002 01:50:22.389986 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-4p78s" Oct 02 01:50:22 crc kubenswrapper[4885]: I1002 01:50:22.390058 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-4p78s" Oct 02 01:50:22 crc kubenswrapper[4885]: I1002 01:50:22.456878 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-4p78s" Oct 02 01:50:22 crc kubenswrapper[4885]: I1002 01:50:22.979128 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-bjxpx" Oct 02 01:50:22 crc kubenswrapper[4885]: I1002 01:50:22.979988 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-bjxpx" Oct 02 01:50:23 crc kubenswrapper[4885]: I1002 01:50:23.350491 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-xq6kh" Oct 02 01:50:23 crc kubenswrapper[4885]: I1002 01:50:23.350815 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-xq6kh" Oct 02 01:50:23 crc kubenswrapper[4885]: I1002 01:50:23.428207 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-4p78s" Oct 02 01:50:24 crc kubenswrapper[4885]: I1002 01:50:24.023198 4885 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-bjxpx" podUID="18a2b456-262c-43fd-97c6-bfaf2d658463" containerName="registry-server" probeResult="failure" output=< Oct 02 01:50:24 crc kubenswrapper[4885]: timeout: failed to connect service ":50051" within 1s Oct 02 01:50:24 crc kubenswrapper[4885]: > Oct 02 01:50:24 crc kubenswrapper[4885]: I1002 01:50:24.406246 4885 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-xq6kh" podUID="21d974e0-8d62-4daf-b383-5a6823df7b8e" containerName="registry-server" probeResult="failure" output=< Oct 02 01:50:24 crc kubenswrapper[4885]: timeout: failed to connect service ":50051" within 1s Oct 02 01:50:24 crc kubenswrapper[4885]: > Oct 02 01:50:24 crc kubenswrapper[4885]: I1002 01:50:24.529294 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4p78s"] Oct 02 01:50:25 crc kubenswrapper[4885]: I1002 01:50:25.377965 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-4p78s" podUID="adaf0543-043d-48b8-829f-928df3bf6e90" containerName="registry-server" containerID="cri-o://9b43b72ac1181cb67f6156477158dacffb97fa65274f1c6447b638a9c7e9aadb" gracePeriod=2 Oct 02 01:50:25 crc kubenswrapper[4885]: I1002 01:50:25.768382 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4p78s" Oct 02 01:50:25 crc kubenswrapper[4885]: I1002 01:50:25.879483 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/adaf0543-043d-48b8-829f-928df3bf6e90-utilities\") pod \"adaf0543-043d-48b8-829f-928df3bf6e90\" (UID: \"adaf0543-043d-48b8-829f-928df3bf6e90\") " Oct 02 01:50:25 crc kubenswrapper[4885]: I1002 01:50:25.879578 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/adaf0543-043d-48b8-829f-928df3bf6e90-catalog-content\") pod \"adaf0543-043d-48b8-829f-928df3bf6e90\" (UID: \"adaf0543-043d-48b8-829f-928df3bf6e90\") " Oct 02 01:50:25 crc kubenswrapper[4885]: I1002 01:50:25.879704 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4hm5\" (UniqueName: \"kubernetes.io/projected/adaf0543-043d-48b8-829f-928df3bf6e90-kube-api-access-s4hm5\") pod \"adaf0543-043d-48b8-829f-928df3bf6e90\" (UID: \"adaf0543-043d-48b8-829f-928df3bf6e90\") " Oct 02 01:50:25 crc kubenswrapper[4885]: I1002 01:50:25.880189 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/adaf0543-043d-48b8-829f-928df3bf6e90-utilities" (OuterVolumeSpecName: "utilities") pod "adaf0543-043d-48b8-829f-928df3bf6e90" (UID: "adaf0543-043d-48b8-829f-928df3bf6e90"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 01:50:25 crc kubenswrapper[4885]: I1002 01:50:25.888449 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/adaf0543-043d-48b8-829f-928df3bf6e90-kube-api-access-s4hm5" (OuterVolumeSpecName: "kube-api-access-s4hm5") pod "adaf0543-043d-48b8-829f-928df3bf6e90" (UID: "adaf0543-043d-48b8-829f-928df3bf6e90"). InnerVolumeSpecName "kube-api-access-s4hm5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:50:25 crc kubenswrapper[4885]: I1002 01:50:25.893682 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/adaf0543-043d-48b8-829f-928df3bf6e90-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "adaf0543-043d-48b8-829f-928df3bf6e90" (UID: "adaf0543-043d-48b8-829f-928df3bf6e90"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 01:50:25 crc kubenswrapper[4885]: I1002 01:50:25.981489 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4hm5\" (UniqueName: \"kubernetes.io/projected/adaf0543-043d-48b8-829f-928df3bf6e90-kube-api-access-s4hm5\") on node \"crc\" DevicePath \"\"" Oct 02 01:50:25 crc kubenswrapper[4885]: I1002 01:50:25.981522 4885 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/adaf0543-043d-48b8-829f-928df3bf6e90-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 01:50:25 crc kubenswrapper[4885]: I1002 01:50:25.981532 4885 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/adaf0543-043d-48b8-829f-928df3bf6e90-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 01:50:26 crc kubenswrapper[4885]: I1002 01:50:26.384700 4885 generic.go:334] "Generic (PLEG): container finished" podID="adaf0543-043d-48b8-829f-928df3bf6e90" containerID="9b43b72ac1181cb67f6156477158dacffb97fa65274f1c6447b638a9c7e9aadb" exitCode=0 Oct 02 01:50:26 crc kubenswrapper[4885]: I1002 01:50:26.384772 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4p78s" Oct 02 01:50:26 crc kubenswrapper[4885]: I1002 01:50:26.384794 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4p78s" event={"ID":"adaf0543-043d-48b8-829f-928df3bf6e90","Type":"ContainerDied","Data":"9b43b72ac1181cb67f6156477158dacffb97fa65274f1c6447b638a9c7e9aadb"} Oct 02 01:50:26 crc kubenswrapper[4885]: I1002 01:50:26.385063 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4p78s" event={"ID":"adaf0543-043d-48b8-829f-928df3bf6e90","Type":"ContainerDied","Data":"5ea24c226c89380df68904a299e46b054f2178d9e00a0b3dc5637146f62742a9"} Oct 02 01:50:26 crc kubenswrapper[4885]: I1002 01:50:26.385088 4885 scope.go:117] "RemoveContainer" containerID="9b43b72ac1181cb67f6156477158dacffb97fa65274f1c6447b638a9c7e9aadb" Oct 02 01:50:26 crc kubenswrapper[4885]: I1002 01:50:26.402517 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4p78s"] Oct 02 01:50:26 crc kubenswrapper[4885]: I1002 01:50:26.403710 4885 scope.go:117] "RemoveContainer" containerID="8dcf03eb4611441fba0f138ac8a4bfae7ed29bc220a6344587e4e94cfee94efd" Oct 02 01:50:26 crc kubenswrapper[4885]: I1002 01:50:26.407959 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-4p78s"] Oct 02 01:50:26 crc kubenswrapper[4885]: I1002 01:50:26.419008 4885 scope.go:117] "RemoveContainer" containerID="6759585f89e100d2eec13552ca691f04cac3cd733d99323bed065d6cb55b0b70" Oct 02 01:50:26 crc kubenswrapper[4885]: I1002 01:50:26.440792 4885 scope.go:117] "RemoveContainer" containerID="9b43b72ac1181cb67f6156477158dacffb97fa65274f1c6447b638a9c7e9aadb" Oct 02 01:50:26 crc kubenswrapper[4885]: E1002 01:50:26.441134 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9b43b72ac1181cb67f6156477158dacffb97fa65274f1c6447b638a9c7e9aadb\": container with ID starting with 9b43b72ac1181cb67f6156477158dacffb97fa65274f1c6447b638a9c7e9aadb not found: ID does not exist" containerID="9b43b72ac1181cb67f6156477158dacffb97fa65274f1c6447b638a9c7e9aadb" Oct 02 01:50:26 crc kubenswrapper[4885]: I1002 01:50:26.441174 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b43b72ac1181cb67f6156477158dacffb97fa65274f1c6447b638a9c7e9aadb"} err="failed to get container status \"9b43b72ac1181cb67f6156477158dacffb97fa65274f1c6447b638a9c7e9aadb\": rpc error: code = NotFound desc = could not find container \"9b43b72ac1181cb67f6156477158dacffb97fa65274f1c6447b638a9c7e9aadb\": container with ID starting with 9b43b72ac1181cb67f6156477158dacffb97fa65274f1c6447b638a9c7e9aadb not found: ID does not exist" Oct 02 01:50:26 crc kubenswrapper[4885]: I1002 01:50:26.441223 4885 scope.go:117] "RemoveContainer" containerID="8dcf03eb4611441fba0f138ac8a4bfae7ed29bc220a6344587e4e94cfee94efd" Oct 02 01:50:26 crc kubenswrapper[4885]: E1002 01:50:26.441475 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8dcf03eb4611441fba0f138ac8a4bfae7ed29bc220a6344587e4e94cfee94efd\": container with ID starting with 8dcf03eb4611441fba0f138ac8a4bfae7ed29bc220a6344587e4e94cfee94efd not found: ID does not exist" containerID="8dcf03eb4611441fba0f138ac8a4bfae7ed29bc220a6344587e4e94cfee94efd" Oct 02 01:50:26 crc kubenswrapper[4885]: I1002 01:50:26.441511 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8dcf03eb4611441fba0f138ac8a4bfae7ed29bc220a6344587e4e94cfee94efd"} err="failed to get container status \"8dcf03eb4611441fba0f138ac8a4bfae7ed29bc220a6344587e4e94cfee94efd\": rpc error: code = NotFound desc = could not find container \"8dcf03eb4611441fba0f138ac8a4bfae7ed29bc220a6344587e4e94cfee94efd\": container with ID starting with 8dcf03eb4611441fba0f138ac8a4bfae7ed29bc220a6344587e4e94cfee94efd not found: ID does not exist" Oct 02 01:50:26 crc kubenswrapper[4885]: I1002 01:50:26.441533 4885 scope.go:117] "RemoveContainer" containerID="6759585f89e100d2eec13552ca691f04cac3cd733d99323bed065d6cb55b0b70" Oct 02 01:50:26 crc kubenswrapper[4885]: E1002 01:50:26.441791 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6759585f89e100d2eec13552ca691f04cac3cd733d99323bed065d6cb55b0b70\": container with ID starting with 6759585f89e100d2eec13552ca691f04cac3cd733d99323bed065d6cb55b0b70 not found: ID does not exist" containerID="6759585f89e100d2eec13552ca691f04cac3cd733d99323bed065d6cb55b0b70" Oct 02 01:50:26 crc kubenswrapper[4885]: I1002 01:50:26.441820 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6759585f89e100d2eec13552ca691f04cac3cd733d99323bed065d6cb55b0b70"} err="failed to get container status \"6759585f89e100d2eec13552ca691f04cac3cd733d99323bed065d6cb55b0b70\": rpc error: code = NotFound desc = could not find container \"6759585f89e100d2eec13552ca691f04cac3cd733d99323bed065d6cb55b0b70\": container with ID starting with 6759585f89e100d2eec13552ca691f04cac3cd733d99323bed065d6cb55b0b70 not found: ID does not exist" Oct 02 01:50:28 crc kubenswrapper[4885]: I1002 01:50:28.058900 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="adaf0543-043d-48b8-829f-928df3bf6e90" path="/var/lib/kubelet/pods/adaf0543-043d-48b8-829f-928df3bf6e90/volumes" Oct 02 01:50:29 crc kubenswrapper[4885]: I1002 01:50:29.842942 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-x44t6" Oct 02 01:50:29 crc kubenswrapper[4885]: I1002 01:50:29.902276 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-x44t6" Oct 02 01:50:30 crc kubenswrapper[4885]: I1002 01:50:30.022567 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-vb2j4" Oct 02 01:50:30 crc kubenswrapper[4885]: I1002 01:50:30.070799 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-vb2j4" Oct 02 01:50:30 crc kubenswrapper[4885]: I1002 01:50:30.234124 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-vgkg6" Oct 02 01:50:30 crc kubenswrapper[4885]: I1002 01:50:30.283488 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-vgkg6" Oct 02 01:50:30 crc kubenswrapper[4885]: I1002 01:50:30.455355 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-bj65p" Oct 02 01:50:30 crc kubenswrapper[4885]: I1002 01:50:30.493169 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-bj65p" Oct 02 01:50:31 crc kubenswrapper[4885]: I1002 01:50:31.880521 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-bj65p"] Oct 02 01:50:32 crc kubenswrapper[4885]: I1002 01:50:32.045369 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-bz9hg" Oct 02 01:50:32 crc kubenswrapper[4885]: I1002 01:50:32.423089 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-bj65p" podUID="7d3cc54c-759c-43a7-b88c-4c5079b42e45" containerName="registry-server" containerID="cri-o://b3a9fb027da67394537ee2d2c52fc9b9b7f26ae078aa1519c69d227596136a1a" gracePeriod=2 Oct 02 01:50:32 crc kubenswrapper[4885]: I1002 01:50:32.482945 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vgkg6"] Oct 02 01:50:32 crc kubenswrapper[4885]: I1002 01:50:32.483314 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-vgkg6" podUID="64a7eacf-552b-4a87-91db-4f6f72ee26b1" containerName="registry-server" containerID="cri-o://02d3da012e388fa50bb19ce84d485cb149bf727561d6b60ae1cdf7de944b9a3e" gracePeriod=2 Oct 02 01:50:33 crc kubenswrapper[4885]: I1002 01:50:33.036224 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-bjxpx" Oct 02 01:50:33 crc kubenswrapper[4885]: I1002 01:50:33.086216 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-bjxpx" Oct 02 01:50:33 crc kubenswrapper[4885]: I1002 01:50:33.404202 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-xq6kh" Oct 02 01:50:33 crc kubenswrapper[4885]: I1002 01:50:33.463905 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-xq6kh" Oct 02 01:50:33 crc kubenswrapper[4885]: I1002 01:50:33.501472 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-mc59z"] Oct 02 01:50:34 crc kubenswrapper[4885]: I1002 01:50:34.432315 4885 generic.go:334] "Generic (PLEG): container finished" podID="7d3cc54c-759c-43a7-b88c-4c5079b42e45" containerID="b3a9fb027da67394537ee2d2c52fc9b9b7f26ae078aa1519c69d227596136a1a" exitCode=0 Oct 02 01:50:34 crc kubenswrapper[4885]: I1002 01:50:34.432366 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bj65p" event={"ID":"7d3cc54c-759c-43a7-b88c-4c5079b42e45","Type":"ContainerDied","Data":"b3a9fb027da67394537ee2d2c52fc9b9b7f26ae078aa1519c69d227596136a1a"} Oct 02 01:50:34 crc kubenswrapper[4885]: I1002 01:50:34.435372 4885 generic.go:334] "Generic (PLEG): container finished" podID="64a7eacf-552b-4a87-91db-4f6f72ee26b1" containerID="02d3da012e388fa50bb19ce84d485cb149bf727561d6b60ae1cdf7de944b9a3e" exitCode=0 Oct 02 01:50:34 crc kubenswrapper[4885]: I1002 01:50:34.435411 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vgkg6" event={"ID":"64a7eacf-552b-4a87-91db-4f6f72ee26b1","Type":"ContainerDied","Data":"02d3da012e388fa50bb19ce84d485cb149bf727561d6b60ae1cdf7de944b9a3e"} Oct 02 01:50:35 crc kubenswrapper[4885]: I1002 01:50:35.162779 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bj65p" Oct 02 01:50:35 crc kubenswrapper[4885]: I1002 01:50:35.236182 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vgkg6" Oct 02 01:50:35 crc kubenswrapper[4885]: I1002 01:50:35.290946 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hjnzp\" (UniqueName: \"kubernetes.io/projected/7d3cc54c-759c-43a7-b88c-4c5079b42e45-kube-api-access-hjnzp\") pod \"7d3cc54c-759c-43a7-b88c-4c5079b42e45\" (UID: \"7d3cc54c-759c-43a7-b88c-4c5079b42e45\") " Oct 02 01:50:35 crc kubenswrapper[4885]: I1002 01:50:35.291242 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d3cc54c-759c-43a7-b88c-4c5079b42e45-catalog-content\") pod \"7d3cc54c-759c-43a7-b88c-4c5079b42e45\" (UID: \"7d3cc54c-759c-43a7-b88c-4c5079b42e45\") " Oct 02 01:50:35 crc kubenswrapper[4885]: I1002 01:50:35.291362 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d3cc54c-759c-43a7-b88c-4c5079b42e45-utilities\") pod \"7d3cc54c-759c-43a7-b88c-4c5079b42e45\" (UID: \"7d3cc54c-759c-43a7-b88c-4c5079b42e45\") " Oct 02 01:50:35 crc kubenswrapper[4885]: I1002 01:50:35.292155 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7d3cc54c-759c-43a7-b88c-4c5079b42e45-utilities" (OuterVolumeSpecName: "utilities") pod "7d3cc54c-759c-43a7-b88c-4c5079b42e45" (UID: "7d3cc54c-759c-43a7-b88c-4c5079b42e45"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 01:50:35 crc kubenswrapper[4885]: I1002 01:50:35.298558 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d3cc54c-759c-43a7-b88c-4c5079b42e45-kube-api-access-hjnzp" (OuterVolumeSpecName: "kube-api-access-hjnzp") pod "7d3cc54c-759c-43a7-b88c-4c5079b42e45" (UID: "7d3cc54c-759c-43a7-b88c-4c5079b42e45"). InnerVolumeSpecName "kube-api-access-hjnzp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:50:35 crc kubenswrapper[4885]: I1002 01:50:35.330805 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7d3cc54c-759c-43a7-b88c-4c5079b42e45-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7d3cc54c-759c-43a7-b88c-4c5079b42e45" (UID: "7d3cc54c-759c-43a7-b88c-4c5079b42e45"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 01:50:35 crc kubenswrapper[4885]: I1002 01:50:35.392340 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f94th\" (UniqueName: \"kubernetes.io/projected/64a7eacf-552b-4a87-91db-4f6f72ee26b1-kube-api-access-f94th\") pod \"64a7eacf-552b-4a87-91db-4f6f72ee26b1\" (UID: \"64a7eacf-552b-4a87-91db-4f6f72ee26b1\") " Oct 02 01:50:35 crc kubenswrapper[4885]: I1002 01:50:35.392439 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/64a7eacf-552b-4a87-91db-4f6f72ee26b1-utilities\") pod \"64a7eacf-552b-4a87-91db-4f6f72ee26b1\" (UID: \"64a7eacf-552b-4a87-91db-4f6f72ee26b1\") " Oct 02 01:50:35 crc kubenswrapper[4885]: I1002 01:50:35.392484 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/64a7eacf-552b-4a87-91db-4f6f72ee26b1-catalog-content\") pod \"64a7eacf-552b-4a87-91db-4f6f72ee26b1\" (UID: \"64a7eacf-552b-4a87-91db-4f6f72ee26b1\") " Oct 02 01:50:35 crc kubenswrapper[4885]: I1002 01:50:35.392721 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hjnzp\" (UniqueName: \"kubernetes.io/projected/7d3cc54c-759c-43a7-b88c-4c5079b42e45-kube-api-access-hjnzp\") on node \"crc\" DevicePath \"\"" Oct 02 01:50:35 crc kubenswrapper[4885]: I1002 01:50:35.392742 4885 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d3cc54c-759c-43a7-b88c-4c5079b42e45-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 01:50:35 crc kubenswrapper[4885]: I1002 01:50:35.392754 4885 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d3cc54c-759c-43a7-b88c-4c5079b42e45-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 01:50:35 crc kubenswrapper[4885]: I1002 01:50:35.393965 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/64a7eacf-552b-4a87-91db-4f6f72ee26b1-utilities" (OuterVolumeSpecName: "utilities") pod "64a7eacf-552b-4a87-91db-4f6f72ee26b1" (UID: "64a7eacf-552b-4a87-91db-4f6f72ee26b1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 01:50:35 crc kubenswrapper[4885]: I1002 01:50:35.396536 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/64a7eacf-552b-4a87-91db-4f6f72ee26b1-kube-api-access-f94th" (OuterVolumeSpecName: "kube-api-access-f94th") pod "64a7eacf-552b-4a87-91db-4f6f72ee26b1" (UID: "64a7eacf-552b-4a87-91db-4f6f72ee26b1"). InnerVolumeSpecName "kube-api-access-f94th". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:50:35 crc kubenswrapper[4885]: I1002 01:50:35.443979 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vgkg6" event={"ID":"64a7eacf-552b-4a87-91db-4f6f72ee26b1","Type":"ContainerDied","Data":"9d0cdc00b0c36d81cf64f9787b38b25b56156fac9ffba5fc555fcaadacdf7333"} Oct 02 01:50:35 crc kubenswrapper[4885]: I1002 01:50:35.444020 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vgkg6" Oct 02 01:50:35 crc kubenswrapper[4885]: I1002 01:50:35.444072 4885 scope.go:117] "RemoveContainer" containerID="02d3da012e388fa50bb19ce84d485cb149bf727561d6b60ae1cdf7de944b9a3e" Oct 02 01:50:35 crc kubenswrapper[4885]: I1002 01:50:35.447168 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/64a7eacf-552b-4a87-91db-4f6f72ee26b1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "64a7eacf-552b-4a87-91db-4f6f72ee26b1" (UID: "64a7eacf-552b-4a87-91db-4f6f72ee26b1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 01:50:35 crc kubenswrapper[4885]: I1002 01:50:35.450093 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bj65p" event={"ID":"7d3cc54c-759c-43a7-b88c-4c5079b42e45","Type":"ContainerDied","Data":"7f12910935271cd413d9748a6c302ee159ee903fc91afb8fa8b128b8b7352739"} Oct 02 01:50:35 crc kubenswrapper[4885]: I1002 01:50:35.450168 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bj65p" Oct 02 01:50:35 crc kubenswrapper[4885]: I1002 01:50:35.474182 4885 scope.go:117] "RemoveContainer" containerID="bf75d1be318ecc5be4b9fd1a749a99c9a6d3ff4bd6a55c4bb9af17cc54ef6380" Oct 02 01:50:35 crc kubenswrapper[4885]: I1002 01:50:35.484485 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-bj65p"] Oct 02 01:50:35 crc kubenswrapper[4885]: I1002 01:50:35.486853 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-bj65p"] Oct 02 01:50:35 crc kubenswrapper[4885]: I1002 01:50:35.493759 4885 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/64a7eacf-552b-4a87-91db-4f6f72ee26b1-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 01:50:35 crc kubenswrapper[4885]: I1002 01:50:35.493804 4885 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/64a7eacf-552b-4a87-91db-4f6f72ee26b1-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 01:50:35 crc kubenswrapper[4885]: I1002 01:50:35.493823 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f94th\" (UniqueName: \"kubernetes.io/projected/64a7eacf-552b-4a87-91db-4f6f72ee26b1-kube-api-access-f94th\") on node \"crc\" DevicePath \"\"" Oct 02 01:50:35 crc kubenswrapper[4885]: I1002 01:50:35.501564 4885 scope.go:117] "RemoveContainer" containerID="a6a30ee6e64ab13e1b7f07dd3d2a9163a36b9d194dcf05693ef9ec4f453688b7" Oct 02 01:50:35 crc kubenswrapper[4885]: I1002 01:50:35.522026 4885 scope.go:117] "RemoveContainer" containerID="b3a9fb027da67394537ee2d2c52fc9b9b7f26ae078aa1519c69d227596136a1a" Oct 02 01:50:35 crc kubenswrapper[4885]: I1002 01:50:35.538828 4885 scope.go:117] "RemoveContainer" containerID="2e955f61a17e0c87f71927faece6588958f98246589c3a9407d58dfb4c054422" Oct 02 01:50:35 crc kubenswrapper[4885]: I1002 01:50:35.562233 4885 scope.go:117] "RemoveContainer" containerID="7487c5f977be380c0eebe25d7111c9a0ab055e691c511a081b77dedca2881541" Oct 02 01:50:35 crc kubenswrapper[4885]: I1002 01:50:35.783126 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vgkg6"] Oct 02 01:50:35 crc kubenswrapper[4885]: I1002 01:50:35.785370 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-vgkg6"] Oct 02 01:50:36 crc kubenswrapper[4885]: I1002 01:50:36.057567 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="64a7eacf-552b-4a87-91db-4f6f72ee26b1" path="/var/lib/kubelet/pods/64a7eacf-552b-4a87-91db-4f6f72ee26b1/volumes" Oct 02 01:50:36 crc kubenswrapper[4885]: I1002 01:50:36.059428 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7d3cc54c-759c-43a7-b88c-4c5079b42e45" path="/var/lib/kubelet/pods/7d3cc54c-759c-43a7-b88c-4c5079b42e45/volumes" Oct 02 01:50:36 crc kubenswrapper[4885]: I1002 01:50:36.880161 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xq6kh"] Oct 02 01:50:36 crc kubenswrapper[4885]: I1002 01:50:36.880422 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-xq6kh" podUID="21d974e0-8d62-4daf-b383-5a6823df7b8e" containerName="registry-server" containerID="cri-o://9061f900e18e6b5622b77850166496ba5b75f7f46f9570ef6a593e6f66e7a2ed" gracePeriod=2 Oct 02 01:50:37 crc kubenswrapper[4885]: I1002 01:50:37.329080 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xq6kh" Oct 02 01:50:37 crc kubenswrapper[4885]: I1002 01:50:37.417482 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/21d974e0-8d62-4daf-b383-5a6823df7b8e-catalog-content\") pod \"21d974e0-8d62-4daf-b383-5a6823df7b8e\" (UID: \"21d974e0-8d62-4daf-b383-5a6823df7b8e\") " Oct 02 01:50:37 crc kubenswrapper[4885]: I1002 01:50:37.417636 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/21d974e0-8d62-4daf-b383-5a6823df7b8e-utilities\") pod \"21d974e0-8d62-4daf-b383-5a6823df7b8e\" (UID: \"21d974e0-8d62-4daf-b383-5a6823df7b8e\") " Oct 02 01:50:37 crc kubenswrapper[4885]: I1002 01:50:37.417811 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-24dnt\" (UniqueName: \"kubernetes.io/projected/21d974e0-8d62-4daf-b383-5a6823df7b8e-kube-api-access-24dnt\") pod \"21d974e0-8d62-4daf-b383-5a6823df7b8e\" (UID: \"21d974e0-8d62-4daf-b383-5a6823df7b8e\") " Oct 02 01:50:37 crc kubenswrapper[4885]: I1002 01:50:37.418289 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/21d974e0-8d62-4daf-b383-5a6823df7b8e-utilities" (OuterVolumeSpecName: "utilities") pod "21d974e0-8d62-4daf-b383-5a6823df7b8e" (UID: "21d974e0-8d62-4daf-b383-5a6823df7b8e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 01:50:37 crc kubenswrapper[4885]: I1002 01:50:37.423059 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/21d974e0-8d62-4daf-b383-5a6823df7b8e-kube-api-access-24dnt" (OuterVolumeSpecName: "kube-api-access-24dnt") pod "21d974e0-8d62-4daf-b383-5a6823df7b8e" (UID: "21d974e0-8d62-4daf-b383-5a6823df7b8e"). InnerVolumeSpecName "kube-api-access-24dnt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:50:37 crc kubenswrapper[4885]: I1002 01:50:37.465583 4885 generic.go:334] "Generic (PLEG): container finished" podID="21d974e0-8d62-4daf-b383-5a6823df7b8e" containerID="9061f900e18e6b5622b77850166496ba5b75f7f46f9570ef6a593e6f66e7a2ed" exitCode=0 Oct 02 01:50:37 crc kubenswrapper[4885]: I1002 01:50:37.465647 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xq6kh" event={"ID":"21d974e0-8d62-4daf-b383-5a6823df7b8e","Type":"ContainerDied","Data":"9061f900e18e6b5622b77850166496ba5b75f7f46f9570ef6a593e6f66e7a2ed"} Oct 02 01:50:37 crc kubenswrapper[4885]: I1002 01:50:37.465674 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xq6kh" Oct 02 01:50:37 crc kubenswrapper[4885]: I1002 01:50:37.465696 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xq6kh" event={"ID":"21d974e0-8d62-4daf-b383-5a6823df7b8e","Type":"ContainerDied","Data":"78033f0e604969f00abf669646a13ff7757a6a01337a3d9270fe18522c086da0"} Oct 02 01:50:37 crc kubenswrapper[4885]: I1002 01:50:37.465727 4885 scope.go:117] "RemoveContainer" containerID="9061f900e18e6b5622b77850166496ba5b75f7f46f9570ef6a593e6f66e7a2ed" Oct 02 01:50:37 crc kubenswrapper[4885]: I1002 01:50:37.480252 4885 scope.go:117] "RemoveContainer" containerID="9536fd1813dfd410759d213206f08c30e3bec894a291f9f250baae317d8f06ca" Oct 02 01:50:37 crc kubenswrapper[4885]: I1002 01:50:37.496801 4885 scope.go:117] "RemoveContainer" containerID="7e5d8fe53556ae09ff185262c2de844873acae065f72e9740d554d72ff3968e1" Oct 02 01:50:37 crc kubenswrapper[4885]: I1002 01:50:37.505124 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/21d974e0-8d62-4daf-b383-5a6823df7b8e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "21d974e0-8d62-4daf-b383-5a6823df7b8e" (UID: "21d974e0-8d62-4daf-b383-5a6823df7b8e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 01:50:37 crc kubenswrapper[4885]: I1002 01:50:37.517944 4885 scope.go:117] "RemoveContainer" containerID="9061f900e18e6b5622b77850166496ba5b75f7f46f9570ef6a593e6f66e7a2ed" Oct 02 01:50:37 crc kubenswrapper[4885]: E1002 01:50:37.518303 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9061f900e18e6b5622b77850166496ba5b75f7f46f9570ef6a593e6f66e7a2ed\": container with ID starting with 9061f900e18e6b5622b77850166496ba5b75f7f46f9570ef6a593e6f66e7a2ed not found: ID does not exist" containerID="9061f900e18e6b5622b77850166496ba5b75f7f46f9570ef6a593e6f66e7a2ed" Oct 02 01:50:37 crc kubenswrapper[4885]: I1002 01:50:37.518343 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9061f900e18e6b5622b77850166496ba5b75f7f46f9570ef6a593e6f66e7a2ed"} err="failed to get container status \"9061f900e18e6b5622b77850166496ba5b75f7f46f9570ef6a593e6f66e7a2ed\": rpc error: code = NotFound desc = could not find container \"9061f900e18e6b5622b77850166496ba5b75f7f46f9570ef6a593e6f66e7a2ed\": container with ID starting with 9061f900e18e6b5622b77850166496ba5b75f7f46f9570ef6a593e6f66e7a2ed not found: ID does not exist" Oct 02 01:50:37 crc kubenswrapper[4885]: I1002 01:50:37.518370 4885 scope.go:117] "RemoveContainer" containerID="9536fd1813dfd410759d213206f08c30e3bec894a291f9f250baae317d8f06ca" Oct 02 01:50:37 crc kubenswrapper[4885]: E1002 01:50:37.518688 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9536fd1813dfd410759d213206f08c30e3bec894a291f9f250baae317d8f06ca\": container with ID starting with 9536fd1813dfd410759d213206f08c30e3bec894a291f9f250baae317d8f06ca not found: ID does not exist" containerID="9536fd1813dfd410759d213206f08c30e3bec894a291f9f250baae317d8f06ca" Oct 02 01:50:37 crc kubenswrapper[4885]: I1002 01:50:37.518710 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9536fd1813dfd410759d213206f08c30e3bec894a291f9f250baae317d8f06ca"} err="failed to get container status \"9536fd1813dfd410759d213206f08c30e3bec894a291f9f250baae317d8f06ca\": rpc error: code = NotFound desc = could not find container \"9536fd1813dfd410759d213206f08c30e3bec894a291f9f250baae317d8f06ca\": container with ID starting with 9536fd1813dfd410759d213206f08c30e3bec894a291f9f250baae317d8f06ca not found: ID does not exist" Oct 02 01:50:37 crc kubenswrapper[4885]: I1002 01:50:37.518722 4885 scope.go:117] "RemoveContainer" containerID="7e5d8fe53556ae09ff185262c2de844873acae065f72e9740d554d72ff3968e1" Oct 02 01:50:37 crc kubenswrapper[4885]: E1002 01:50:37.518882 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7e5d8fe53556ae09ff185262c2de844873acae065f72e9740d554d72ff3968e1\": container with ID starting with 7e5d8fe53556ae09ff185262c2de844873acae065f72e9740d554d72ff3968e1 not found: ID does not exist" containerID="7e5d8fe53556ae09ff185262c2de844873acae065f72e9740d554d72ff3968e1" Oct 02 01:50:37 crc kubenswrapper[4885]: I1002 01:50:37.518896 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7e5d8fe53556ae09ff185262c2de844873acae065f72e9740d554d72ff3968e1"} err="failed to get container status \"7e5d8fe53556ae09ff185262c2de844873acae065f72e9740d554d72ff3968e1\": rpc error: code = NotFound desc = could not find container \"7e5d8fe53556ae09ff185262c2de844873acae065f72e9740d554d72ff3968e1\": container with ID starting with 7e5d8fe53556ae09ff185262c2de844873acae065f72e9740d554d72ff3968e1 not found: ID does not exist" Oct 02 01:50:37 crc kubenswrapper[4885]: I1002 01:50:37.521338 4885 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/21d974e0-8d62-4daf-b383-5a6823df7b8e-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 01:50:37 crc kubenswrapper[4885]: I1002 01:50:37.521382 4885 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/21d974e0-8d62-4daf-b383-5a6823df7b8e-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 01:50:37 crc kubenswrapper[4885]: I1002 01:50:37.521403 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-24dnt\" (UniqueName: \"kubernetes.io/projected/21d974e0-8d62-4daf-b383-5a6823df7b8e-kube-api-access-24dnt\") on node \"crc\" DevicePath \"\"" Oct 02 01:50:37 crc kubenswrapper[4885]: I1002 01:50:37.816784 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xq6kh"] Oct 02 01:50:37 crc kubenswrapper[4885]: I1002 01:50:37.822727 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-xq6kh"] Oct 02 01:50:38 crc kubenswrapper[4885]: I1002 01:50:38.057664 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="21d974e0-8d62-4daf-b383-5a6823df7b8e" path="/var/lib/kubelet/pods/21d974e0-8d62-4daf-b383-5a6823df7b8e/volumes" Oct 02 01:50:43 crc kubenswrapper[4885]: I1002 01:50:43.266041 4885 patch_prober.go:28] interesting pod/machine-config-daemon-rttx8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 01:50:43 crc kubenswrapper[4885]: I1002 01:50:43.266581 4885 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 01:50:43 crc kubenswrapper[4885]: I1002 01:50:43.266650 4885 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" Oct 02 01:50:43 crc kubenswrapper[4885]: I1002 01:50:43.267836 4885 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"09536e182a46e77318c10812d1777e551f834907c006cab5c726dc66f5bce88a"} pod="openshift-machine-config-operator/machine-config-daemon-rttx8" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 01:50:43 crc kubenswrapper[4885]: I1002 01:50:43.268023 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" containerName="machine-config-daemon" containerID="cri-o://09536e182a46e77318c10812d1777e551f834907c006cab5c726dc66f5bce88a" gracePeriod=600 Oct 02 01:50:43 crc kubenswrapper[4885]: I1002 01:50:43.517805 4885 generic.go:334] "Generic (PLEG): container finished" podID="d169c8ba-000a-4218-8e3b-4ae53035b110" containerID="09536e182a46e77318c10812d1777e551f834907c006cab5c726dc66f5bce88a" exitCode=0 Oct 02 01:50:43 crc kubenswrapper[4885]: I1002 01:50:43.517914 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" event={"ID":"d169c8ba-000a-4218-8e3b-4ae53035b110","Type":"ContainerDied","Data":"09536e182a46e77318c10812d1777e551f834907c006cab5c726dc66f5bce88a"} Oct 02 01:50:44 crc kubenswrapper[4885]: I1002 01:50:44.545563 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" event={"ID":"d169c8ba-000a-4218-8e3b-4ae53035b110","Type":"ContainerStarted","Data":"edcfaf9e26fcac2c575da14a9bd0f19b12460248f9779d8e37fc41a2db9b606d"} Oct 02 01:50:58 crc kubenswrapper[4885]: I1002 01:50:58.522574 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-mc59z" podUID="b6e3c195-a333-4c4f-8b75-71edbfcc26c3" containerName="oauth-openshift" containerID="cri-o://4199acd0bb43e86b61474db1810b27dd89d8d2bd044322f903c037b271703dc7" gracePeriod=15 Oct 02 01:50:58 crc kubenswrapper[4885]: I1002 01:50:58.963947 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-mc59z" Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.016153 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-5f7dff4c75-6ggrd"] Oct 02 01:50:59 crc kubenswrapper[4885]: E1002 01:50:59.016597 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d3cc54c-759c-43a7-b88c-4c5079b42e45" containerName="registry-server" Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.016621 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d3cc54c-759c-43a7-b88c-4c5079b42e45" containerName="registry-server" Oct 02 01:50:59 crc kubenswrapper[4885]: E1002 01:50:59.016637 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64a7eacf-552b-4a87-91db-4f6f72ee26b1" containerName="extract-content" Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.016650 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="64a7eacf-552b-4a87-91db-4f6f72ee26b1" containerName="extract-content" Oct 02 01:50:59 crc kubenswrapper[4885]: E1002 01:50:59.016670 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21d974e0-8d62-4daf-b383-5a6823df7b8e" containerName="registry-server" Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.016683 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="21d974e0-8d62-4daf-b383-5a6823df7b8e" containerName="registry-server" Oct 02 01:50:59 crc kubenswrapper[4885]: E1002 01:50:59.016699 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64a7eacf-552b-4a87-91db-4f6f72ee26b1" containerName="extract-utilities" Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.016711 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="64a7eacf-552b-4a87-91db-4f6f72ee26b1" containerName="extract-utilities" Oct 02 01:50:59 crc kubenswrapper[4885]: E1002 01:50:59.016733 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d3cc54c-759c-43a7-b88c-4c5079b42e45" containerName="extract-content" Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.016745 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d3cc54c-759c-43a7-b88c-4c5079b42e45" containerName="extract-content" Oct 02 01:50:59 crc kubenswrapper[4885]: E1002 01:50:59.016763 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64a7eacf-552b-4a87-91db-4f6f72ee26b1" containerName="registry-server" Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.016775 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="64a7eacf-552b-4a87-91db-4f6f72ee26b1" containerName="registry-server" Oct 02 01:50:59 crc kubenswrapper[4885]: E1002 01:50:59.016793 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d3cc54c-759c-43a7-b88c-4c5079b42e45" containerName="extract-utilities" Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.017072 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d3cc54c-759c-43a7-b88c-4c5079b42e45" containerName="extract-utilities" Oct 02 01:50:59 crc kubenswrapper[4885]: E1002 01:50:59.017107 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="adaf0543-043d-48b8-829f-928df3bf6e90" containerName="registry-server" Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.017125 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="adaf0543-043d-48b8-829f-928df3bf6e90" containerName="registry-server" Oct 02 01:50:59 crc kubenswrapper[4885]: E1002 01:50:59.017146 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="adaf0543-043d-48b8-829f-928df3bf6e90" containerName="extract-utilities" Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.017163 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="adaf0543-043d-48b8-829f-928df3bf6e90" containerName="extract-utilities" Oct 02 01:50:59 crc kubenswrapper[4885]: E1002 01:50:59.017188 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6e3c195-a333-4c4f-8b75-71edbfcc26c3" containerName="oauth-openshift" Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.017204 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6e3c195-a333-4c4f-8b75-71edbfcc26c3" containerName="oauth-openshift" Oct 02 01:50:59 crc kubenswrapper[4885]: E1002 01:50:59.017225 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="adaf0543-043d-48b8-829f-928df3bf6e90" containerName="extract-content" Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.017240 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="adaf0543-043d-48b8-829f-928df3bf6e90" containerName="extract-content" Oct 02 01:50:59 crc kubenswrapper[4885]: E1002 01:50:59.017301 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21d974e0-8d62-4daf-b383-5a6823df7b8e" containerName="extract-utilities" Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.017321 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="21d974e0-8d62-4daf-b383-5a6823df7b8e" containerName="extract-utilities" Oct 02 01:50:59 crc kubenswrapper[4885]: E1002 01:50:59.017346 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11b7ddd4-611b-41ee-9e12-b436939be852" containerName="pruner" Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.017362 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="11b7ddd4-611b-41ee-9e12-b436939be852" containerName="pruner" Oct 02 01:50:59 crc kubenswrapper[4885]: E1002 01:50:59.017390 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21d974e0-8d62-4daf-b383-5a6823df7b8e" containerName="extract-content" Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.017407 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="21d974e0-8d62-4daf-b383-5a6823df7b8e" containerName="extract-content" Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.017774 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d3cc54c-759c-43a7-b88c-4c5079b42e45" containerName="registry-server" Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.017810 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6e3c195-a333-4c4f-8b75-71edbfcc26c3" containerName="oauth-openshift" Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.017829 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="11b7ddd4-611b-41ee-9e12-b436939be852" containerName="pruner" Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.017855 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="64a7eacf-552b-4a87-91db-4f6f72ee26b1" containerName="registry-server" Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.017885 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="21d974e0-8d62-4daf-b383-5a6823df7b8e" containerName="registry-server" Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.017905 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="adaf0543-043d-48b8-829f-928df3bf6e90" containerName="registry-server" Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.020071 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-5f7dff4c75-6ggrd" Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.028171 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-5f7dff4c75-6ggrd"] Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.145963 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/b6e3c195-a333-4c4f-8b75-71edbfcc26c3-v4-0-config-system-ocp-branding-template\") pod \"b6e3c195-a333-4c4f-8b75-71edbfcc26c3\" (UID: \"b6e3c195-a333-4c4f-8b75-71edbfcc26c3\") " Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.146069 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/b6e3c195-a333-4c4f-8b75-71edbfcc26c3-v4-0-config-user-template-login\") pod \"b6e3c195-a333-4c4f-8b75-71edbfcc26c3\" (UID: \"b6e3c195-a333-4c4f-8b75-71edbfcc26c3\") " Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.146142 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b6e3c195-a333-4c4f-8b75-71edbfcc26c3-audit-dir\") pod \"b6e3c195-a333-4c4f-8b75-71edbfcc26c3\" (UID: \"b6e3c195-a333-4c4f-8b75-71edbfcc26c3\") " Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.146180 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/b6e3c195-a333-4c4f-8b75-71edbfcc26c3-v4-0-config-user-idp-0-file-data\") pod \"b6e3c195-a333-4c4f-8b75-71edbfcc26c3\" (UID: \"b6e3c195-a333-4c4f-8b75-71edbfcc26c3\") " Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.146286 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/b6e3c195-a333-4c4f-8b75-71edbfcc26c3-v4-0-config-system-service-ca\") pod \"b6e3c195-a333-4c4f-8b75-71edbfcc26c3\" (UID: \"b6e3c195-a333-4c4f-8b75-71edbfcc26c3\") " Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.146330 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b6e3c195-a333-4c4f-8b75-71edbfcc26c3-v4-0-config-system-trusted-ca-bundle\") pod \"b6e3c195-a333-4c4f-8b75-71edbfcc26c3\" (UID: \"b6e3c195-a333-4c4f-8b75-71edbfcc26c3\") " Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.146363 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l5s4w\" (UniqueName: \"kubernetes.io/projected/b6e3c195-a333-4c4f-8b75-71edbfcc26c3-kube-api-access-l5s4w\") pod \"b6e3c195-a333-4c4f-8b75-71edbfcc26c3\" (UID: \"b6e3c195-a333-4c4f-8b75-71edbfcc26c3\") " Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.146406 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/b6e3c195-a333-4c4f-8b75-71edbfcc26c3-v4-0-config-system-cliconfig\") pod \"b6e3c195-a333-4c4f-8b75-71edbfcc26c3\" (UID: \"b6e3c195-a333-4c4f-8b75-71edbfcc26c3\") " Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.146452 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/b6e3c195-a333-4c4f-8b75-71edbfcc26c3-v4-0-config-system-router-certs\") pod \"b6e3c195-a333-4c4f-8b75-71edbfcc26c3\" (UID: \"b6e3c195-a333-4c4f-8b75-71edbfcc26c3\") " Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.146489 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/b6e3c195-a333-4c4f-8b75-71edbfcc26c3-v4-0-config-user-template-error\") pod \"b6e3c195-a333-4c4f-8b75-71edbfcc26c3\" (UID: \"b6e3c195-a333-4c4f-8b75-71edbfcc26c3\") " Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.146532 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/b6e3c195-a333-4c4f-8b75-71edbfcc26c3-v4-0-config-system-session\") pod \"b6e3c195-a333-4c4f-8b75-71edbfcc26c3\" (UID: \"b6e3c195-a333-4c4f-8b75-71edbfcc26c3\") " Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.146565 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/b6e3c195-a333-4c4f-8b75-71edbfcc26c3-audit-policies\") pod \"b6e3c195-a333-4c4f-8b75-71edbfcc26c3\" (UID: \"b6e3c195-a333-4c4f-8b75-71edbfcc26c3\") " Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.146609 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/b6e3c195-a333-4c4f-8b75-71edbfcc26c3-v4-0-config-user-template-provider-selection\") pod \"b6e3c195-a333-4c4f-8b75-71edbfcc26c3\" (UID: \"b6e3c195-a333-4c4f-8b75-71edbfcc26c3\") " Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.146640 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/b6e3c195-a333-4c4f-8b75-71edbfcc26c3-v4-0-config-system-serving-cert\") pod \"b6e3c195-a333-4c4f-8b75-71edbfcc26c3\" (UID: \"b6e3c195-a333-4c4f-8b75-71edbfcc26c3\") " Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.146831 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/fe4ff5e6-3d46-406b-aa78-4a3e008ee434-v4-0-config-system-session\") pod \"oauth-openshift-5f7dff4c75-6ggrd\" (UID: \"fe4ff5e6-3d46-406b-aa78-4a3e008ee434\") " pod="openshift-authentication/oauth-openshift-5f7dff4c75-6ggrd" Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.146871 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/fe4ff5e6-3d46-406b-aa78-4a3e008ee434-audit-dir\") pod \"oauth-openshift-5f7dff4c75-6ggrd\" (UID: \"fe4ff5e6-3d46-406b-aa78-4a3e008ee434\") " pod="openshift-authentication/oauth-openshift-5f7dff4c75-6ggrd" Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.146902 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/fe4ff5e6-3d46-406b-aa78-4a3e008ee434-v4-0-config-system-cliconfig\") pod \"oauth-openshift-5f7dff4c75-6ggrd\" (UID: \"fe4ff5e6-3d46-406b-aa78-4a3e008ee434\") " pod="openshift-authentication/oauth-openshift-5f7dff4c75-6ggrd" Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.146943 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/fe4ff5e6-3d46-406b-aa78-4a3e008ee434-v4-0-config-user-template-error\") pod \"oauth-openshift-5f7dff4c75-6ggrd\" (UID: \"fe4ff5e6-3d46-406b-aa78-4a3e008ee434\") " pod="openshift-authentication/oauth-openshift-5f7dff4c75-6ggrd" Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.147002 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lflvn\" (UniqueName: \"kubernetes.io/projected/fe4ff5e6-3d46-406b-aa78-4a3e008ee434-kube-api-access-lflvn\") pod \"oauth-openshift-5f7dff4c75-6ggrd\" (UID: \"fe4ff5e6-3d46-406b-aa78-4a3e008ee434\") " pod="openshift-authentication/oauth-openshift-5f7dff4c75-6ggrd" Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.147052 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/fe4ff5e6-3d46-406b-aa78-4a3e008ee434-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-5f7dff4c75-6ggrd\" (UID: \"fe4ff5e6-3d46-406b-aa78-4a3e008ee434\") " pod="openshift-authentication/oauth-openshift-5f7dff4c75-6ggrd" Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.147090 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/fe4ff5e6-3d46-406b-aa78-4a3e008ee434-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-5f7dff4c75-6ggrd\" (UID: \"fe4ff5e6-3d46-406b-aa78-4a3e008ee434\") " pod="openshift-authentication/oauth-openshift-5f7dff4c75-6ggrd" Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.147128 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/fe4ff5e6-3d46-406b-aa78-4a3e008ee434-audit-policies\") pod \"oauth-openshift-5f7dff4c75-6ggrd\" (UID: \"fe4ff5e6-3d46-406b-aa78-4a3e008ee434\") " pod="openshift-authentication/oauth-openshift-5f7dff4c75-6ggrd" Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.147159 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/fe4ff5e6-3d46-406b-aa78-4a3e008ee434-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-5f7dff4c75-6ggrd\" (UID: \"fe4ff5e6-3d46-406b-aa78-4a3e008ee434\") " pod="openshift-authentication/oauth-openshift-5f7dff4c75-6ggrd" Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.147193 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/fe4ff5e6-3d46-406b-aa78-4a3e008ee434-v4-0-config-system-router-certs\") pod \"oauth-openshift-5f7dff4c75-6ggrd\" (UID: \"fe4ff5e6-3d46-406b-aa78-4a3e008ee434\") " pod="openshift-authentication/oauth-openshift-5f7dff4c75-6ggrd" Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.147225 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/fe4ff5e6-3d46-406b-aa78-4a3e008ee434-v4-0-config-system-serving-cert\") pod \"oauth-openshift-5f7dff4c75-6ggrd\" (UID: \"fe4ff5e6-3d46-406b-aa78-4a3e008ee434\") " pod="openshift-authentication/oauth-openshift-5f7dff4c75-6ggrd" Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.147257 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/fe4ff5e6-3d46-406b-aa78-4a3e008ee434-v4-0-config-system-service-ca\") pod \"oauth-openshift-5f7dff4c75-6ggrd\" (UID: \"fe4ff5e6-3d46-406b-aa78-4a3e008ee434\") " pod="openshift-authentication/oauth-openshift-5f7dff4c75-6ggrd" Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.147342 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fe4ff5e6-3d46-406b-aa78-4a3e008ee434-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-5f7dff4c75-6ggrd\" (UID: \"fe4ff5e6-3d46-406b-aa78-4a3e008ee434\") " pod="openshift-authentication/oauth-openshift-5f7dff4c75-6ggrd" Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.147379 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/fe4ff5e6-3d46-406b-aa78-4a3e008ee434-v4-0-config-user-template-login\") pod \"oauth-openshift-5f7dff4c75-6ggrd\" (UID: \"fe4ff5e6-3d46-406b-aa78-4a3e008ee434\") " pod="openshift-authentication/oauth-openshift-5f7dff4c75-6ggrd" Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.147428 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6e3c195-a333-4c4f-8b75-71edbfcc26c3-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "b6e3c195-a333-4c4f-8b75-71edbfcc26c3" (UID: "b6e3c195-a333-4c4f-8b75-71edbfcc26c3"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.147514 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b6e3c195-a333-4c4f-8b75-71edbfcc26c3-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "b6e3c195-a333-4c4f-8b75-71edbfcc26c3" (UID: "b6e3c195-a333-4c4f-8b75-71edbfcc26c3"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.148614 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6e3c195-a333-4c4f-8b75-71edbfcc26c3-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "b6e3c195-a333-4c4f-8b75-71edbfcc26c3" (UID: "b6e3c195-a333-4c4f-8b75-71edbfcc26c3"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.149137 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6e3c195-a333-4c4f-8b75-71edbfcc26c3-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "b6e3c195-a333-4c4f-8b75-71edbfcc26c3" (UID: "b6e3c195-a333-4c4f-8b75-71edbfcc26c3"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.149251 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6e3c195-a333-4c4f-8b75-71edbfcc26c3-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "b6e3c195-a333-4c4f-8b75-71edbfcc26c3" (UID: "b6e3c195-a333-4c4f-8b75-71edbfcc26c3"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.166613 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6e3c195-a333-4c4f-8b75-71edbfcc26c3-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "b6e3c195-a333-4c4f-8b75-71edbfcc26c3" (UID: "b6e3c195-a333-4c4f-8b75-71edbfcc26c3"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.166812 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6e3c195-a333-4c4f-8b75-71edbfcc26c3-kube-api-access-l5s4w" (OuterVolumeSpecName: "kube-api-access-l5s4w") pod "b6e3c195-a333-4c4f-8b75-71edbfcc26c3" (UID: "b6e3c195-a333-4c4f-8b75-71edbfcc26c3"). InnerVolumeSpecName "kube-api-access-l5s4w". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.167461 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6e3c195-a333-4c4f-8b75-71edbfcc26c3-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "b6e3c195-a333-4c4f-8b75-71edbfcc26c3" (UID: "b6e3c195-a333-4c4f-8b75-71edbfcc26c3"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.168574 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6e3c195-a333-4c4f-8b75-71edbfcc26c3-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "b6e3c195-a333-4c4f-8b75-71edbfcc26c3" (UID: "b6e3c195-a333-4c4f-8b75-71edbfcc26c3"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.168941 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6e3c195-a333-4c4f-8b75-71edbfcc26c3-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "b6e3c195-a333-4c4f-8b75-71edbfcc26c3" (UID: "b6e3c195-a333-4c4f-8b75-71edbfcc26c3"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.179404 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6e3c195-a333-4c4f-8b75-71edbfcc26c3-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "b6e3c195-a333-4c4f-8b75-71edbfcc26c3" (UID: "b6e3c195-a333-4c4f-8b75-71edbfcc26c3"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.180902 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6e3c195-a333-4c4f-8b75-71edbfcc26c3-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "b6e3c195-a333-4c4f-8b75-71edbfcc26c3" (UID: "b6e3c195-a333-4c4f-8b75-71edbfcc26c3"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.181066 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6e3c195-a333-4c4f-8b75-71edbfcc26c3-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "b6e3c195-a333-4c4f-8b75-71edbfcc26c3" (UID: "b6e3c195-a333-4c4f-8b75-71edbfcc26c3"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.181571 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6e3c195-a333-4c4f-8b75-71edbfcc26c3-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "b6e3c195-a333-4c4f-8b75-71edbfcc26c3" (UID: "b6e3c195-a333-4c4f-8b75-71edbfcc26c3"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.248908 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lflvn\" (UniqueName: \"kubernetes.io/projected/fe4ff5e6-3d46-406b-aa78-4a3e008ee434-kube-api-access-lflvn\") pod \"oauth-openshift-5f7dff4c75-6ggrd\" (UID: \"fe4ff5e6-3d46-406b-aa78-4a3e008ee434\") " pod="openshift-authentication/oauth-openshift-5f7dff4c75-6ggrd" Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.249006 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/fe4ff5e6-3d46-406b-aa78-4a3e008ee434-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-5f7dff4c75-6ggrd\" (UID: \"fe4ff5e6-3d46-406b-aa78-4a3e008ee434\") " pod="openshift-authentication/oauth-openshift-5f7dff4c75-6ggrd" Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.249051 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/fe4ff5e6-3d46-406b-aa78-4a3e008ee434-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-5f7dff4c75-6ggrd\" (UID: \"fe4ff5e6-3d46-406b-aa78-4a3e008ee434\") " pod="openshift-authentication/oauth-openshift-5f7dff4c75-6ggrd" Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.249096 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/fe4ff5e6-3d46-406b-aa78-4a3e008ee434-audit-policies\") pod \"oauth-openshift-5f7dff4c75-6ggrd\" (UID: \"fe4ff5e6-3d46-406b-aa78-4a3e008ee434\") " pod="openshift-authentication/oauth-openshift-5f7dff4c75-6ggrd" Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.249132 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/fe4ff5e6-3d46-406b-aa78-4a3e008ee434-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-5f7dff4c75-6ggrd\" (UID: \"fe4ff5e6-3d46-406b-aa78-4a3e008ee434\") " pod="openshift-authentication/oauth-openshift-5f7dff4c75-6ggrd" Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.249168 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/fe4ff5e6-3d46-406b-aa78-4a3e008ee434-v4-0-config-system-router-certs\") pod \"oauth-openshift-5f7dff4c75-6ggrd\" (UID: \"fe4ff5e6-3d46-406b-aa78-4a3e008ee434\") " pod="openshift-authentication/oauth-openshift-5f7dff4c75-6ggrd" Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.249201 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/fe4ff5e6-3d46-406b-aa78-4a3e008ee434-v4-0-config-system-serving-cert\") pod \"oauth-openshift-5f7dff4c75-6ggrd\" (UID: \"fe4ff5e6-3d46-406b-aa78-4a3e008ee434\") " pod="openshift-authentication/oauth-openshift-5f7dff4c75-6ggrd" Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.249231 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/fe4ff5e6-3d46-406b-aa78-4a3e008ee434-v4-0-config-system-service-ca\") pod \"oauth-openshift-5f7dff4c75-6ggrd\" (UID: \"fe4ff5e6-3d46-406b-aa78-4a3e008ee434\") " pod="openshift-authentication/oauth-openshift-5f7dff4c75-6ggrd" Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.249354 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fe4ff5e6-3d46-406b-aa78-4a3e008ee434-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-5f7dff4c75-6ggrd\" (UID: \"fe4ff5e6-3d46-406b-aa78-4a3e008ee434\") " pod="openshift-authentication/oauth-openshift-5f7dff4c75-6ggrd" Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.249409 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/fe4ff5e6-3d46-406b-aa78-4a3e008ee434-v4-0-config-user-template-login\") pod \"oauth-openshift-5f7dff4c75-6ggrd\" (UID: \"fe4ff5e6-3d46-406b-aa78-4a3e008ee434\") " pod="openshift-authentication/oauth-openshift-5f7dff4c75-6ggrd" Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.249483 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/fe4ff5e6-3d46-406b-aa78-4a3e008ee434-v4-0-config-system-session\") pod \"oauth-openshift-5f7dff4c75-6ggrd\" (UID: \"fe4ff5e6-3d46-406b-aa78-4a3e008ee434\") " pod="openshift-authentication/oauth-openshift-5f7dff4c75-6ggrd" Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.249533 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/fe4ff5e6-3d46-406b-aa78-4a3e008ee434-audit-dir\") pod \"oauth-openshift-5f7dff4c75-6ggrd\" (UID: \"fe4ff5e6-3d46-406b-aa78-4a3e008ee434\") " pod="openshift-authentication/oauth-openshift-5f7dff4c75-6ggrd" Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.249575 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/fe4ff5e6-3d46-406b-aa78-4a3e008ee434-v4-0-config-system-cliconfig\") pod \"oauth-openshift-5f7dff4c75-6ggrd\" (UID: \"fe4ff5e6-3d46-406b-aa78-4a3e008ee434\") " pod="openshift-authentication/oauth-openshift-5f7dff4c75-6ggrd" Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.249620 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/fe4ff5e6-3d46-406b-aa78-4a3e008ee434-v4-0-config-user-template-error\") pod \"oauth-openshift-5f7dff4c75-6ggrd\" (UID: \"fe4ff5e6-3d46-406b-aa78-4a3e008ee434\") " pod="openshift-authentication/oauth-openshift-5f7dff4c75-6ggrd" Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.249690 4885 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/b6e3c195-a333-4c4f-8b75-71edbfcc26c3-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.249720 4885 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/b6e3c195-a333-4c4f-8b75-71edbfcc26c3-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.249751 4885 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/b6e3c195-a333-4c4f-8b75-71edbfcc26c3-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.249780 4885 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/b6e3c195-a333-4c4f-8b75-71edbfcc26c3-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.249807 4885 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/b6e3c195-a333-4c4f-8b75-71edbfcc26c3-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.249832 4885 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/b6e3c195-a333-4c4f-8b75-71edbfcc26c3-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.249862 4885 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b6e3c195-a333-4c4f-8b75-71edbfcc26c3-audit-dir\") on node \"crc\" DevicePath \"\"" Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.249886 4885 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/b6e3c195-a333-4c4f-8b75-71edbfcc26c3-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.249912 4885 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/b6e3c195-a333-4c4f-8b75-71edbfcc26c3-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.249942 4885 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b6e3c195-a333-4c4f-8b75-71edbfcc26c3-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.249967 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l5s4w\" (UniqueName: \"kubernetes.io/projected/b6e3c195-a333-4c4f-8b75-71edbfcc26c3-kube-api-access-l5s4w\") on node \"crc\" DevicePath \"\"" Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.249993 4885 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/b6e3c195-a333-4c4f-8b75-71edbfcc26c3-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.250019 4885 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/b6e3c195-a333-4c4f-8b75-71edbfcc26c3-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.250044 4885 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/b6e3c195-a333-4c4f-8b75-71edbfcc26c3-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.250665 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/fe4ff5e6-3d46-406b-aa78-4a3e008ee434-audit-dir\") pod \"oauth-openshift-5f7dff4c75-6ggrd\" (UID: \"fe4ff5e6-3d46-406b-aa78-4a3e008ee434\") " pod="openshift-authentication/oauth-openshift-5f7dff4c75-6ggrd" Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.251516 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fe4ff5e6-3d46-406b-aa78-4a3e008ee434-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-5f7dff4c75-6ggrd\" (UID: \"fe4ff5e6-3d46-406b-aa78-4a3e008ee434\") " pod="openshift-authentication/oauth-openshift-5f7dff4c75-6ggrd" Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.251898 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/fe4ff5e6-3d46-406b-aa78-4a3e008ee434-v4-0-config-system-service-ca\") pod \"oauth-openshift-5f7dff4c75-6ggrd\" (UID: \"fe4ff5e6-3d46-406b-aa78-4a3e008ee434\") " pod="openshift-authentication/oauth-openshift-5f7dff4c75-6ggrd" Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.252549 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/fe4ff5e6-3d46-406b-aa78-4a3e008ee434-audit-policies\") pod \"oauth-openshift-5f7dff4c75-6ggrd\" (UID: \"fe4ff5e6-3d46-406b-aa78-4a3e008ee434\") " pod="openshift-authentication/oauth-openshift-5f7dff4c75-6ggrd" Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.252899 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/fe4ff5e6-3d46-406b-aa78-4a3e008ee434-v4-0-config-system-cliconfig\") pod \"oauth-openshift-5f7dff4c75-6ggrd\" (UID: \"fe4ff5e6-3d46-406b-aa78-4a3e008ee434\") " pod="openshift-authentication/oauth-openshift-5f7dff4c75-6ggrd" Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.255473 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/fe4ff5e6-3d46-406b-aa78-4a3e008ee434-v4-0-config-system-router-certs\") pod \"oauth-openshift-5f7dff4c75-6ggrd\" (UID: \"fe4ff5e6-3d46-406b-aa78-4a3e008ee434\") " pod="openshift-authentication/oauth-openshift-5f7dff4c75-6ggrd" Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.255638 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/fe4ff5e6-3d46-406b-aa78-4a3e008ee434-v4-0-config-user-template-error\") pod \"oauth-openshift-5f7dff4c75-6ggrd\" (UID: \"fe4ff5e6-3d46-406b-aa78-4a3e008ee434\") " pod="openshift-authentication/oauth-openshift-5f7dff4c75-6ggrd" Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.255831 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/fe4ff5e6-3d46-406b-aa78-4a3e008ee434-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-5f7dff4c75-6ggrd\" (UID: \"fe4ff5e6-3d46-406b-aa78-4a3e008ee434\") " pod="openshift-authentication/oauth-openshift-5f7dff4c75-6ggrd" Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.255892 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/fe4ff5e6-3d46-406b-aa78-4a3e008ee434-v4-0-config-system-session\") pod \"oauth-openshift-5f7dff4c75-6ggrd\" (UID: \"fe4ff5e6-3d46-406b-aa78-4a3e008ee434\") " pod="openshift-authentication/oauth-openshift-5f7dff4c75-6ggrd" Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.257067 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/fe4ff5e6-3d46-406b-aa78-4a3e008ee434-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-5f7dff4c75-6ggrd\" (UID: \"fe4ff5e6-3d46-406b-aa78-4a3e008ee434\") " pod="openshift-authentication/oauth-openshift-5f7dff4c75-6ggrd" Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.257677 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/fe4ff5e6-3d46-406b-aa78-4a3e008ee434-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-5f7dff4c75-6ggrd\" (UID: \"fe4ff5e6-3d46-406b-aa78-4a3e008ee434\") " pod="openshift-authentication/oauth-openshift-5f7dff4c75-6ggrd" Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.257784 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/fe4ff5e6-3d46-406b-aa78-4a3e008ee434-v4-0-config-user-template-login\") pod \"oauth-openshift-5f7dff4c75-6ggrd\" (UID: \"fe4ff5e6-3d46-406b-aa78-4a3e008ee434\") " pod="openshift-authentication/oauth-openshift-5f7dff4c75-6ggrd" Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.258444 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/fe4ff5e6-3d46-406b-aa78-4a3e008ee434-v4-0-config-system-serving-cert\") pod \"oauth-openshift-5f7dff4c75-6ggrd\" (UID: \"fe4ff5e6-3d46-406b-aa78-4a3e008ee434\") " pod="openshift-authentication/oauth-openshift-5f7dff4c75-6ggrd" Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.278845 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lflvn\" (UniqueName: \"kubernetes.io/projected/fe4ff5e6-3d46-406b-aa78-4a3e008ee434-kube-api-access-lflvn\") pod \"oauth-openshift-5f7dff4c75-6ggrd\" (UID: \"fe4ff5e6-3d46-406b-aa78-4a3e008ee434\") " pod="openshift-authentication/oauth-openshift-5f7dff4c75-6ggrd" Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.351037 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-5f7dff4c75-6ggrd" Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.659714 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-5f7dff4c75-6ggrd"] Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.661918 4885 generic.go:334] "Generic (PLEG): container finished" podID="b6e3c195-a333-4c4f-8b75-71edbfcc26c3" containerID="4199acd0bb43e86b61474db1810b27dd89d8d2bd044322f903c037b271703dc7" exitCode=0 Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.661984 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-mc59z" event={"ID":"b6e3c195-a333-4c4f-8b75-71edbfcc26c3","Type":"ContainerDied","Data":"4199acd0bb43e86b61474db1810b27dd89d8d2bd044322f903c037b271703dc7"} Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.662039 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-mc59z" event={"ID":"b6e3c195-a333-4c4f-8b75-71edbfcc26c3","Type":"ContainerDied","Data":"240fb0c7b8e78aab6f935fa72be12ed1cd0c0d37c3359c5ce3772b71db33943d"} Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.662076 4885 scope.go:117] "RemoveContainer" containerID="4199acd0bb43e86b61474db1810b27dd89d8d2bd044322f903c037b271703dc7" Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.662310 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-mc59z" Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.725167 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-mc59z"] Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.728479 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-mc59z"] Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.733176 4885 scope.go:117] "RemoveContainer" containerID="4199acd0bb43e86b61474db1810b27dd89d8d2bd044322f903c037b271703dc7" Oct 02 01:50:59 crc kubenswrapper[4885]: E1002 01:50:59.735097 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4199acd0bb43e86b61474db1810b27dd89d8d2bd044322f903c037b271703dc7\": container with ID starting with 4199acd0bb43e86b61474db1810b27dd89d8d2bd044322f903c037b271703dc7 not found: ID does not exist" containerID="4199acd0bb43e86b61474db1810b27dd89d8d2bd044322f903c037b271703dc7" Oct 02 01:50:59 crc kubenswrapper[4885]: I1002 01:50:59.735153 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4199acd0bb43e86b61474db1810b27dd89d8d2bd044322f903c037b271703dc7"} err="failed to get container status \"4199acd0bb43e86b61474db1810b27dd89d8d2bd044322f903c037b271703dc7\": rpc error: code = NotFound desc = could not find container \"4199acd0bb43e86b61474db1810b27dd89d8d2bd044322f903c037b271703dc7\": container with ID starting with 4199acd0bb43e86b61474db1810b27dd89d8d2bd044322f903c037b271703dc7 not found: ID does not exist" Oct 02 01:51:00 crc kubenswrapper[4885]: I1002 01:51:00.054943 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6e3c195-a333-4c4f-8b75-71edbfcc26c3" path="/var/lib/kubelet/pods/b6e3c195-a333-4c4f-8b75-71edbfcc26c3/volumes" Oct 02 01:51:00 crc kubenswrapper[4885]: I1002 01:51:00.673965 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-5f7dff4c75-6ggrd" event={"ID":"fe4ff5e6-3d46-406b-aa78-4a3e008ee434","Type":"ContainerStarted","Data":"eddb8104418782c82b7d77a3cdc97190215721593eb8b2d0b9445edc639b1899"} Oct 02 01:51:00 crc kubenswrapper[4885]: I1002 01:51:00.674573 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-5f7dff4c75-6ggrd" Oct 02 01:51:00 crc kubenswrapper[4885]: I1002 01:51:00.674603 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-5f7dff4c75-6ggrd" event={"ID":"fe4ff5e6-3d46-406b-aa78-4a3e008ee434","Type":"ContainerStarted","Data":"1868fe003f4c67947a3c0ba9416f4d828015c992e6baaf4a586fd5c3f7130f4f"} Oct 02 01:51:00 crc kubenswrapper[4885]: I1002 01:51:00.710996 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-5f7dff4c75-6ggrd" Oct 02 01:51:00 crc kubenswrapper[4885]: I1002 01:51:00.714689 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-5f7dff4c75-6ggrd" podStartSLOduration=27.714662013999998 podStartE2EDuration="27.714662014s" podCreationTimestamp="2025-10-02 01:50:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:51:00.706955623 +0000 UTC m=+249.518703122" watchObservedRunningTime="2025-10-02 01:51:00.714662014 +0000 UTC m=+249.526409453" Oct 02 01:51:20 crc kubenswrapper[4885]: I1002 01:51:20.621524 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vb2j4"] Oct 02 01:51:20 crc kubenswrapper[4885]: I1002 01:51:20.622389 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-vb2j4" podUID="a28f881c-dd5e-4bf8-82ad-2c27b98dcceb" containerName="registry-server" containerID="cri-o://fe571f9130c9ad468ffd6ca21eef2aa385a83079b41b04f8e414c3e5450e39aa" gracePeriod=30 Oct 02 01:51:20 crc kubenswrapper[4885]: I1002 01:51:20.638958 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-x44t6"] Oct 02 01:51:20 crc kubenswrapper[4885]: I1002 01:51:20.639889 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-x44t6" podUID="961bc3c9-b6be-4ece-98f1-971a583f9db8" containerName="registry-server" containerID="cri-o://fa91784093ffc8297aea22d51f40aa7fe56e7c4e2ea262606bd94b010b505037" gracePeriod=30 Oct 02 01:51:20 crc kubenswrapper[4885]: I1002 01:51:20.670167 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-z4rlr"] Oct 02 01:51:20 crc kubenswrapper[4885]: I1002 01:51:20.670457 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-z4rlr" podUID="aa89bc07-6d38-40ba-8e90-9ecae726246d" containerName="marketplace-operator" containerID="cri-o://f04387030d901d22dd8ad8245b4b97815ac740667f7a7642348a4bfb750b45f7" gracePeriod=30 Oct 02 01:51:20 crc kubenswrapper[4885]: I1002 01:51:20.682462 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bz9hg"] Oct 02 01:51:20 crc kubenswrapper[4885]: I1002 01:51:20.682670 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-bz9hg" podUID="aecca2a3-40b0-4666-a86d-d67ecb580e99" containerName="registry-server" containerID="cri-o://00335c982112fd4f0ea73d0823b195c015c56ae6305bfcd2e105be3c2c335313" gracePeriod=30 Oct 02 01:51:20 crc kubenswrapper[4885]: I1002 01:51:20.684918 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bjxpx"] Oct 02 01:51:20 crc kubenswrapper[4885]: I1002 01:51:20.685151 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-bjxpx" podUID="18a2b456-262c-43fd-97c6-bfaf2d658463" containerName="registry-server" containerID="cri-o://9248711d02d76c11268438ac426cc0097b451fcdadac403d61ebe565076fff04" gracePeriod=30 Oct 02 01:51:20 crc kubenswrapper[4885]: I1002 01:51:20.687209 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-z7fcl"] Oct 02 01:51:20 crc kubenswrapper[4885]: I1002 01:51:20.687924 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-z7fcl" Oct 02 01:51:20 crc kubenswrapper[4885]: I1002 01:51:20.691498 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-z7fcl"] Oct 02 01:51:20 crc kubenswrapper[4885]: I1002 01:51:20.789693 4885 generic.go:334] "Generic (PLEG): container finished" podID="961bc3c9-b6be-4ece-98f1-971a583f9db8" containerID="fa91784093ffc8297aea22d51f40aa7fe56e7c4e2ea262606bd94b010b505037" exitCode=0 Oct 02 01:51:20 crc kubenswrapper[4885]: I1002 01:51:20.789774 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x44t6" event={"ID":"961bc3c9-b6be-4ece-98f1-971a583f9db8","Type":"ContainerDied","Data":"fa91784093ffc8297aea22d51f40aa7fe56e7c4e2ea262606bd94b010b505037"} Oct 02 01:51:20 crc kubenswrapper[4885]: I1002 01:51:20.795560 4885 generic.go:334] "Generic (PLEG): container finished" podID="a28f881c-dd5e-4bf8-82ad-2c27b98dcceb" containerID="fe571f9130c9ad468ffd6ca21eef2aa385a83079b41b04f8e414c3e5450e39aa" exitCode=0 Oct 02 01:51:20 crc kubenswrapper[4885]: I1002 01:51:20.795611 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vb2j4" event={"ID":"a28f881c-dd5e-4bf8-82ad-2c27b98dcceb","Type":"ContainerDied","Data":"fe571f9130c9ad468ffd6ca21eef2aa385a83079b41b04f8e414c3e5450e39aa"} Oct 02 01:51:20 crc kubenswrapper[4885]: I1002 01:51:20.798478 4885 generic.go:334] "Generic (PLEG): container finished" podID="aa89bc07-6d38-40ba-8e90-9ecae726246d" containerID="f04387030d901d22dd8ad8245b4b97815ac740667f7a7642348a4bfb750b45f7" exitCode=0 Oct 02 01:51:20 crc kubenswrapper[4885]: I1002 01:51:20.798532 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-z4rlr" event={"ID":"aa89bc07-6d38-40ba-8e90-9ecae726246d","Type":"ContainerDied","Data":"f04387030d901d22dd8ad8245b4b97815ac740667f7a7642348a4bfb750b45f7"} Oct 02 01:51:20 crc kubenswrapper[4885]: I1002 01:51:20.802383 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j79s5\" (UniqueName: \"kubernetes.io/projected/b0909ffb-180c-4e74-9916-98de47f2419d-kube-api-access-j79s5\") pod \"marketplace-operator-79b997595-z7fcl\" (UID: \"b0909ffb-180c-4e74-9916-98de47f2419d\") " pod="openshift-marketplace/marketplace-operator-79b997595-z7fcl" Oct 02 01:51:20 crc kubenswrapper[4885]: I1002 01:51:20.802419 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b0909ffb-180c-4e74-9916-98de47f2419d-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-z7fcl\" (UID: \"b0909ffb-180c-4e74-9916-98de47f2419d\") " pod="openshift-marketplace/marketplace-operator-79b997595-z7fcl" Oct 02 01:51:20 crc kubenswrapper[4885]: I1002 01:51:20.802449 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b0909ffb-180c-4e74-9916-98de47f2419d-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-z7fcl\" (UID: \"b0909ffb-180c-4e74-9916-98de47f2419d\") " pod="openshift-marketplace/marketplace-operator-79b997595-z7fcl" Oct 02 01:51:20 crc kubenswrapper[4885]: I1002 01:51:20.903994 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j79s5\" (UniqueName: \"kubernetes.io/projected/b0909ffb-180c-4e74-9916-98de47f2419d-kube-api-access-j79s5\") pod \"marketplace-operator-79b997595-z7fcl\" (UID: \"b0909ffb-180c-4e74-9916-98de47f2419d\") " pod="openshift-marketplace/marketplace-operator-79b997595-z7fcl" Oct 02 01:51:20 crc kubenswrapper[4885]: I1002 01:51:20.904035 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b0909ffb-180c-4e74-9916-98de47f2419d-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-z7fcl\" (UID: \"b0909ffb-180c-4e74-9916-98de47f2419d\") " pod="openshift-marketplace/marketplace-operator-79b997595-z7fcl" Oct 02 01:51:20 crc kubenswrapper[4885]: I1002 01:51:20.904071 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b0909ffb-180c-4e74-9916-98de47f2419d-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-z7fcl\" (UID: \"b0909ffb-180c-4e74-9916-98de47f2419d\") " pod="openshift-marketplace/marketplace-operator-79b997595-z7fcl" Oct 02 01:51:20 crc kubenswrapper[4885]: I1002 01:51:20.905339 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b0909ffb-180c-4e74-9916-98de47f2419d-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-z7fcl\" (UID: \"b0909ffb-180c-4e74-9916-98de47f2419d\") " pod="openshift-marketplace/marketplace-operator-79b997595-z7fcl" Oct 02 01:51:20 crc kubenswrapper[4885]: I1002 01:51:20.909200 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b0909ffb-180c-4e74-9916-98de47f2419d-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-z7fcl\" (UID: \"b0909ffb-180c-4e74-9916-98de47f2419d\") " pod="openshift-marketplace/marketplace-operator-79b997595-z7fcl" Oct 02 01:51:20 crc kubenswrapper[4885]: I1002 01:51:20.918349 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j79s5\" (UniqueName: \"kubernetes.io/projected/b0909ffb-180c-4e74-9916-98de47f2419d-kube-api-access-j79s5\") pod \"marketplace-operator-79b997595-z7fcl\" (UID: \"b0909ffb-180c-4e74-9916-98de47f2419d\") " pod="openshift-marketplace/marketplace-operator-79b997595-z7fcl" Oct 02 01:51:21 crc kubenswrapper[4885]: I1002 01:51:21.095983 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-z7fcl" Oct 02 01:51:21 crc kubenswrapper[4885]: I1002 01:51:21.099822 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vb2j4" Oct 02 01:51:21 crc kubenswrapper[4885]: I1002 01:51:21.104343 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-x44t6" Oct 02 01:51:21 crc kubenswrapper[4885]: I1002 01:51:21.108080 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-z4rlr" Oct 02 01:51:21 crc kubenswrapper[4885]: I1002 01:51:21.158425 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bjxpx" Oct 02 01:51:21 crc kubenswrapper[4885]: I1002 01:51:21.166999 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bz9hg" Oct 02 01:51:21 crc kubenswrapper[4885]: I1002 01:51:21.207523 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/961bc3c9-b6be-4ece-98f1-971a583f9db8-utilities\") pod \"961bc3c9-b6be-4ece-98f1-971a583f9db8\" (UID: \"961bc3c9-b6be-4ece-98f1-971a583f9db8\") " Oct 02 01:51:21 crc kubenswrapper[4885]: I1002 01:51:21.208188 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/961bc3c9-b6be-4ece-98f1-971a583f9db8-utilities" (OuterVolumeSpecName: "utilities") pod "961bc3c9-b6be-4ece-98f1-971a583f9db8" (UID: "961bc3c9-b6be-4ece-98f1-971a583f9db8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 01:51:21 crc kubenswrapper[4885]: I1002 01:51:21.208239 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nc9tk\" (UniqueName: \"kubernetes.io/projected/aa89bc07-6d38-40ba-8e90-9ecae726246d-kube-api-access-nc9tk\") pod \"aa89bc07-6d38-40ba-8e90-9ecae726246d\" (UID: \"aa89bc07-6d38-40ba-8e90-9ecae726246d\") " Oct 02 01:51:21 crc kubenswrapper[4885]: I1002 01:51:21.208988 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18a2b456-262c-43fd-97c6-bfaf2d658463-utilities\") pod \"18a2b456-262c-43fd-97c6-bfaf2d658463\" (UID: \"18a2b456-262c-43fd-97c6-bfaf2d658463\") " Oct 02 01:51:21 crc kubenswrapper[4885]: I1002 01:51:21.209022 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/961bc3c9-b6be-4ece-98f1-971a583f9db8-catalog-content\") pod \"961bc3c9-b6be-4ece-98f1-971a583f9db8\" (UID: \"961bc3c9-b6be-4ece-98f1-971a583f9db8\") " Oct 02 01:51:21 crc kubenswrapper[4885]: I1002 01:51:21.209053 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jqcx5\" (UniqueName: \"kubernetes.io/projected/aecca2a3-40b0-4666-a86d-d67ecb580e99-kube-api-access-jqcx5\") pod \"aecca2a3-40b0-4666-a86d-d67ecb580e99\" (UID: \"aecca2a3-40b0-4666-a86d-d67ecb580e99\") " Oct 02 01:51:21 crc kubenswrapper[4885]: I1002 01:51:21.209079 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-svczt\" (UniqueName: \"kubernetes.io/projected/961bc3c9-b6be-4ece-98f1-971a583f9db8-kube-api-access-svczt\") pod \"961bc3c9-b6be-4ece-98f1-971a583f9db8\" (UID: \"961bc3c9-b6be-4ece-98f1-971a583f9db8\") " Oct 02 01:51:21 crc kubenswrapper[4885]: I1002 01:51:21.209108 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/aa89bc07-6d38-40ba-8e90-9ecae726246d-marketplace-trusted-ca\") pod \"aa89bc07-6d38-40ba-8e90-9ecae726246d\" (UID: \"aa89bc07-6d38-40ba-8e90-9ecae726246d\") " Oct 02 01:51:21 crc kubenswrapper[4885]: I1002 01:51:21.209678 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aa89bc07-6d38-40ba-8e90-9ecae726246d-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "aa89bc07-6d38-40ba-8e90-9ecae726246d" (UID: "aa89bc07-6d38-40ba-8e90-9ecae726246d"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:51:21 crc kubenswrapper[4885]: I1002 01:51:21.209694 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/18a2b456-262c-43fd-97c6-bfaf2d658463-utilities" (OuterVolumeSpecName: "utilities") pod "18a2b456-262c-43fd-97c6-bfaf2d658463" (UID: "18a2b456-262c-43fd-97c6-bfaf2d658463"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 01:51:21 crc kubenswrapper[4885]: I1002 01:51:21.209819 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/aa89bc07-6d38-40ba-8e90-9ecae726246d-marketplace-operator-metrics\") pod \"aa89bc07-6d38-40ba-8e90-9ecae726246d\" (UID: \"aa89bc07-6d38-40ba-8e90-9ecae726246d\") " Oct 02 01:51:21 crc kubenswrapper[4885]: I1002 01:51:21.209911 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a28f881c-dd5e-4bf8-82ad-2c27b98dcceb-catalog-content\") pod \"a28f881c-dd5e-4bf8-82ad-2c27b98dcceb\" (UID: \"a28f881c-dd5e-4bf8-82ad-2c27b98dcceb\") " Oct 02 01:51:21 crc kubenswrapper[4885]: I1002 01:51:21.210001 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cdnq6\" (UniqueName: \"kubernetes.io/projected/a28f881c-dd5e-4bf8-82ad-2c27b98dcceb-kube-api-access-cdnq6\") pod \"a28f881c-dd5e-4bf8-82ad-2c27b98dcceb\" (UID: \"a28f881c-dd5e-4bf8-82ad-2c27b98dcceb\") " Oct 02 01:51:21 crc kubenswrapper[4885]: I1002 01:51:21.210032 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aecca2a3-40b0-4666-a86d-d67ecb580e99-utilities\") pod \"aecca2a3-40b0-4666-a86d-d67ecb580e99\" (UID: \"aecca2a3-40b0-4666-a86d-d67ecb580e99\") " Oct 02 01:51:21 crc kubenswrapper[4885]: I1002 01:51:21.210057 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aecca2a3-40b0-4666-a86d-d67ecb580e99-catalog-content\") pod \"aecca2a3-40b0-4666-a86d-d67ecb580e99\" (UID: \"aecca2a3-40b0-4666-a86d-d67ecb580e99\") " Oct 02 01:51:21 crc kubenswrapper[4885]: I1002 01:51:21.210102 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-smmgf\" (UniqueName: \"kubernetes.io/projected/18a2b456-262c-43fd-97c6-bfaf2d658463-kube-api-access-smmgf\") pod \"18a2b456-262c-43fd-97c6-bfaf2d658463\" (UID: \"18a2b456-262c-43fd-97c6-bfaf2d658463\") " Oct 02 01:51:21 crc kubenswrapper[4885]: I1002 01:51:21.210131 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a28f881c-dd5e-4bf8-82ad-2c27b98dcceb-utilities\") pod \"a28f881c-dd5e-4bf8-82ad-2c27b98dcceb\" (UID: \"a28f881c-dd5e-4bf8-82ad-2c27b98dcceb\") " Oct 02 01:51:21 crc kubenswrapper[4885]: I1002 01:51:21.210172 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18a2b456-262c-43fd-97c6-bfaf2d658463-catalog-content\") pod \"18a2b456-262c-43fd-97c6-bfaf2d658463\" (UID: \"18a2b456-262c-43fd-97c6-bfaf2d658463\") " Oct 02 01:51:21 crc kubenswrapper[4885]: I1002 01:51:21.210626 4885 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/961bc3c9-b6be-4ece-98f1-971a583f9db8-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 01:51:21 crc kubenswrapper[4885]: I1002 01:51:21.210642 4885 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18a2b456-262c-43fd-97c6-bfaf2d658463-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 01:51:21 crc kubenswrapper[4885]: I1002 01:51:21.210654 4885 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/aa89bc07-6d38-40ba-8e90-9ecae726246d-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 02 01:51:21 crc kubenswrapper[4885]: I1002 01:51:21.212015 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa89bc07-6d38-40ba-8e90-9ecae726246d-kube-api-access-nc9tk" (OuterVolumeSpecName: "kube-api-access-nc9tk") pod "aa89bc07-6d38-40ba-8e90-9ecae726246d" (UID: "aa89bc07-6d38-40ba-8e90-9ecae726246d"). InnerVolumeSpecName "kube-api-access-nc9tk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:51:21 crc kubenswrapper[4885]: I1002 01:51:21.212138 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aecca2a3-40b0-4666-a86d-d67ecb580e99-kube-api-access-jqcx5" (OuterVolumeSpecName: "kube-api-access-jqcx5") pod "aecca2a3-40b0-4666-a86d-d67ecb580e99" (UID: "aecca2a3-40b0-4666-a86d-d67ecb580e99"). InnerVolumeSpecName "kube-api-access-jqcx5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:51:21 crc kubenswrapper[4885]: I1002 01:51:21.212664 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/961bc3c9-b6be-4ece-98f1-971a583f9db8-kube-api-access-svczt" (OuterVolumeSpecName: "kube-api-access-svczt") pod "961bc3c9-b6be-4ece-98f1-971a583f9db8" (UID: "961bc3c9-b6be-4ece-98f1-971a583f9db8"). InnerVolumeSpecName "kube-api-access-svczt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:51:21 crc kubenswrapper[4885]: I1002 01:51:21.213198 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aecca2a3-40b0-4666-a86d-d67ecb580e99-utilities" (OuterVolumeSpecName: "utilities") pod "aecca2a3-40b0-4666-a86d-d67ecb580e99" (UID: "aecca2a3-40b0-4666-a86d-d67ecb580e99"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 01:51:21 crc kubenswrapper[4885]: I1002 01:51:21.214395 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a28f881c-dd5e-4bf8-82ad-2c27b98dcceb-kube-api-access-cdnq6" (OuterVolumeSpecName: "kube-api-access-cdnq6") pod "a28f881c-dd5e-4bf8-82ad-2c27b98dcceb" (UID: "a28f881c-dd5e-4bf8-82ad-2c27b98dcceb"). InnerVolumeSpecName "kube-api-access-cdnq6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:51:21 crc kubenswrapper[4885]: I1002 01:51:21.214817 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a28f881c-dd5e-4bf8-82ad-2c27b98dcceb-utilities" (OuterVolumeSpecName: "utilities") pod "a28f881c-dd5e-4bf8-82ad-2c27b98dcceb" (UID: "a28f881c-dd5e-4bf8-82ad-2c27b98dcceb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 01:51:21 crc kubenswrapper[4885]: I1002 01:51:21.215126 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa89bc07-6d38-40ba-8e90-9ecae726246d-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "aa89bc07-6d38-40ba-8e90-9ecae726246d" (UID: "aa89bc07-6d38-40ba-8e90-9ecae726246d"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:51:21 crc kubenswrapper[4885]: I1002 01:51:21.219631 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18a2b456-262c-43fd-97c6-bfaf2d658463-kube-api-access-smmgf" (OuterVolumeSpecName: "kube-api-access-smmgf") pod "18a2b456-262c-43fd-97c6-bfaf2d658463" (UID: "18a2b456-262c-43fd-97c6-bfaf2d658463"). InnerVolumeSpecName "kube-api-access-smmgf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:51:21 crc kubenswrapper[4885]: I1002 01:51:21.234536 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aecca2a3-40b0-4666-a86d-d67ecb580e99-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "aecca2a3-40b0-4666-a86d-d67ecb580e99" (UID: "aecca2a3-40b0-4666-a86d-d67ecb580e99"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 01:51:21 crc kubenswrapper[4885]: I1002 01:51:21.276619 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/961bc3c9-b6be-4ece-98f1-971a583f9db8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "961bc3c9-b6be-4ece-98f1-971a583f9db8" (UID: "961bc3c9-b6be-4ece-98f1-971a583f9db8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 01:51:21 crc kubenswrapper[4885]: I1002 01:51:21.293028 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a28f881c-dd5e-4bf8-82ad-2c27b98dcceb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a28f881c-dd5e-4bf8-82ad-2c27b98dcceb" (UID: "a28f881c-dd5e-4bf8-82ad-2c27b98dcceb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 01:51:21 crc kubenswrapper[4885]: I1002 01:51:21.307914 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/18a2b456-262c-43fd-97c6-bfaf2d658463-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "18a2b456-262c-43fd-97c6-bfaf2d658463" (UID: "18a2b456-262c-43fd-97c6-bfaf2d658463"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 01:51:21 crc kubenswrapper[4885]: I1002 01:51:21.311542 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nc9tk\" (UniqueName: \"kubernetes.io/projected/aa89bc07-6d38-40ba-8e90-9ecae726246d-kube-api-access-nc9tk\") on node \"crc\" DevicePath \"\"" Oct 02 01:51:21 crc kubenswrapper[4885]: I1002 01:51:21.311564 4885 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/961bc3c9-b6be-4ece-98f1-971a583f9db8-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 01:51:21 crc kubenswrapper[4885]: I1002 01:51:21.311574 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jqcx5\" (UniqueName: \"kubernetes.io/projected/aecca2a3-40b0-4666-a86d-d67ecb580e99-kube-api-access-jqcx5\") on node \"crc\" DevicePath \"\"" Oct 02 01:51:21 crc kubenswrapper[4885]: I1002 01:51:21.311588 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-svczt\" (UniqueName: \"kubernetes.io/projected/961bc3c9-b6be-4ece-98f1-971a583f9db8-kube-api-access-svczt\") on node \"crc\" DevicePath \"\"" Oct 02 01:51:21 crc kubenswrapper[4885]: I1002 01:51:21.311600 4885 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/aa89bc07-6d38-40ba-8e90-9ecae726246d-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 02 01:51:21 crc kubenswrapper[4885]: I1002 01:51:21.311611 4885 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a28f881c-dd5e-4bf8-82ad-2c27b98dcceb-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 01:51:21 crc kubenswrapper[4885]: I1002 01:51:21.311622 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cdnq6\" (UniqueName: \"kubernetes.io/projected/a28f881c-dd5e-4bf8-82ad-2c27b98dcceb-kube-api-access-cdnq6\") on node \"crc\" DevicePath \"\"" Oct 02 01:51:21 crc kubenswrapper[4885]: I1002 01:51:21.311632 4885 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aecca2a3-40b0-4666-a86d-d67ecb580e99-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 01:51:21 crc kubenswrapper[4885]: I1002 01:51:21.311641 4885 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aecca2a3-40b0-4666-a86d-d67ecb580e99-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 01:51:21 crc kubenswrapper[4885]: I1002 01:51:21.311652 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-smmgf\" (UniqueName: \"kubernetes.io/projected/18a2b456-262c-43fd-97c6-bfaf2d658463-kube-api-access-smmgf\") on node \"crc\" DevicePath \"\"" Oct 02 01:51:21 crc kubenswrapper[4885]: I1002 01:51:21.311663 4885 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a28f881c-dd5e-4bf8-82ad-2c27b98dcceb-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 01:51:21 crc kubenswrapper[4885]: I1002 01:51:21.311673 4885 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18a2b456-262c-43fd-97c6-bfaf2d658463-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 01:51:21 crc kubenswrapper[4885]: I1002 01:51:21.515378 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-z7fcl"] Oct 02 01:51:21 crc kubenswrapper[4885]: I1002 01:51:21.806137 4885 generic.go:334] "Generic (PLEG): container finished" podID="18a2b456-262c-43fd-97c6-bfaf2d658463" containerID="9248711d02d76c11268438ac426cc0097b451fcdadac403d61ebe565076fff04" exitCode=0 Oct 02 01:51:21 crc kubenswrapper[4885]: I1002 01:51:21.806427 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bjxpx" event={"ID":"18a2b456-262c-43fd-97c6-bfaf2d658463","Type":"ContainerDied","Data":"9248711d02d76c11268438ac426cc0097b451fcdadac403d61ebe565076fff04"} Oct 02 01:51:21 crc kubenswrapper[4885]: I1002 01:51:21.806455 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bjxpx" event={"ID":"18a2b456-262c-43fd-97c6-bfaf2d658463","Type":"ContainerDied","Data":"e78614ca8f60a5b44f38c8025dd93bb1c8d5d7b8f35750263ca0a38471bcc53e"} Oct 02 01:51:21 crc kubenswrapper[4885]: I1002 01:51:21.806472 4885 scope.go:117] "RemoveContainer" containerID="9248711d02d76c11268438ac426cc0097b451fcdadac403d61ebe565076fff04" Oct 02 01:51:21 crc kubenswrapper[4885]: I1002 01:51:21.806582 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bjxpx" Oct 02 01:51:21 crc kubenswrapper[4885]: I1002 01:51:21.817871 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-z7fcl" event={"ID":"b0909ffb-180c-4e74-9916-98de47f2419d","Type":"ContainerStarted","Data":"4ae0ca1c961a3c520cbd1a6e4cc97c20ccb51bbfca7b707644e4c53f687ed8a5"} Oct 02 01:51:21 crc kubenswrapper[4885]: I1002 01:51:21.817914 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-z7fcl" event={"ID":"b0909ffb-180c-4e74-9916-98de47f2419d","Type":"ContainerStarted","Data":"8b6d099ee2e882a79acc32c05400732b278ba83f47f20616d5d0a7b8c5dbdb4f"} Oct 02 01:51:21 crc kubenswrapper[4885]: I1002 01:51:21.823378 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-z7fcl" Oct 02 01:51:21 crc kubenswrapper[4885]: I1002 01:51:21.825468 4885 scope.go:117] "RemoveContainer" containerID="7da19d71cc38343e6a7f281f23d03e3b06ebb566c68d8495c0bfa7179aa213d5" Oct 02 01:51:21 crc kubenswrapper[4885]: I1002 01:51:21.827877 4885 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-z7fcl container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.55:8080/healthz\": dial tcp 10.217.0.55:8080: connect: connection refused" start-of-body= Oct 02 01:51:21 crc kubenswrapper[4885]: I1002 01:51:21.827930 4885 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-z7fcl" podUID="b0909ffb-180c-4e74-9916-98de47f2419d" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.55:8080/healthz\": dial tcp 10.217.0.55:8080: connect: connection refused" Oct 02 01:51:21 crc kubenswrapper[4885]: I1002 01:51:21.830154 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vb2j4" event={"ID":"a28f881c-dd5e-4bf8-82ad-2c27b98dcceb","Type":"ContainerDied","Data":"168bb5104de006a02899dc8905b192d43d958565115989616801f8349aa9e56f"} Oct 02 01:51:21 crc kubenswrapper[4885]: I1002 01:51:21.830238 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vb2j4" Oct 02 01:51:21 crc kubenswrapper[4885]: I1002 01:51:21.833315 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-z4rlr" event={"ID":"aa89bc07-6d38-40ba-8e90-9ecae726246d","Type":"ContainerDied","Data":"99bdb8b94a6ada7de44e3b7bef3ef8e258caabc772eefeaa6ac77cb8b466dcb4"} Oct 02 01:51:21 crc kubenswrapper[4885]: I1002 01:51:21.833416 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-z4rlr" Oct 02 01:51:21 crc kubenswrapper[4885]: I1002 01:51:21.845028 4885 generic.go:334] "Generic (PLEG): container finished" podID="aecca2a3-40b0-4666-a86d-d67ecb580e99" containerID="00335c982112fd4f0ea73d0823b195c015c56ae6305bfcd2e105be3c2c335313" exitCode=0 Oct 02 01:51:21 crc kubenswrapper[4885]: I1002 01:51:21.845142 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bz9hg" Oct 02 01:51:21 crc kubenswrapper[4885]: I1002 01:51:21.845588 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bz9hg" event={"ID":"aecca2a3-40b0-4666-a86d-d67ecb580e99","Type":"ContainerDied","Data":"00335c982112fd4f0ea73d0823b195c015c56ae6305bfcd2e105be3c2c335313"} Oct 02 01:51:21 crc kubenswrapper[4885]: I1002 01:51:21.845633 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bz9hg" event={"ID":"aecca2a3-40b0-4666-a86d-d67ecb580e99","Type":"ContainerDied","Data":"782ea6d3943d5c7111e54f12edf8a09d145a4689c9943241adbb6da0379ff934"} Oct 02 01:51:21 crc kubenswrapper[4885]: I1002 01:51:21.853971 4885 scope.go:117] "RemoveContainer" containerID="7e2a6b3be7e1a960b7e630282ad27b151752fa01cd3c4af7d6808b5f6e302500" Oct 02 01:51:21 crc kubenswrapper[4885]: I1002 01:51:21.859005 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x44t6" event={"ID":"961bc3c9-b6be-4ece-98f1-971a583f9db8","Type":"ContainerDied","Data":"24036c008472e8e838156422e286e72d19432b6e007bf9d20ad0da5750f5f8d9"} Oct 02 01:51:21 crc kubenswrapper[4885]: I1002 01:51:21.859100 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-x44t6" Oct 02 01:51:21 crc kubenswrapper[4885]: I1002 01:51:21.859867 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-z7fcl" podStartSLOduration=1.8598554219999999 podStartE2EDuration="1.859855422s" podCreationTimestamp="2025-10-02 01:51:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:51:21.844495426 +0000 UTC m=+270.656242825" watchObservedRunningTime="2025-10-02 01:51:21.859855422 +0000 UTC m=+270.671602821" Oct 02 01:51:21 crc kubenswrapper[4885]: I1002 01:51:21.861977 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bjxpx"] Oct 02 01:51:21 crc kubenswrapper[4885]: I1002 01:51:21.866616 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-bjxpx"] Oct 02 01:51:21 crc kubenswrapper[4885]: I1002 01:51:21.890401 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-z4rlr"] Oct 02 01:51:21 crc kubenswrapper[4885]: I1002 01:51:21.893073 4885 scope.go:117] "RemoveContainer" containerID="9248711d02d76c11268438ac426cc0097b451fcdadac403d61ebe565076fff04" Oct 02 01:51:21 crc kubenswrapper[4885]: E1002 01:51:21.894510 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9248711d02d76c11268438ac426cc0097b451fcdadac403d61ebe565076fff04\": container with ID starting with 9248711d02d76c11268438ac426cc0097b451fcdadac403d61ebe565076fff04 not found: ID does not exist" containerID="9248711d02d76c11268438ac426cc0097b451fcdadac403d61ebe565076fff04" Oct 02 01:51:21 crc kubenswrapper[4885]: I1002 01:51:21.894559 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9248711d02d76c11268438ac426cc0097b451fcdadac403d61ebe565076fff04"} err="failed to get container status \"9248711d02d76c11268438ac426cc0097b451fcdadac403d61ebe565076fff04\": rpc error: code = NotFound desc = could not find container \"9248711d02d76c11268438ac426cc0097b451fcdadac403d61ebe565076fff04\": container with ID starting with 9248711d02d76c11268438ac426cc0097b451fcdadac403d61ebe565076fff04 not found: ID does not exist" Oct 02 01:51:21 crc kubenswrapper[4885]: I1002 01:51:21.894594 4885 scope.go:117] "RemoveContainer" containerID="7da19d71cc38343e6a7f281f23d03e3b06ebb566c68d8495c0bfa7179aa213d5" Oct 02 01:51:21 crc kubenswrapper[4885]: E1002 01:51:21.895168 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7da19d71cc38343e6a7f281f23d03e3b06ebb566c68d8495c0bfa7179aa213d5\": container with ID starting with 7da19d71cc38343e6a7f281f23d03e3b06ebb566c68d8495c0bfa7179aa213d5 not found: ID does not exist" containerID="7da19d71cc38343e6a7f281f23d03e3b06ebb566c68d8495c0bfa7179aa213d5" Oct 02 01:51:21 crc kubenswrapper[4885]: I1002 01:51:21.895201 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7da19d71cc38343e6a7f281f23d03e3b06ebb566c68d8495c0bfa7179aa213d5"} err="failed to get container status \"7da19d71cc38343e6a7f281f23d03e3b06ebb566c68d8495c0bfa7179aa213d5\": rpc error: code = NotFound desc = could not find container \"7da19d71cc38343e6a7f281f23d03e3b06ebb566c68d8495c0bfa7179aa213d5\": container with ID starting with 7da19d71cc38343e6a7f281f23d03e3b06ebb566c68d8495c0bfa7179aa213d5 not found: ID does not exist" Oct 02 01:51:21 crc kubenswrapper[4885]: I1002 01:51:21.895215 4885 scope.go:117] "RemoveContainer" containerID="7e2a6b3be7e1a960b7e630282ad27b151752fa01cd3c4af7d6808b5f6e302500" Oct 02 01:51:21 crc kubenswrapper[4885]: E1002 01:51:21.895554 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7e2a6b3be7e1a960b7e630282ad27b151752fa01cd3c4af7d6808b5f6e302500\": container with ID starting with 7e2a6b3be7e1a960b7e630282ad27b151752fa01cd3c4af7d6808b5f6e302500 not found: ID does not exist" containerID="7e2a6b3be7e1a960b7e630282ad27b151752fa01cd3c4af7d6808b5f6e302500" Oct 02 01:51:21 crc kubenswrapper[4885]: I1002 01:51:21.895595 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7e2a6b3be7e1a960b7e630282ad27b151752fa01cd3c4af7d6808b5f6e302500"} err="failed to get container status \"7e2a6b3be7e1a960b7e630282ad27b151752fa01cd3c4af7d6808b5f6e302500\": rpc error: code = NotFound desc = could not find container \"7e2a6b3be7e1a960b7e630282ad27b151752fa01cd3c4af7d6808b5f6e302500\": container with ID starting with 7e2a6b3be7e1a960b7e630282ad27b151752fa01cd3c4af7d6808b5f6e302500 not found: ID does not exist" Oct 02 01:51:21 crc kubenswrapper[4885]: I1002 01:51:21.895625 4885 scope.go:117] "RemoveContainer" containerID="fe571f9130c9ad468ffd6ca21eef2aa385a83079b41b04f8e414c3e5450e39aa" Oct 02 01:51:21 crc kubenswrapper[4885]: I1002 01:51:21.895917 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-z4rlr"] Oct 02 01:51:21 crc kubenswrapper[4885]: I1002 01:51:21.915451 4885 scope.go:117] "RemoveContainer" containerID="d93f7aa02729b304652b4e31bacbaa2bfd913a75cf17811af175d9bee828a608" Oct 02 01:51:21 crc kubenswrapper[4885]: I1002 01:51:21.919124 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bz9hg"] Oct 02 01:51:21 crc kubenswrapper[4885]: I1002 01:51:21.921521 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-bz9hg"] Oct 02 01:51:21 crc kubenswrapper[4885]: I1002 01:51:21.931048 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-x44t6"] Oct 02 01:51:21 crc kubenswrapper[4885]: I1002 01:51:21.934507 4885 scope.go:117] "RemoveContainer" containerID="2efd1cee25f69426399bbace390e0354e4fe87fc61ac719c7e5bd32126074664" Oct 02 01:51:21 crc kubenswrapper[4885]: I1002 01:51:21.936765 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-x44t6"] Oct 02 01:51:21 crc kubenswrapper[4885]: I1002 01:51:21.944801 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vb2j4"] Oct 02 01:51:21 crc kubenswrapper[4885]: I1002 01:51:21.947593 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-vb2j4"] Oct 02 01:51:21 crc kubenswrapper[4885]: I1002 01:51:21.950117 4885 scope.go:117] "RemoveContainer" containerID="f04387030d901d22dd8ad8245b4b97815ac740667f7a7642348a4bfb750b45f7" Oct 02 01:51:21 crc kubenswrapper[4885]: I1002 01:51:21.966540 4885 scope.go:117] "RemoveContainer" containerID="00335c982112fd4f0ea73d0823b195c015c56ae6305bfcd2e105be3c2c335313" Oct 02 01:51:21 crc kubenswrapper[4885]: I1002 01:51:21.986539 4885 scope.go:117] "RemoveContainer" containerID="21589b19918287bd89fce80ba0975e43acbe59e01dba132d76542afd723ef0ed" Oct 02 01:51:22 crc kubenswrapper[4885]: I1002 01:51:22.004197 4885 scope.go:117] "RemoveContainer" containerID="f2bc855d60ae2f644fbcb6ce6f352db3d78ca46e877869ee645e59ef859d6482" Oct 02 01:51:22 crc kubenswrapper[4885]: I1002 01:51:22.026550 4885 scope.go:117] "RemoveContainer" containerID="00335c982112fd4f0ea73d0823b195c015c56ae6305bfcd2e105be3c2c335313" Oct 02 01:51:22 crc kubenswrapper[4885]: E1002 01:51:22.027330 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"00335c982112fd4f0ea73d0823b195c015c56ae6305bfcd2e105be3c2c335313\": container with ID starting with 00335c982112fd4f0ea73d0823b195c015c56ae6305bfcd2e105be3c2c335313 not found: ID does not exist" containerID="00335c982112fd4f0ea73d0823b195c015c56ae6305bfcd2e105be3c2c335313" Oct 02 01:51:22 crc kubenswrapper[4885]: I1002 01:51:22.027362 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00335c982112fd4f0ea73d0823b195c015c56ae6305bfcd2e105be3c2c335313"} err="failed to get container status \"00335c982112fd4f0ea73d0823b195c015c56ae6305bfcd2e105be3c2c335313\": rpc error: code = NotFound desc = could not find container \"00335c982112fd4f0ea73d0823b195c015c56ae6305bfcd2e105be3c2c335313\": container with ID starting with 00335c982112fd4f0ea73d0823b195c015c56ae6305bfcd2e105be3c2c335313 not found: ID does not exist" Oct 02 01:51:22 crc kubenswrapper[4885]: I1002 01:51:22.027383 4885 scope.go:117] "RemoveContainer" containerID="21589b19918287bd89fce80ba0975e43acbe59e01dba132d76542afd723ef0ed" Oct 02 01:51:22 crc kubenswrapper[4885]: E1002 01:51:22.027698 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"21589b19918287bd89fce80ba0975e43acbe59e01dba132d76542afd723ef0ed\": container with ID starting with 21589b19918287bd89fce80ba0975e43acbe59e01dba132d76542afd723ef0ed not found: ID does not exist" containerID="21589b19918287bd89fce80ba0975e43acbe59e01dba132d76542afd723ef0ed" Oct 02 01:51:22 crc kubenswrapper[4885]: I1002 01:51:22.027756 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"21589b19918287bd89fce80ba0975e43acbe59e01dba132d76542afd723ef0ed"} err="failed to get container status \"21589b19918287bd89fce80ba0975e43acbe59e01dba132d76542afd723ef0ed\": rpc error: code = NotFound desc = could not find container \"21589b19918287bd89fce80ba0975e43acbe59e01dba132d76542afd723ef0ed\": container with ID starting with 21589b19918287bd89fce80ba0975e43acbe59e01dba132d76542afd723ef0ed not found: ID does not exist" Oct 02 01:51:22 crc kubenswrapper[4885]: I1002 01:51:22.027784 4885 scope.go:117] "RemoveContainer" containerID="f2bc855d60ae2f644fbcb6ce6f352db3d78ca46e877869ee645e59ef859d6482" Oct 02 01:51:22 crc kubenswrapper[4885]: E1002 01:51:22.028200 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f2bc855d60ae2f644fbcb6ce6f352db3d78ca46e877869ee645e59ef859d6482\": container with ID starting with f2bc855d60ae2f644fbcb6ce6f352db3d78ca46e877869ee645e59ef859d6482 not found: ID does not exist" containerID="f2bc855d60ae2f644fbcb6ce6f352db3d78ca46e877869ee645e59ef859d6482" Oct 02 01:51:22 crc kubenswrapper[4885]: I1002 01:51:22.028220 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f2bc855d60ae2f644fbcb6ce6f352db3d78ca46e877869ee645e59ef859d6482"} err="failed to get container status \"f2bc855d60ae2f644fbcb6ce6f352db3d78ca46e877869ee645e59ef859d6482\": rpc error: code = NotFound desc = could not find container \"f2bc855d60ae2f644fbcb6ce6f352db3d78ca46e877869ee645e59ef859d6482\": container with ID starting with f2bc855d60ae2f644fbcb6ce6f352db3d78ca46e877869ee645e59ef859d6482 not found: ID does not exist" Oct 02 01:51:22 crc kubenswrapper[4885]: I1002 01:51:22.028235 4885 scope.go:117] "RemoveContainer" containerID="fa91784093ffc8297aea22d51f40aa7fe56e7c4e2ea262606bd94b010b505037" Oct 02 01:51:22 crc kubenswrapper[4885]: I1002 01:51:22.039235 4885 scope.go:117] "RemoveContainer" containerID="25eb9e8df43c4aec5ea67236b9ef11962fafa9abcfcb04b1451581f12d298a85" Oct 02 01:51:22 crc kubenswrapper[4885]: I1002 01:51:22.065529 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="18a2b456-262c-43fd-97c6-bfaf2d658463" path="/var/lib/kubelet/pods/18a2b456-262c-43fd-97c6-bfaf2d658463/volumes" Oct 02 01:51:22 crc kubenswrapper[4885]: I1002 01:51:22.066108 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="961bc3c9-b6be-4ece-98f1-971a583f9db8" path="/var/lib/kubelet/pods/961bc3c9-b6be-4ece-98f1-971a583f9db8/volumes" Oct 02 01:51:22 crc kubenswrapper[4885]: I1002 01:51:22.066663 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a28f881c-dd5e-4bf8-82ad-2c27b98dcceb" path="/var/lib/kubelet/pods/a28f881c-dd5e-4bf8-82ad-2c27b98dcceb/volumes" Oct 02 01:51:22 crc kubenswrapper[4885]: I1002 01:51:22.071362 4885 scope.go:117] "RemoveContainer" containerID="90606a90b0518a25cced263a6b69e97903effb88b507ade600dace77b974890e" Oct 02 01:51:22 crc kubenswrapper[4885]: I1002 01:51:22.072989 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aa89bc07-6d38-40ba-8e90-9ecae726246d" path="/var/lib/kubelet/pods/aa89bc07-6d38-40ba-8e90-9ecae726246d/volumes" Oct 02 01:51:22 crc kubenswrapper[4885]: I1002 01:51:22.075758 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aecca2a3-40b0-4666-a86d-d67ecb580e99" path="/var/lib/kubelet/pods/aecca2a3-40b0-4666-a86d-d67ecb580e99/volumes" Oct 02 01:51:22 crc kubenswrapper[4885]: I1002 01:51:22.442851 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-sn6xg"] Oct 02 01:51:22 crc kubenswrapper[4885]: E1002 01:51:22.443201 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a28f881c-dd5e-4bf8-82ad-2c27b98dcceb" containerName="extract-utilities" Oct 02 01:51:22 crc kubenswrapper[4885]: I1002 01:51:22.443227 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="a28f881c-dd5e-4bf8-82ad-2c27b98dcceb" containerName="extract-utilities" Oct 02 01:51:22 crc kubenswrapper[4885]: E1002 01:51:22.443255 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="961bc3c9-b6be-4ece-98f1-971a583f9db8" containerName="extract-utilities" Oct 02 01:51:22 crc kubenswrapper[4885]: I1002 01:51:22.443303 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="961bc3c9-b6be-4ece-98f1-971a583f9db8" containerName="extract-utilities" Oct 02 01:51:22 crc kubenswrapper[4885]: E1002 01:51:22.443323 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="961bc3c9-b6be-4ece-98f1-971a583f9db8" containerName="registry-server" Oct 02 01:51:22 crc kubenswrapper[4885]: I1002 01:51:22.443339 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="961bc3c9-b6be-4ece-98f1-971a583f9db8" containerName="registry-server" Oct 02 01:51:22 crc kubenswrapper[4885]: E1002 01:51:22.443359 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18a2b456-262c-43fd-97c6-bfaf2d658463" containerName="registry-server" Oct 02 01:51:22 crc kubenswrapper[4885]: I1002 01:51:22.443375 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="18a2b456-262c-43fd-97c6-bfaf2d658463" containerName="registry-server" Oct 02 01:51:22 crc kubenswrapper[4885]: E1002 01:51:22.443400 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aecca2a3-40b0-4666-a86d-d67ecb580e99" containerName="registry-server" Oct 02 01:51:22 crc kubenswrapper[4885]: I1002 01:51:22.443416 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="aecca2a3-40b0-4666-a86d-d67ecb580e99" containerName="registry-server" Oct 02 01:51:22 crc kubenswrapper[4885]: E1002 01:51:22.443444 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aecca2a3-40b0-4666-a86d-d67ecb580e99" containerName="extract-utilities" Oct 02 01:51:22 crc kubenswrapper[4885]: I1002 01:51:22.443460 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="aecca2a3-40b0-4666-a86d-d67ecb580e99" containerName="extract-utilities" Oct 02 01:51:22 crc kubenswrapper[4885]: E1002 01:51:22.443482 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18a2b456-262c-43fd-97c6-bfaf2d658463" containerName="extract-utilities" Oct 02 01:51:22 crc kubenswrapper[4885]: I1002 01:51:22.443497 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="18a2b456-262c-43fd-97c6-bfaf2d658463" containerName="extract-utilities" Oct 02 01:51:22 crc kubenswrapper[4885]: E1002 01:51:22.443534 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="961bc3c9-b6be-4ece-98f1-971a583f9db8" containerName="extract-content" Oct 02 01:51:22 crc kubenswrapper[4885]: I1002 01:51:22.443550 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="961bc3c9-b6be-4ece-98f1-971a583f9db8" containerName="extract-content" Oct 02 01:51:22 crc kubenswrapper[4885]: E1002 01:51:22.443575 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa89bc07-6d38-40ba-8e90-9ecae726246d" containerName="marketplace-operator" Oct 02 01:51:22 crc kubenswrapper[4885]: I1002 01:51:22.443591 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa89bc07-6d38-40ba-8e90-9ecae726246d" containerName="marketplace-operator" Oct 02 01:51:22 crc kubenswrapper[4885]: E1002 01:51:22.443612 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aecca2a3-40b0-4666-a86d-d67ecb580e99" containerName="extract-content" Oct 02 01:51:22 crc kubenswrapper[4885]: I1002 01:51:22.443629 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="aecca2a3-40b0-4666-a86d-d67ecb580e99" containerName="extract-content" Oct 02 01:51:22 crc kubenswrapper[4885]: E1002 01:51:22.443649 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a28f881c-dd5e-4bf8-82ad-2c27b98dcceb" containerName="registry-server" Oct 02 01:51:22 crc kubenswrapper[4885]: I1002 01:51:22.443666 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="a28f881c-dd5e-4bf8-82ad-2c27b98dcceb" containerName="registry-server" Oct 02 01:51:22 crc kubenswrapper[4885]: E1002 01:51:22.443691 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18a2b456-262c-43fd-97c6-bfaf2d658463" containerName="extract-content" Oct 02 01:51:22 crc kubenswrapper[4885]: I1002 01:51:22.443708 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="18a2b456-262c-43fd-97c6-bfaf2d658463" containerName="extract-content" Oct 02 01:51:22 crc kubenswrapper[4885]: E1002 01:51:22.443732 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a28f881c-dd5e-4bf8-82ad-2c27b98dcceb" containerName="extract-content" Oct 02 01:51:22 crc kubenswrapper[4885]: I1002 01:51:22.443749 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="a28f881c-dd5e-4bf8-82ad-2c27b98dcceb" containerName="extract-content" Oct 02 01:51:22 crc kubenswrapper[4885]: I1002 01:51:22.443962 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="aecca2a3-40b0-4666-a86d-d67ecb580e99" containerName="registry-server" Oct 02 01:51:22 crc kubenswrapper[4885]: I1002 01:51:22.443997 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="961bc3c9-b6be-4ece-98f1-971a583f9db8" containerName="registry-server" Oct 02 01:51:22 crc kubenswrapper[4885]: I1002 01:51:22.444020 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="a28f881c-dd5e-4bf8-82ad-2c27b98dcceb" containerName="registry-server" Oct 02 01:51:22 crc kubenswrapper[4885]: I1002 01:51:22.444041 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="18a2b456-262c-43fd-97c6-bfaf2d658463" containerName="registry-server" Oct 02 01:51:22 crc kubenswrapper[4885]: I1002 01:51:22.444063 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa89bc07-6d38-40ba-8e90-9ecae726246d" containerName="marketplace-operator" Oct 02 01:51:22 crc kubenswrapper[4885]: I1002 01:51:22.445588 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sn6xg" Oct 02 01:51:22 crc kubenswrapper[4885]: I1002 01:51:22.447520 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-sn6xg"] Oct 02 01:51:22 crc kubenswrapper[4885]: I1002 01:51:22.448220 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 02 01:51:22 crc kubenswrapper[4885]: I1002 01:51:22.534002 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56c1dc36-0146-45ea-9396-990a988a6d9e-catalog-content\") pod \"certified-operators-sn6xg\" (UID: \"56c1dc36-0146-45ea-9396-990a988a6d9e\") " pod="openshift-marketplace/certified-operators-sn6xg" Oct 02 01:51:22 crc kubenswrapper[4885]: I1002 01:51:22.534209 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b8nm6\" (UniqueName: \"kubernetes.io/projected/56c1dc36-0146-45ea-9396-990a988a6d9e-kube-api-access-b8nm6\") pod \"certified-operators-sn6xg\" (UID: \"56c1dc36-0146-45ea-9396-990a988a6d9e\") " pod="openshift-marketplace/certified-operators-sn6xg" Oct 02 01:51:22 crc kubenswrapper[4885]: I1002 01:51:22.534311 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56c1dc36-0146-45ea-9396-990a988a6d9e-utilities\") pod \"certified-operators-sn6xg\" (UID: \"56c1dc36-0146-45ea-9396-990a988a6d9e\") " pod="openshift-marketplace/certified-operators-sn6xg" Oct 02 01:51:22 crc kubenswrapper[4885]: I1002 01:51:22.635156 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b8nm6\" (UniqueName: \"kubernetes.io/projected/56c1dc36-0146-45ea-9396-990a988a6d9e-kube-api-access-b8nm6\") pod \"certified-operators-sn6xg\" (UID: \"56c1dc36-0146-45ea-9396-990a988a6d9e\") " pod="openshift-marketplace/certified-operators-sn6xg" Oct 02 01:51:22 crc kubenswrapper[4885]: I1002 01:51:22.635198 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56c1dc36-0146-45ea-9396-990a988a6d9e-utilities\") pod \"certified-operators-sn6xg\" (UID: \"56c1dc36-0146-45ea-9396-990a988a6d9e\") " pod="openshift-marketplace/certified-operators-sn6xg" Oct 02 01:51:22 crc kubenswrapper[4885]: I1002 01:51:22.635242 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56c1dc36-0146-45ea-9396-990a988a6d9e-catalog-content\") pod \"certified-operators-sn6xg\" (UID: \"56c1dc36-0146-45ea-9396-990a988a6d9e\") " pod="openshift-marketplace/certified-operators-sn6xg" Oct 02 01:51:22 crc kubenswrapper[4885]: I1002 01:51:22.635729 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56c1dc36-0146-45ea-9396-990a988a6d9e-catalog-content\") pod \"certified-operators-sn6xg\" (UID: \"56c1dc36-0146-45ea-9396-990a988a6d9e\") " pod="openshift-marketplace/certified-operators-sn6xg" Oct 02 01:51:22 crc kubenswrapper[4885]: I1002 01:51:22.635848 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56c1dc36-0146-45ea-9396-990a988a6d9e-utilities\") pod \"certified-operators-sn6xg\" (UID: \"56c1dc36-0146-45ea-9396-990a988a6d9e\") " pod="openshift-marketplace/certified-operators-sn6xg" Oct 02 01:51:22 crc kubenswrapper[4885]: I1002 01:51:22.651725 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b8nm6\" (UniqueName: \"kubernetes.io/projected/56c1dc36-0146-45ea-9396-990a988a6d9e-kube-api-access-b8nm6\") pod \"certified-operators-sn6xg\" (UID: \"56c1dc36-0146-45ea-9396-990a988a6d9e\") " pod="openshift-marketplace/certified-operators-sn6xg" Oct 02 01:51:22 crc kubenswrapper[4885]: I1002 01:51:22.776442 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sn6xg" Oct 02 01:51:22 crc kubenswrapper[4885]: I1002 01:51:22.878981 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-z7fcl" Oct 02 01:51:22 crc kubenswrapper[4885]: I1002 01:51:22.981934 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-sn6xg"] Oct 02 01:51:23 crc kubenswrapper[4885]: I1002 01:51:23.436847 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-sbsjt"] Oct 02 01:51:23 crc kubenswrapper[4885]: I1002 01:51:23.438027 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sbsjt" Oct 02 01:51:23 crc kubenswrapper[4885]: I1002 01:51:23.440558 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 02 01:51:23 crc kubenswrapper[4885]: I1002 01:51:23.446832 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-sbsjt"] Oct 02 01:51:23 crc kubenswrapper[4885]: I1002 01:51:23.547372 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eb7219fa-b79f-48be-9c5d-6453330d9527-utilities\") pod \"redhat-marketplace-sbsjt\" (UID: \"eb7219fa-b79f-48be-9c5d-6453330d9527\") " pod="openshift-marketplace/redhat-marketplace-sbsjt" Oct 02 01:51:23 crc kubenswrapper[4885]: I1002 01:51:23.547445 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-962gf\" (UniqueName: \"kubernetes.io/projected/eb7219fa-b79f-48be-9c5d-6453330d9527-kube-api-access-962gf\") pod \"redhat-marketplace-sbsjt\" (UID: \"eb7219fa-b79f-48be-9c5d-6453330d9527\") " pod="openshift-marketplace/redhat-marketplace-sbsjt" Oct 02 01:51:23 crc kubenswrapper[4885]: I1002 01:51:23.547560 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eb7219fa-b79f-48be-9c5d-6453330d9527-catalog-content\") pod \"redhat-marketplace-sbsjt\" (UID: \"eb7219fa-b79f-48be-9c5d-6453330d9527\") " pod="openshift-marketplace/redhat-marketplace-sbsjt" Oct 02 01:51:23 crc kubenswrapper[4885]: I1002 01:51:23.648417 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eb7219fa-b79f-48be-9c5d-6453330d9527-utilities\") pod \"redhat-marketplace-sbsjt\" (UID: \"eb7219fa-b79f-48be-9c5d-6453330d9527\") " pod="openshift-marketplace/redhat-marketplace-sbsjt" Oct 02 01:51:23 crc kubenswrapper[4885]: I1002 01:51:23.648459 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-962gf\" (UniqueName: \"kubernetes.io/projected/eb7219fa-b79f-48be-9c5d-6453330d9527-kube-api-access-962gf\") pod \"redhat-marketplace-sbsjt\" (UID: \"eb7219fa-b79f-48be-9c5d-6453330d9527\") " pod="openshift-marketplace/redhat-marketplace-sbsjt" Oct 02 01:51:23 crc kubenswrapper[4885]: I1002 01:51:23.648488 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eb7219fa-b79f-48be-9c5d-6453330d9527-catalog-content\") pod \"redhat-marketplace-sbsjt\" (UID: \"eb7219fa-b79f-48be-9c5d-6453330d9527\") " pod="openshift-marketplace/redhat-marketplace-sbsjt" Oct 02 01:51:23 crc kubenswrapper[4885]: I1002 01:51:23.648847 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eb7219fa-b79f-48be-9c5d-6453330d9527-catalog-content\") pod \"redhat-marketplace-sbsjt\" (UID: \"eb7219fa-b79f-48be-9c5d-6453330d9527\") " pod="openshift-marketplace/redhat-marketplace-sbsjt" Oct 02 01:51:23 crc kubenswrapper[4885]: I1002 01:51:23.649050 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eb7219fa-b79f-48be-9c5d-6453330d9527-utilities\") pod \"redhat-marketplace-sbsjt\" (UID: \"eb7219fa-b79f-48be-9c5d-6453330d9527\") " pod="openshift-marketplace/redhat-marketplace-sbsjt" Oct 02 01:51:23 crc kubenswrapper[4885]: I1002 01:51:23.679110 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-962gf\" (UniqueName: \"kubernetes.io/projected/eb7219fa-b79f-48be-9c5d-6453330d9527-kube-api-access-962gf\") pod \"redhat-marketplace-sbsjt\" (UID: \"eb7219fa-b79f-48be-9c5d-6453330d9527\") " pod="openshift-marketplace/redhat-marketplace-sbsjt" Oct 02 01:51:23 crc kubenswrapper[4885]: I1002 01:51:23.762901 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sbsjt" Oct 02 01:51:23 crc kubenswrapper[4885]: I1002 01:51:23.879745 4885 generic.go:334] "Generic (PLEG): container finished" podID="56c1dc36-0146-45ea-9396-990a988a6d9e" containerID="998e7f87b6c7b4be831abc48629c418bc1c25e66b29266084c0162d72a74fe80" exitCode=0 Oct 02 01:51:23 crc kubenswrapper[4885]: I1002 01:51:23.879818 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sn6xg" event={"ID":"56c1dc36-0146-45ea-9396-990a988a6d9e","Type":"ContainerDied","Data":"998e7f87b6c7b4be831abc48629c418bc1c25e66b29266084c0162d72a74fe80"} Oct 02 01:51:23 crc kubenswrapper[4885]: I1002 01:51:23.879901 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sn6xg" event={"ID":"56c1dc36-0146-45ea-9396-990a988a6d9e","Type":"ContainerStarted","Data":"f98c2d983728e4ea22a880ad78eba520df501ada0556869c83cf0c077f9112ac"} Oct 02 01:51:24 crc kubenswrapper[4885]: I1002 01:51:24.026909 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-sbsjt"] Oct 02 01:51:24 crc kubenswrapper[4885]: I1002 01:51:24.839768 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-9tmpc"] Oct 02 01:51:24 crc kubenswrapper[4885]: I1002 01:51:24.842011 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9tmpc" Oct 02 01:51:24 crc kubenswrapper[4885]: I1002 01:51:24.844127 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 02 01:51:24 crc kubenswrapper[4885]: I1002 01:51:24.849142 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9tmpc"] Oct 02 01:51:24 crc kubenswrapper[4885]: I1002 01:51:24.898547 4885 generic.go:334] "Generic (PLEG): container finished" podID="eb7219fa-b79f-48be-9c5d-6453330d9527" containerID="d94d9d0c244141c2ecc5c835e0eb231da2a48817474320fda085067c113a9139" exitCode=0 Oct 02 01:51:24 crc kubenswrapper[4885]: I1002 01:51:24.898621 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sbsjt" event={"ID":"eb7219fa-b79f-48be-9c5d-6453330d9527","Type":"ContainerDied","Data":"d94d9d0c244141c2ecc5c835e0eb231da2a48817474320fda085067c113a9139"} Oct 02 01:51:24 crc kubenswrapper[4885]: I1002 01:51:24.898708 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sbsjt" event={"ID":"eb7219fa-b79f-48be-9c5d-6453330d9527","Type":"ContainerStarted","Data":"cc63919eae635b487fa1256ace7cf6a75db726aed3c29b2daebe788377407c68"} Oct 02 01:51:24 crc kubenswrapper[4885]: I1002 01:51:24.903361 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sn6xg" event={"ID":"56c1dc36-0146-45ea-9396-990a988a6d9e","Type":"ContainerStarted","Data":"304afc6ec4e90ba8f8a18f46d0fd19be51b159aa197963879198f257dc630739"} Oct 02 01:51:24 crc kubenswrapper[4885]: I1002 01:51:24.966432 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/747efce9-9b31-4e1b-a7b5-94019bcce7f4-utilities\") pod \"redhat-operators-9tmpc\" (UID: \"747efce9-9b31-4e1b-a7b5-94019bcce7f4\") " pod="openshift-marketplace/redhat-operators-9tmpc" Oct 02 01:51:24 crc kubenswrapper[4885]: I1002 01:51:24.966501 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ncb2f\" (UniqueName: \"kubernetes.io/projected/747efce9-9b31-4e1b-a7b5-94019bcce7f4-kube-api-access-ncb2f\") pod \"redhat-operators-9tmpc\" (UID: \"747efce9-9b31-4e1b-a7b5-94019bcce7f4\") " pod="openshift-marketplace/redhat-operators-9tmpc" Oct 02 01:51:24 crc kubenswrapper[4885]: I1002 01:51:24.966563 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/747efce9-9b31-4e1b-a7b5-94019bcce7f4-catalog-content\") pod \"redhat-operators-9tmpc\" (UID: \"747efce9-9b31-4e1b-a7b5-94019bcce7f4\") " pod="openshift-marketplace/redhat-operators-9tmpc" Oct 02 01:51:25 crc kubenswrapper[4885]: I1002 01:51:25.067339 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/747efce9-9b31-4e1b-a7b5-94019bcce7f4-catalog-content\") pod \"redhat-operators-9tmpc\" (UID: \"747efce9-9b31-4e1b-a7b5-94019bcce7f4\") " pod="openshift-marketplace/redhat-operators-9tmpc" Oct 02 01:51:25 crc kubenswrapper[4885]: I1002 01:51:25.067420 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/747efce9-9b31-4e1b-a7b5-94019bcce7f4-utilities\") pod \"redhat-operators-9tmpc\" (UID: \"747efce9-9b31-4e1b-a7b5-94019bcce7f4\") " pod="openshift-marketplace/redhat-operators-9tmpc" Oct 02 01:51:25 crc kubenswrapper[4885]: I1002 01:51:25.067451 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ncb2f\" (UniqueName: \"kubernetes.io/projected/747efce9-9b31-4e1b-a7b5-94019bcce7f4-kube-api-access-ncb2f\") pod \"redhat-operators-9tmpc\" (UID: \"747efce9-9b31-4e1b-a7b5-94019bcce7f4\") " pod="openshift-marketplace/redhat-operators-9tmpc" Oct 02 01:51:25 crc kubenswrapper[4885]: I1002 01:51:25.068012 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/747efce9-9b31-4e1b-a7b5-94019bcce7f4-utilities\") pod \"redhat-operators-9tmpc\" (UID: \"747efce9-9b31-4e1b-a7b5-94019bcce7f4\") " pod="openshift-marketplace/redhat-operators-9tmpc" Oct 02 01:51:25 crc kubenswrapper[4885]: I1002 01:51:25.069309 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/747efce9-9b31-4e1b-a7b5-94019bcce7f4-catalog-content\") pod \"redhat-operators-9tmpc\" (UID: \"747efce9-9b31-4e1b-a7b5-94019bcce7f4\") " pod="openshift-marketplace/redhat-operators-9tmpc" Oct 02 01:51:25 crc kubenswrapper[4885]: I1002 01:51:25.091205 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ncb2f\" (UniqueName: \"kubernetes.io/projected/747efce9-9b31-4e1b-a7b5-94019bcce7f4-kube-api-access-ncb2f\") pod \"redhat-operators-9tmpc\" (UID: \"747efce9-9b31-4e1b-a7b5-94019bcce7f4\") " pod="openshift-marketplace/redhat-operators-9tmpc" Oct 02 01:51:25 crc kubenswrapper[4885]: I1002 01:51:25.179464 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9tmpc" Oct 02 01:51:25 crc kubenswrapper[4885]: I1002 01:51:25.618855 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9tmpc"] Oct 02 01:51:25 crc kubenswrapper[4885]: I1002 01:51:25.840718 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-xbmgl"] Oct 02 01:51:25 crc kubenswrapper[4885]: I1002 01:51:25.842429 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xbmgl" Oct 02 01:51:25 crc kubenswrapper[4885]: I1002 01:51:25.849198 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 02 01:51:25 crc kubenswrapper[4885]: I1002 01:51:25.851587 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xbmgl"] Oct 02 01:51:25 crc kubenswrapper[4885]: I1002 01:51:25.892131 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ls6jj\" (UniqueName: \"kubernetes.io/projected/d8e59d2f-c3ac-4a7a-b8a1-b1de59eb0724-kube-api-access-ls6jj\") pod \"community-operators-xbmgl\" (UID: \"d8e59d2f-c3ac-4a7a-b8a1-b1de59eb0724\") " pod="openshift-marketplace/community-operators-xbmgl" Oct 02 01:51:25 crc kubenswrapper[4885]: I1002 01:51:25.892205 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d8e59d2f-c3ac-4a7a-b8a1-b1de59eb0724-utilities\") pod \"community-operators-xbmgl\" (UID: \"d8e59d2f-c3ac-4a7a-b8a1-b1de59eb0724\") " pod="openshift-marketplace/community-operators-xbmgl" Oct 02 01:51:25 crc kubenswrapper[4885]: I1002 01:51:25.892311 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d8e59d2f-c3ac-4a7a-b8a1-b1de59eb0724-catalog-content\") pod \"community-operators-xbmgl\" (UID: \"d8e59d2f-c3ac-4a7a-b8a1-b1de59eb0724\") " pod="openshift-marketplace/community-operators-xbmgl" Oct 02 01:51:25 crc kubenswrapper[4885]: I1002 01:51:25.909129 4885 generic.go:334] "Generic (PLEG): container finished" podID="eb7219fa-b79f-48be-9c5d-6453330d9527" containerID="ed8a08055b50c28a759ed9efc3601473b9c7ff6b1b6ed0351286285b861884a6" exitCode=0 Oct 02 01:51:25 crc kubenswrapper[4885]: I1002 01:51:25.909197 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sbsjt" event={"ID":"eb7219fa-b79f-48be-9c5d-6453330d9527","Type":"ContainerDied","Data":"ed8a08055b50c28a759ed9efc3601473b9c7ff6b1b6ed0351286285b861884a6"} Oct 02 01:51:25 crc kubenswrapper[4885]: I1002 01:51:25.922788 4885 generic.go:334] "Generic (PLEG): container finished" podID="56c1dc36-0146-45ea-9396-990a988a6d9e" containerID="304afc6ec4e90ba8f8a18f46d0fd19be51b159aa197963879198f257dc630739" exitCode=0 Oct 02 01:51:25 crc kubenswrapper[4885]: I1002 01:51:25.922824 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sn6xg" event={"ID":"56c1dc36-0146-45ea-9396-990a988a6d9e","Type":"ContainerDied","Data":"304afc6ec4e90ba8f8a18f46d0fd19be51b159aa197963879198f257dc630739"} Oct 02 01:51:25 crc kubenswrapper[4885]: I1002 01:51:25.925218 4885 generic.go:334] "Generic (PLEG): container finished" podID="747efce9-9b31-4e1b-a7b5-94019bcce7f4" containerID="570411e1a75b1d80a1f2398b85c54a6738c136d18f1eecbf604b4ffeaaa5aa94" exitCode=0 Oct 02 01:51:25 crc kubenswrapper[4885]: I1002 01:51:25.925272 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9tmpc" event={"ID":"747efce9-9b31-4e1b-a7b5-94019bcce7f4","Type":"ContainerDied","Data":"570411e1a75b1d80a1f2398b85c54a6738c136d18f1eecbf604b4ffeaaa5aa94"} Oct 02 01:51:25 crc kubenswrapper[4885]: I1002 01:51:25.925302 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9tmpc" event={"ID":"747efce9-9b31-4e1b-a7b5-94019bcce7f4","Type":"ContainerStarted","Data":"0a0cb0edc97797955210faa3060b9bf60e92b5e9c6f7847cff53d2217f0fbc65"} Oct 02 01:51:25 crc kubenswrapper[4885]: I1002 01:51:25.993461 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ls6jj\" (UniqueName: \"kubernetes.io/projected/d8e59d2f-c3ac-4a7a-b8a1-b1de59eb0724-kube-api-access-ls6jj\") pod \"community-operators-xbmgl\" (UID: \"d8e59d2f-c3ac-4a7a-b8a1-b1de59eb0724\") " pod="openshift-marketplace/community-operators-xbmgl" Oct 02 01:51:25 crc kubenswrapper[4885]: I1002 01:51:25.993517 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d8e59d2f-c3ac-4a7a-b8a1-b1de59eb0724-utilities\") pod \"community-operators-xbmgl\" (UID: \"d8e59d2f-c3ac-4a7a-b8a1-b1de59eb0724\") " pod="openshift-marketplace/community-operators-xbmgl" Oct 02 01:51:25 crc kubenswrapper[4885]: I1002 01:51:25.993561 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d8e59d2f-c3ac-4a7a-b8a1-b1de59eb0724-catalog-content\") pod \"community-operators-xbmgl\" (UID: \"d8e59d2f-c3ac-4a7a-b8a1-b1de59eb0724\") " pod="openshift-marketplace/community-operators-xbmgl" Oct 02 01:51:25 crc kubenswrapper[4885]: I1002 01:51:25.994565 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d8e59d2f-c3ac-4a7a-b8a1-b1de59eb0724-catalog-content\") pod \"community-operators-xbmgl\" (UID: \"d8e59d2f-c3ac-4a7a-b8a1-b1de59eb0724\") " pod="openshift-marketplace/community-operators-xbmgl" Oct 02 01:51:25 crc kubenswrapper[4885]: I1002 01:51:25.994664 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d8e59d2f-c3ac-4a7a-b8a1-b1de59eb0724-utilities\") pod \"community-operators-xbmgl\" (UID: \"d8e59d2f-c3ac-4a7a-b8a1-b1de59eb0724\") " pod="openshift-marketplace/community-operators-xbmgl" Oct 02 01:51:26 crc kubenswrapper[4885]: I1002 01:51:26.016722 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ls6jj\" (UniqueName: \"kubernetes.io/projected/d8e59d2f-c3ac-4a7a-b8a1-b1de59eb0724-kube-api-access-ls6jj\") pod \"community-operators-xbmgl\" (UID: \"d8e59d2f-c3ac-4a7a-b8a1-b1de59eb0724\") " pod="openshift-marketplace/community-operators-xbmgl" Oct 02 01:51:26 crc kubenswrapper[4885]: I1002 01:51:26.197194 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xbmgl" Oct 02 01:51:26 crc kubenswrapper[4885]: I1002 01:51:26.612805 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xbmgl"] Oct 02 01:51:26 crc kubenswrapper[4885]: W1002 01:51:26.620115 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd8e59d2f_c3ac_4a7a_b8a1_b1de59eb0724.slice/crio-57ee1ae63045def09b2a2873b2a711978e5dadba12e46b71c161c55168372433 WatchSource:0}: Error finding container 57ee1ae63045def09b2a2873b2a711978e5dadba12e46b71c161c55168372433: Status 404 returned error can't find the container with id 57ee1ae63045def09b2a2873b2a711978e5dadba12e46b71c161c55168372433 Oct 02 01:51:26 crc kubenswrapper[4885]: I1002 01:51:26.931191 4885 generic.go:334] "Generic (PLEG): container finished" podID="d8e59d2f-c3ac-4a7a-b8a1-b1de59eb0724" containerID="a243c5573b5f9434a2cba53a4eab1357686bf0813c324ce1bb5e129e783dcb08" exitCode=0 Oct 02 01:51:26 crc kubenswrapper[4885]: I1002 01:51:26.931998 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xbmgl" event={"ID":"d8e59d2f-c3ac-4a7a-b8a1-b1de59eb0724","Type":"ContainerDied","Data":"a243c5573b5f9434a2cba53a4eab1357686bf0813c324ce1bb5e129e783dcb08"} Oct 02 01:51:26 crc kubenswrapper[4885]: I1002 01:51:26.932050 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xbmgl" event={"ID":"d8e59d2f-c3ac-4a7a-b8a1-b1de59eb0724","Type":"ContainerStarted","Data":"57ee1ae63045def09b2a2873b2a711978e5dadba12e46b71c161c55168372433"} Oct 02 01:51:26 crc kubenswrapper[4885]: I1002 01:51:26.937217 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sn6xg" event={"ID":"56c1dc36-0146-45ea-9396-990a988a6d9e","Type":"ContainerStarted","Data":"ba09b74d64572913c1f91336e9d759e1bdb3c913c7edbd42add86a447e85304d"} Oct 02 01:51:26 crc kubenswrapper[4885]: I1002 01:51:26.943833 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sbsjt" event={"ID":"eb7219fa-b79f-48be-9c5d-6453330d9527","Type":"ContainerStarted","Data":"053cd75a12470e09f93265ce06cdafe9f95590aeac13fd9b18bb43f85a55b66c"} Oct 02 01:51:26 crc kubenswrapper[4885]: I1002 01:51:26.976801 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-sbsjt" podStartSLOduration=2.486206707 podStartE2EDuration="3.976781056s" podCreationTimestamp="2025-10-02 01:51:23 +0000 UTC" firstStartedPulling="2025-10-02 01:51:24.901126968 +0000 UTC m=+273.712874377" lastFinishedPulling="2025-10-02 01:51:26.391701327 +0000 UTC m=+275.203448726" observedRunningTime="2025-10-02 01:51:26.975013548 +0000 UTC m=+275.786760947" watchObservedRunningTime="2025-10-02 01:51:26.976781056 +0000 UTC m=+275.788528455" Oct 02 01:51:27 crc kubenswrapper[4885]: I1002 01:51:27.001740 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-sn6xg" podStartSLOduration=2.502608769 podStartE2EDuration="5.00172079s" podCreationTimestamp="2025-10-02 01:51:22 +0000 UTC" firstStartedPulling="2025-10-02 01:51:23.881796357 +0000 UTC m=+272.693543756" lastFinishedPulling="2025-10-02 01:51:26.380908378 +0000 UTC m=+275.192655777" observedRunningTime="2025-10-02 01:51:27.001551125 +0000 UTC m=+275.813298524" watchObservedRunningTime="2025-10-02 01:51:27.00172079 +0000 UTC m=+275.813468189" Oct 02 01:51:27 crc kubenswrapper[4885]: I1002 01:51:27.949185 4885 generic.go:334] "Generic (PLEG): container finished" podID="747efce9-9b31-4e1b-a7b5-94019bcce7f4" containerID="257f185d0127275bb8101e9046f4c83f4e96edbe347d70f8578cd832d383fed0" exitCode=0 Oct 02 01:51:27 crc kubenswrapper[4885]: I1002 01:51:27.950549 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9tmpc" event={"ID":"747efce9-9b31-4e1b-a7b5-94019bcce7f4","Type":"ContainerDied","Data":"257f185d0127275bb8101e9046f4c83f4e96edbe347d70f8578cd832d383fed0"} Oct 02 01:51:28 crc kubenswrapper[4885]: I1002 01:51:28.959556 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xbmgl" event={"ID":"d8e59d2f-c3ac-4a7a-b8a1-b1de59eb0724","Type":"ContainerStarted","Data":"0eeda0ee5ad6e996e6dee63cee14734f6c68d5b9c989e9fde8e3fdddb56f93ce"} Oct 02 01:51:28 crc kubenswrapper[4885]: I1002 01:51:28.961602 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9tmpc" event={"ID":"747efce9-9b31-4e1b-a7b5-94019bcce7f4","Type":"ContainerStarted","Data":"0fb377d5ce3f85c37f74ccdaee9c467cac67295021a4bd326ec3bd4e20d6166a"} Oct 02 01:51:29 crc kubenswrapper[4885]: I1002 01:51:29.970871 4885 generic.go:334] "Generic (PLEG): container finished" podID="d8e59d2f-c3ac-4a7a-b8a1-b1de59eb0724" containerID="0eeda0ee5ad6e996e6dee63cee14734f6c68d5b9c989e9fde8e3fdddb56f93ce" exitCode=0 Oct 02 01:51:29 crc kubenswrapper[4885]: I1002 01:51:29.970970 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xbmgl" event={"ID":"d8e59d2f-c3ac-4a7a-b8a1-b1de59eb0724","Type":"ContainerDied","Data":"0eeda0ee5ad6e996e6dee63cee14734f6c68d5b9c989e9fde8e3fdddb56f93ce"} Oct 02 01:51:29 crc kubenswrapper[4885]: I1002 01:51:29.992019 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-9tmpc" podStartSLOduration=3.48638721 podStartE2EDuration="5.992003162s" podCreationTimestamp="2025-10-02 01:51:24 +0000 UTC" firstStartedPulling="2025-10-02 01:51:25.929710849 +0000 UTC m=+274.741458248" lastFinishedPulling="2025-10-02 01:51:28.435326771 +0000 UTC m=+277.247074200" observedRunningTime="2025-10-02 01:51:29.006813511 +0000 UTC m=+277.818560910" watchObservedRunningTime="2025-10-02 01:51:29.992003162 +0000 UTC m=+278.803750561" Oct 02 01:51:31 crc kubenswrapper[4885]: I1002 01:51:31.986235 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xbmgl" event={"ID":"d8e59d2f-c3ac-4a7a-b8a1-b1de59eb0724","Type":"ContainerStarted","Data":"8f80893f65e2a0cd18185a0245c5f27ffc1fa4a0ec3b943c8d628c0ade1c1220"} Oct 02 01:51:32 crc kubenswrapper[4885]: I1002 01:51:32.005367 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-xbmgl" podStartSLOduration=2.994421544 podStartE2EDuration="7.005347699s" podCreationTimestamp="2025-10-02 01:51:25 +0000 UTC" firstStartedPulling="2025-10-02 01:51:26.932648152 +0000 UTC m=+275.744395551" lastFinishedPulling="2025-10-02 01:51:30.943574277 +0000 UTC m=+279.755321706" observedRunningTime="2025-10-02 01:51:32.004236483 +0000 UTC m=+280.815983952" watchObservedRunningTime="2025-10-02 01:51:32.005347699 +0000 UTC m=+280.817095128" Oct 02 01:51:32 crc kubenswrapper[4885]: I1002 01:51:32.777370 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-sn6xg" Oct 02 01:51:32 crc kubenswrapper[4885]: I1002 01:51:32.777423 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-sn6xg" Oct 02 01:51:32 crc kubenswrapper[4885]: I1002 01:51:32.820900 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-sn6xg" Oct 02 01:51:33 crc kubenswrapper[4885]: I1002 01:51:33.051357 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-sn6xg" Oct 02 01:51:33 crc kubenswrapper[4885]: I1002 01:51:33.763858 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-sbsjt" Oct 02 01:51:33 crc kubenswrapper[4885]: I1002 01:51:33.764433 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-sbsjt" Oct 02 01:51:33 crc kubenswrapper[4885]: I1002 01:51:33.814568 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-sbsjt" Oct 02 01:51:34 crc kubenswrapper[4885]: I1002 01:51:34.058340 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-sbsjt" Oct 02 01:51:35 crc kubenswrapper[4885]: I1002 01:51:35.179739 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-9tmpc" Oct 02 01:51:35 crc kubenswrapper[4885]: I1002 01:51:35.179804 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-9tmpc" Oct 02 01:51:35 crc kubenswrapper[4885]: I1002 01:51:35.243861 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-9tmpc" Oct 02 01:51:36 crc kubenswrapper[4885]: I1002 01:51:36.063109 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-9tmpc" Oct 02 01:51:36 crc kubenswrapper[4885]: I1002 01:51:36.198153 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-xbmgl" Oct 02 01:51:36 crc kubenswrapper[4885]: I1002 01:51:36.198245 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-xbmgl" Oct 02 01:51:36 crc kubenswrapper[4885]: I1002 01:51:36.236249 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-xbmgl" Oct 02 01:51:37 crc kubenswrapper[4885]: I1002 01:51:37.064294 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-xbmgl" Oct 02 01:52:43 crc kubenswrapper[4885]: I1002 01:52:43.265574 4885 patch_prober.go:28] interesting pod/machine-config-daemon-rttx8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 01:52:43 crc kubenswrapper[4885]: I1002 01:52:43.266628 4885 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 01:53:13 crc kubenswrapper[4885]: I1002 01:53:13.266362 4885 patch_prober.go:28] interesting pod/machine-config-daemon-rttx8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 01:53:13 crc kubenswrapper[4885]: I1002 01:53:13.267116 4885 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 01:53:43 crc kubenswrapper[4885]: I1002 01:53:43.266032 4885 patch_prober.go:28] interesting pod/machine-config-daemon-rttx8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 01:53:43 crc kubenswrapper[4885]: I1002 01:53:43.266880 4885 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 01:53:43 crc kubenswrapper[4885]: I1002 01:53:43.266958 4885 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" Oct 02 01:53:43 crc kubenswrapper[4885]: I1002 01:53:43.268149 4885 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"edcfaf9e26fcac2c575da14a9bd0f19b12460248f9779d8e37fc41a2db9b606d"} pod="openshift-machine-config-operator/machine-config-daemon-rttx8" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 01:53:43 crc kubenswrapper[4885]: I1002 01:53:43.268859 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" containerName="machine-config-daemon" containerID="cri-o://edcfaf9e26fcac2c575da14a9bd0f19b12460248f9779d8e37fc41a2db9b606d" gracePeriod=600 Oct 02 01:53:43 crc kubenswrapper[4885]: I1002 01:53:43.848155 4885 generic.go:334] "Generic (PLEG): container finished" podID="d169c8ba-000a-4218-8e3b-4ae53035b110" containerID="edcfaf9e26fcac2c575da14a9bd0f19b12460248f9779d8e37fc41a2db9b606d" exitCode=0 Oct 02 01:53:43 crc kubenswrapper[4885]: I1002 01:53:43.848255 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" event={"ID":"d169c8ba-000a-4218-8e3b-4ae53035b110","Type":"ContainerDied","Data":"edcfaf9e26fcac2c575da14a9bd0f19b12460248f9779d8e37fc41a2db9b606d"} Oct 02 01:53:43 crc kubenswrapper[4885]: I1002 01:53:43.848669 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" event={"ID":"d169c8ba-000a-4218-8e3b-4ae53035b110","Type":"ContainerStarted","Data":"a1c63629d5a2ca3023ebf7324c43c17375f971212dce224761d6f69e9b4ea0ce"} Oct 02 01:53:43 crc kubenswrapper[4885]: I1002 01:53:43.848712 4885 scope.go:117] "RemoveContainer" containerID="09536e182a46e77318c10812d1777e551f834907c006cab5c726dc66f5bce88a" Oct 02 01:54:54 crc kubenswrapper[4885]: I1002 01:54:54.615384 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-sgr2r"] Oct 02 01:54:54 crc kubenswrapper[4885]: I1002 01:54:54.617240 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-sgr2r" Oct 02 01:54:54 crc kubenswrapper[4885]: I1002 01:54:54.638923 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-sgr2r"] Oct 02 01:54:54 crc kubenswrapper[4885]: I1002 01:54:54.818647 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/717bf496-3a04-4777-82fc-97a6072b9c7a-registry-tls\") pod \"image-registry-66df7c8f76-sgr2r\" (UID: \"717bf496-3a04-4777-82fc-97a6072b9c7a\") " pod="openshift-image-registry/image-registry-66df7c8f76-sgr2r" Oct 02 01:54:54 crc kubenswrapper[4885]: I1002 01:54:54.818794 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/717bf496-3a04-4777-82fc-97a6072b9c7a-trusted-ca\") pod \"image-registry-66df7c8f76-sgr2r\" (UID: \"717bf496-3a04-4777-82fc-97a6072b9c7a\") " pod="openshift-image-registry/image-registry-66df7c8f76-sgr2r" Oct 02 01:54:54 crc kubenswrapper[4885]: I1002 01:54:54.818921 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/717bf496-3a04-4777-82fc-97a6072b9c7a-installation-pull-secrets\") pod \"image-registry-66df7c8f76-sgr2r\" (UID: \"717bf496-3a04-4777-82fc-97a6072b9c7a\") " pod="openshift-image-registry/image-registry-66df7c8f76-sgr2r" Oct 02 01:54:54 crc kubenswrapper[4885]: I1002 01:54:54.819004 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vwc9m\" (UniqueName: \"kubernetes.io/projected/717bf496-3a04-4777-82fc-97a6072b9c7a-kube-api-access-vwc9m\") pod \"image-registry-66df7c8f76-sgr2r\" (UID: \"717bf496-3a04-4777-82fc-97a6072b9c7a\") " pod="openshift-image-registry/image-registry-66df7c8f76-sgr2r" Oct 02 01:54:54 crc kubenswrapper[4885]: I1002 01:54:54.819042 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-sgr2r\" (UID: \"717bf496-3a04-4777-82fc-97a6072b9c7a\") " pod="openshift-image-registry/image-registry-66df7c8f76-sgr2r" Oct 02 01:54:54 crc kubenswrapper[4885]: I1002 01:54:54.819125 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/717bf496-3a04-4777-82fc-97a6072b9c7a-bound-sa-token\") pod \"image-registry-66df7c8f76-sgr2r\" (UID: \"717bf496-3a04-4777-82fc-97a6072b9c7a\") " pod="openshift-image-registry/image-registry-66df7c8f76-sgr2r" Oct 02 01:54:54 crc kubenswrapper[4885]: I1002 01:54:54.819191 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/717bf496-3a04-4777-82fc-97a6072b9c7a-ca-trust-extracted\") pod \"image-registry-66df7c8f76-sgr2r\" (UID: \"717bf496-3a04-4777-82fc-97a6072b9c7a\") " pod="openshift-image-registry/image-registry-66df7c8f76-sgr2r" Oct 02 01:54:54 crc kubenswrapper[4885]: I1002 01:54:54.819232 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/717bf496-3a04-4777-82fc-97a6072b9c7a-registry-certificates\") pod \"image-registry-66df7c8f76-sgr2r\" (UID: \"717bf496-3a04-4777-82fc-97a6072b9c7a\") " pod="openshift-image-registry/image-registry-66df7c8f76-sgr2r" Oct 02 01:54:54 crc kubenswrapper[4885]: I1002 01:54:54.852809 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-sgr2r\" (UID: \"717bf496-3a04-4777-82fc-97a6072b9c7a\") " pod="openshift-image-registry/image-registry-66df7c8f76-sgr2r" Oct 02 01:54:54 crc kubenswrapper[4885]: I1002 01:54:54.920380 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/717bf496-3a04-4777-82fc-97a6072b9c7a-trusted-ca\") pod \"image-registry-66df7c8f76-sgr2r\" (UID: \"717bf496-3a04-4777-82fc-97a6072b9c7a\") " pod="openshift-image-registry/image-registry-66df7c8f76-sgr2r" Oct 02 01:54:54 crc kubenswrapper[4885]: I1002 01:54:54.920466 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/717bf496-3a04-4777-82fc-97a6072b9c7a-installation-pull-secrets\") pod \"image-registry-66df7c8f76-sgr2r\" (UID: \"717bf496-3a04-4777-82fc-97a6072b9c7a\") " pod="openshift-image-registry/image-registry-66df7c8f76-sgr2r" Oct 02 01:54:54 crc kubenswrapper[4885]: I1002 01:54:54.920524 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vwc9m\" (UniqueName: \"kubernetes.io/projected/717bf496-3a04-4777-82fc-97a6072b9c7a-kube-api-access-vwc9m\") pod \"image-registry-66df7c8f76-sgr2r\" (UID: \"717bf496-3a04-4777-82fc-97a6072b9c7a\") " pod="openshift-image-registry/image-registry-66df7c8f76-sgr2r" Oct 02 01:54:54 crc kubenswrapper[4885]: I1002 01:54:54.920578 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/717bf496-3a04-4777-82fc-97a6072b9c7a-bound-sa-token\") pod \"image-registry-66df7c8f76-sgr2r\" (UID: \"717bf496-3a04-4777-82fc-97a6072b9c7a\") " pod="openshift-image-registry/image-registry-66df7c8f76-sgr2r" Oct 02 01:54:54 crc kubenswrapper[4885]: I1002 01:54:54.920630 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/717bf496-3a04-4777-82fc-97a6072b9c7a-ca-trust-extracted\") pod \"image-registry-66df7c8f76-sgr2r\" (UID: \"717bf496-3a04-4777-82fc-97a6072b9c7a\") " pod="openshift-image-registry/image-registry-66df7c8f76-sgr2r" Oct 02 01:54:54 crc kubenswrapper[4885]: I1002 01:54:54.920665 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/717bf496-3a04-4777-82fc-97a6072b9c7a-registry-certificates\") pod \"image-registry-66df7c8f76-sgr2r\" (UID: \"717bf496-3a04-4777-82fc-97a6072b9c7a\") " pod="openshift-image-registry/image-registry-66df7c8f76-sgr2r" Oct 02 01:54:54 crc kubenswrapper[4885]: I1002 01:54:54.920720 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/717bf496-3a04-4777-82fc-97a6072b9c7a-registry-tls\") pod \"image-registry-66df7c8f76-sgr2r\" (UID: \"717bf496-3a04-4777-82fc-97a6072b9c7a\") " pod="openshift-image-registry/image-registry-66df7c8f76-sgr2r" Oct 02 01:54:54 crc kubenswrapper[4885]: I1002 01:54:54.921810 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/717bf496-3a04-4777-82fc-97a6072b9c7a-ca-trust-extracted\") pod \"image-registry-66df7c8f76-sgr2r\" (UID: \"717bf496-3a04-4777-82fc-97a6072b9c7a\") " pod="openshift-image-registry/image-registry-66df7c8f76-sgr2r" Oct 02 01:54:54 crc kubenswrapper[4885]: I1002 01:54:54.922835 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/717bf496-3a04-4777-82fc-97a6072b9c7a-trusted-ca\") pod \"image-registry-66df7c8f76-sgr2r\" (UID: \"717bf496-3a04-4777-82fc-97a6072b9c7a\") " pod="openshift-image-registry/image-registry-66df7c8f76-sgr2r" Oct 02 01:54:54 crc kubenswrapper[4885]: I1002 01:54:54.922980 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/717bf496-3a04-4777-82fc-97a6072b9c7a-registry-certificates\") pod \"image-registry-66df7c8f76-sgr2r\" (UID: \"717bf496-3a04-4777-82fc-97a6072b9c7a\") " pod="openshift-image-registry/image-registry-66df7c8f76-sgr2r" Oct 02 01:54:54 crc kubenswrapper[4885]: I1002 01:54:54.930982 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/717bf496-3a04-4777-82fc-97a6072b9c7a-registry-tls\") pod \"image-registry-66df7c8f76-sgr2r\" (UID: \"717bf496-3a04-4777-82fc-97a6072b9c7a\") " pod="openshift-image-registry/image-registry-66df7c8f76-sgr2r" Oct 02 01:54:54 crc kubenswrapper[4885]: I1002 01:54:54.931005 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/717bf496-3a04-4777-82fc-97a6072b9c7a-installation-pull-secrets\") pod \"image-registry-66df7c8f76-sgr2r\" (UID: \"717bf496-3a04-4777-82fc-97a6072b9c7a\") " pod="openshift-image-registry/image-registry-66df7c8f76-sgr2r" Oct 02 01:54:54 crc kubenswrapper[4885]: I1002 01:54:54.949375 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/717bf496-3a04-4777-82fc-97a6072b9c7a-bound-sa-token\") pod \"image-registry-66df7c8f76-sgr2r\" (UID: \"717bf496-3a04-4777-82fc-97a6072b9c7a\") " pod="openshift-image-registry/image-registry-66df7c8f76-sgr2r" Oct 02 01:54:54 crc kubenswrapper[4885]: I1002 01:54:54.952716 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vwc9m\" (UniqueName: \"kubernetes.io/projected/717bf496-3a04-4777-82fc-97a6072b9c7a-kube-api-access-vwc9m\") pod \"image-registry-66df7c8f76-sgr2r\" (UID: \"717bf496-3a04-4777-82fc-97a6072b9c7a\") " pod="openshift-image-registry/image-registry-66df7c8f76-sgr2r" Oct 02 01:54:55 crc kubenswrapper[4885]: I1002 01:54:55.235794 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-sgr2r" Oct 02 01:54:55 crc kubenswrapper[4885]: I1002 01:54:55.508589 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-sgr2r"] Oct 02 01:54:56 crc kubenswrapper[4885]: I1002 01:54:56.344315 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-sgr2r" event={"ID":"717bf496-3a04-4777-82fc-97a6072b9c7a","Type":"ContainerStarted","Data":"999f23132236677961544637cb7919d594671f2d81f5bbcd313245f5effcc554"} Oct 02 01:54:56 crc kubenswrapper[4885]: I1002 01:54:56.344740 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-sgr2r" Oct 02 01:54:56 crc kubenswrapper[4885]: I1002 01:54:56.344762 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-sgr2r" event={"ID":"717bf496-3a04-4777-82fc-97a6072b9c7a","Type":"ContainerStarted","Data":"3ba9a942b550f6a26e9dabf312872a245e52554424eac6dc93866e88d4138d04"} Oct 02 01:54:56 crc kubenswrapper[4885]: I1002 01:54:56.370893 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-sgr2r" podStartSLOduration=2.370871945 podStartE2EDuration="2.370871945s" podCreationTimestamp="2025-10-02 01:54:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:54:56.370125602 +0000 UTC m=+485.181873031" watchObservedRunningTime="2025-10-02 01:54:56.370871945 +0000 UTC m=+485.182619374" Oct 02 01:55:15 crc kubenswrapper[4885]: I1002 01:55:15.245024 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-sgr2r" Oct 02 01:55:15 crc kubenswrapper[4885]: I1002 01:55:15.317660 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-gdxhh"] Oct 02 01:55:40 crc kubenswrapper[4885]: I1002 01:55:40.400192 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-gdxhh" podUID="35380309-be1e-4d61-a2f2-e40a09ff4814" containerName="registry" containerID="cri-o://bc1b9e337176117271850d975cd2912508089832ab08adbeed14222591ad3db4" gracePeriod=30 Oct 02 01:55:40 crc kubenswrapper[4885]: I1002 01:55:40.628682 4885 generic.go:334] "Generic (PLEG): container finished" podID="35380309-be1e-4d61-a2f2-e40a09ff4814" containerID="bc1b9e337176117271850d975cd2912508089832ab08adbeed14222591ad3db4" exitCode=0 Oct 02 01:55:40 crc kubenswrapper[4885]: I1002 01:55:40.628732 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-gdxhh" event={"ID":"35380309-be1e-4d61-a2f2-e40a09ff4814","Type":"ContainerDied","Data":"bc1b9e337176117271850d975cd2912508089832ab08adbeed14222591ad3db4"} Oct 02 01:55:40 crc kubenswrapper[4885]: I1002 01:55:40.901717 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-gdxhh" Oct 02 01:55:41 crc kubenswrapper[4885]: I1002 01:55:41.056813 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/35380309-be1e-4d61-a2f2-e40a09ff4814-installation-pull-secrets\") pod \"35380309-be1e-4d61-a2f2-e40a09ff4814\" (UID: \"35380309-be1e-4d61-a2f2-e40a09ff4814\") " Oct 02 01:55:41 crc kubenswrapper[4885]: I1002 01:55:41.057006 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/35380309-be1e-4d61-a2f2-e40a09ff4814-ca-trust-extracted\") pod \"35380309-be1e-4d61-a2f2-e40a09ff4814\" (UID: \"35380309-be1e-4d61-a2f2-e40a09ff4814\") " Oct 02 01:55:41 crc kubenswrapper[4885]: I1002 01:55:41.057068 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/35380309-be1e-4d61-a2f2-e40a09ff4814-registry-tls\") pod \"35380309-be1e-4d61-a2f2-e40a09ff4814\" (UID: \"35380309-be1e-4d61-a2f2-e40a09ff4814\") " Oct 02 01:55:41 crc kubenswrapper[4885]: I1002 01:55:41.057125 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/35380309-be1e-4d61-a2f2-e40a09ff4814-bound-sa-token\") pod \"35380309-be1e-4d61-a2f2-e40a09ff4814\" (UID: \"35380309-be1e-4d61-a2f2-e40a09ff4814\") " Oct 02 01:55:41 crc kubenswrapper[4885]: I1002 01:55:41.057385 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"35380309-be1e-4d61-a2f2-e40a09ff4814\" (UID: \"35380309-be1e-4d61-a2f2-e40a09ff4814\") " Oct 02 01:55:41 crc kubenswrapper[4885]: I1002 01:55:41.057460 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/35380309-be1e-4d61-a2f2-e40a09ff4814-trusted-ca\") pod \"35380309-be1e-4d61-a2f2-e40a09ff4814\" (UID: \"35380309-be1e-4d61-a2f2-e40a09ff4814\") " Oct 02 01:55:41 crc kubenswrapper[4885]: I1002 01:55:41.057536 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/35380309-be1e-4d61-a2f2-e40a09ff4814-registry-certificates\") pod \"35380309-be1e-4d61-a2f2-e40a09ff4814\" (UID: \"35380309-be1e-4d61-a2f2-e40a09ff4814\") " Oct 02 01:55:41 crc kubenswrapper[4885]: I1002 01:55:41.057609 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-98nfr\" (UniqueName: \"kubernetes.io/projected/35380309-be1e-4d61-a2f2-e40a09ff4814-kube-api-access-98nfr\") pod \"35380309-be1e-4d61-a2f2-e40a09ff4814\" (UID: \"35380309-be1e-4d61-a2f2-e40a09ff4814\") " Oct 02 01:55:41 crc kubenswrapper[4885]: I1002 01:55:41.058889 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/35380309-be1e-4d61-a2f2-e40a09ff4814-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "35380309-be1e-4d61-a2f2-e40a09ff4814" (UID: "35380309-be1e-4d61-a2f2-e40a09ff4814"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:55:41 crc kubenswrapper[4885]: I1002 01:55:41.058994 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/35380309-be1e-4d61-a2f2-e40a09ff4814-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "35380309-be1e-4d61-a2f2-e40a09ff4814" (UID: "35380309-be1e-4d61-a2f2-e40a09ff4814"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:55:41 crc kubenswrapper[4885]: I1002 01:55:41.066421 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35380309-be1e-4d61-a2f2-e40a09ff4814-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "35380309-be1e-4d61-a2f2-e40a09ff4814" (UID: "35380309-be1e-4d61-a2f2-e40a09ff4814"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:55:41 crc kubenswrapper[4885]: I1002 01:55:41.067016 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35380309-be1e-4d61-a2f2-e40a09ff4814-kube-api-access-98nfr" (OuterVolumeSpecName: "kube-api-access-98nfr") pod "35380309-be1e-4d61-a2f2-e40a09ff4814" (UID: "35380309-be1e-4d61-a2f2-e40a09ff4814"). InnerVolumeSpecName "kube-api-access-98nfr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:55:41 crc kubenswrapper[4885]: I1002 01:55:41.067321 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35380309-be1e-4d61-a2f2-e40a09ff4814-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "35380309-be1e-4d61-a2f2-e40a09ff4814" (UID: "35380309-be1e-4d61-a2f2-e40a09ff4814"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:55:41 crc kubenswrapper[4885]: I1002 01:55:41.069802 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35380309-be1e-4d61-a2f2-e40a09ff4814-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "35380309-be1e-4d61-a2f2-e40a09ff4814" (UID: "35380309-be1e-4d61-a2f2-e40a09ff4814"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:55:41 crc kubenswrapper[4885]: I1002 01:55:41.075836 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "35380309-be1e-4d61-a2f2-e40a09ff4814" (UID: "35380309-be1e-4d61-a2f2-e40a09ff4814"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 02 01:55:41 crc kubenswrapper[4885]: I1002 01:55:41.089342 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/35380309-be1e-4d61-a2f2-e40a09ff4814-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "35380309-be1e-4d61-a2f2-e40a09ff4814" (UID: "35380309-be1e-4d61-a2f2-e40a09ff4814"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 01:55:41 crc kubenswrapper[4885]: I1002 01:55:41.159888 4885 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/35380309-be1e-4d61-a2f2-e40a09ff4814-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 02 01:55:41 crc kubenswrapper[4885]: I1002 01:55:41.159998 4885 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/35380309-be1e-4d61-a2f2-e40a09ff4814-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 02 01:55:41 crc kubenswrapper[4885]: I1002 01:55:41.160030 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-98nfr\" (UniqueName: \"kubernetes.io/projected/35380309-be1e-4d61-a2f2-e40a09ff4814-kube-api-access-98nfr\") on node \"crc\" DevicePath \"\"" Oct 02 01:55:41 crc kubenswrapper[4885]: I1002 01:55:41.160057 4885 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/35380309-be1e-4d61-a2f2-e40a09ff4814-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 02 01:55:41 crc kubenswrapper[4885]: I1002 01:55:41.160082 4885 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/35380309-be1e-4d61-a2f2-e40a09ff4814-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 02 01:55:41 crc kubenswrapper[4885]: I1002 01:55:41.160178 4885 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/35380309-be1e-4d61-a2f2-e40a09ff4814-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 02 01:55:41 crc kubenswrapper[4885]: I1002 01:55:41.160209 4885 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/35380309-be1e-4d61-a2f2-e40a09ff4814-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 02 01:55:41 crc kubenswrapper[4885]: I1002 01:55:41.639810 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-gdxhh" event={"ID":"35380309-be1e-4d61-a2f2-e40a09ff4814","Type":"ContainerDied","Data":"3daaff641431a9adaf830a1a9cecd02cac7be521fe571b5b42f7b7dc957b2865"} Oct 02 01:55:41 crc kubenswrapper[4885]: I1002 01:55:41.639881 4885 scope.go:117] "RemoveContainer" containerID="bc1b9e337176117271850d975cd2912508089832ab08adbeed14222591ad3db4" Oct 02 01:55:41 crc kubenswrapper[4885]: I1002 01:55:41.639895 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-gdxhh" Oct 02 01:55:41 crc kubenswrapper[4885]: I1002 01:55:41.703921 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-gdxhh"] Oct 02 01:55:41 crc kubenswrapper[4885]: I1002 01:55:41.711036 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-gdxhh"] Oct 02 01:55:42 crc kubenswrapper[4885]: I1002 01:55:42.061048 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="35380309-be1e-4d61-a2f2-e40a09ff4814" path="/var/lib/kubelet/pods/35380309-be1e-4d61-a2f2-e40a09ff4814/volumes" Oct 02 01:55:43 crc kubenswrapper[4885]: I1002 01:55:43.266203 4885 patch_prober.go:28] interesting pod/machine-config-daemon-rttx8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 01:55:43 crc kubenswrapper[4885]: I1002 01:55:43.267518 4885 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 01:55:45 crc kubenswrapper[4885]: I1002 01:55:45.708048 4885 patch_prober.go:28] interesting pod/image-registry-697d97f7c8-gdxhh container/registry namespace/openshift-image-registry: Readiness probe status=failure output="Get \"https://10.217.0.26:5000/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 02 01:55:45 crc kubenswrapper[4885]: I1002 01:55:45.708142 4885 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-image-registry/image-registry-697d97f7c8-gdxhh" podUID="35380309-be1e-4d61-a2f2-e40a09ff4814" containerName="registry" probeResult="failure" output="Get \"https://10.217.0.26:5000/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 02 01:56:13 crc kubenswrapper[4885]: I1002 01:56:13.266578 4885 patch_prober.go:28] interesting pod/machine-config-daemon-rttx8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 01:56:13 crc kubenswrapper[4885]: I1002 01:56:13.267573 4885 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 01:56:28 crc kubenswrapper[4885]: I1002 01:56:28.459948 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-rbccc"] Oct 02 01:56:28 crc kubenswrapper[4885]: E1002 01:56:28.460830 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35380309-be1e-4d61-a2f2-e40a09ff4814" containerName="registry" Oct 02 01:56:28 crc kubenswrapper[4885]: I1002 01:56:28.460853 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="35380309-be1e-4d61-a2f2-e40a09ff4814" containerName="registry" Oct 02 01:56:28 crc kubenswrapper[4885]: I1002 01:56:28.461047 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="35380309-be1e-4d61-a2f2-e40a09ff4814" containerName="registry" Oct 02 01:56:28 crc kubenswrapper[4885]: I1002 01:56:28.461657 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-rbccc" Oct 02 01:56:28 crc kubenswrapper[4885]: I1002 01:56:28.467067 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Oct 02 01:56:28 crc kubenswrapper[4885]: I1002 01:56:28.467412 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Oct 02 01:56:28 crc kubenswrapper[4885]: I1002 01:56:28.467132 4885 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-tkdmg" Oct 02 01:56:28 crc kubenswrapper[4885]: I1002 01:56:28.480579 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-pj866"] Oct 02 01:56:28 crc kubenswrapper[4885]: I1002 01:56:28.481209 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-pj866" Oct 02 01:56:28 crc kubenswrapper[4885]: W1002 01:56:28.483557 4885 reflector.go:561] object-"cert-manager"/"cert-manager-dockercfg-gr9bj": failed to list *v1.Secret: secrets "cert-manager-dockercfg-gr9bj" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "cert-manager": no relationship found between node 'crc' and this object Oct 02 01:56:28 crc kubenswrapper[4885]: E1002 01:56:28.483637 4885 reflector.go:158] "Unhandled Error" err="object-\"cert-manager\"/\"cert-manager-dockercfg-gr9bj\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"cert-manager-dockercfg-gr9bj\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"cert-manager\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 02 01:56:28 crc kubenswrapper[4885]: I1002 01:56:28.484683 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-rbccc"] Oct 02 01:56:28 crc kubenswrapper[4885]: I1002 01:56:28.489207 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-vh2n2"] Oct 02 01:56:28 crc kubenswrapper[4885]: I1002 01:56:28.490279 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-vh2n2" Oct 02 01:56:28 crc kubenswrapper[4885]: I1002 01:56:28.492149 4885 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-wsbgg" Oct 02 01:56:28 crc kubenswrapper[4885]: I1002 01:56:28.502319 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-pj866"] Oct 02 01:56:28 crc kubenswrapper[4885]: I1002 01:56:28.505240 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-vh2n2"] Oct 02 01:56:28 crc kubenswrapper[4885]: I1002 01:56:28.585653 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cmp9s\" (UniqueName: \"kubernetes.io/projected/0b6eafa9-42ce-4e5b-9ded-180a669d26e2-kube-api-access-cmp9s\") pod \"cert-manager-5b446d88c5-pj866\" (UID: \"0b6eafa9-42ce-4e5b-9ded-180a669d26e2\") " pod="cert-manager/cert-manager-5b446d88c5-pj866" Oct 02 01:56:28 crc kubenswrapper[4885]: I1002 01:56:28.585708 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9qf5f\" (UniqueName: \"kubernetes.io/projected/e50cfe49-c47f-4f2a-9799-0c79123fac70-kube-api-access-9qf5f\") pod \"cert-manager-cainjector-7f985d654d-rbccc\" (UID: \"e50cfe49-c47f-4f2a-9799-0c79123fac70\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-rbccc" Oct 02 01:56:28 crc kubenswrapper[4885]: I1002 01:56:28.687475 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cmp9s\" (UniqueName: \"kubernetes.io/projected/0b6eafa9-42ce-4e5b-9ded-180a669d26e2-kube-api-access-cmp9s\") pod \"cert-manager-5b446d88c5-pj866\" (UID: \"0b6eafa9-42ce-4e5b-9ded-180a669d26e2\") " pod="cert-manager/cert-manager-5b446d88c5-pj866" Oct 02 01:56:28 crc kubenswrapper[4885]: I1002 01:56:28.687552 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9qf5f\" (UniqueName: \"kubernetes.io/projected/e50cfe49-c47f-4f2a-9799-0c79123fac70-kube-api-access-9qf5f\") pod \"cert-manager-cainjector-7f985d654d-rbccc\" (UID: \"e50cfe49-c47f-4f2a-9799-0c79123fac70\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-rbccc" Oct 02 01:56:28 crc kubenswrapper[4885]: I1002 01:56:28.687613 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jw84c\" (UniqueName: \"kubernetes.io/projected/18c66206-ad33-47d2-a37a-b3b50c2bf984-kube-api-access-jw84c\") pod \"cert-manager-webhook-5655c58dd6-vh2n2\" (UID: \"18c66206-ad33-47d2-a37a-b3b50c2bf984\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-vh2n2" Oct 02 01:56:28 crc kubenswrapper[4885]: I1002 01:56:28.715867 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9qf5f\" (UniqueName: \"kubernetes.io/projected/e50cfe49-c47f-4f2a-9799-0c79123fac70-kube-api-access-9qf5f\") pod \"cert-manager-cainjector-7f985d654d-rbccc\" (UID: \"e50cfe49-c47f-4f2a-9799-0c79123fac70\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-rbccc" Oct 02 01:56:28 crc kubenswrapper[4885]: I1002 01:56:28.722626 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cmp9s\" (UniqueName: \"kubernetes.io/projected/0b6eafa9-42ce-4e5b-9ded-180a669d26e2-kube-api-access-cmp9s\") pod \"cert-manager-5b446d88c5-pj866\" (UID: \"0b6eafa9-42ce-4e5b-9ded-180a669d26e2\") " pod="cert-manager/cert-manager-5b446d88c5-pj866" Oct 02 01:56:28 crc kubenswrapper[4885]: I1002 01:56:28.788662 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jw84c\" (UniqueName: \"kubernetes.io/projected/18c66206-ad33-47d2-a37a-b3b50c2bf984-kube-api-access-jw84c\") pod \"cert-manager-webhook-5655c58dd6-vh2n2\" (UID: \"18c66206-ad33-47d2-a37a-b3b50c2bf984\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-vh2n2" Oct 02 01:56:28 crc kubenswrapper[4885]: I1002 01:56:28.791416 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-rbccc" Oct 02 01:56:28 crc kubenswrapper[4885]: I1002 01:56:28.810403 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jw84c\" (UniqueName: \"kubernetes.io/projected/18c66206-ad33-47d2-a37a-b3b50c2bf984-kube-api-access-jw84c\") pod \"cert-manager-webhook-5655c58dd6-vh2n2\" (UID: \"18c66206-ad33-47d2-a37a-b3b50c2bf984\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-vh2n2" Oct 02 01:56:28 crc kubenswrapper[4885]: I1002 01:56:28.816880 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-vh2n2" Oct 02 01:56:29 crc kubenswrapper[4885]: I1002 01:56:29.029959 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-vh2n2"] Oct 02 01:56:29 crc kubenswrapper[4885]: I1002 01:56:29.048365 4885 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 01:56:29 crc kubenswrapper[4885]: I1002 01:56:29.073717 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-rbccc"] Oct 02 01:56:29 crc kubenswrapper[4885]: W1002 01:56:29.084595 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode50cfe49_c47f_4f2a_9799_0c79123fac70.slice/crio-2f48a6d0daae453f166b6c79452470379f87fea43b99437f7b72b065b21a8506 WatchSource:0}: Error finding container 2f48a6d0daae453f166b6c79452470379f87fea43b99437f7b72b065b21a8506: Status 404 returned error can't find the container with id 2f48a6d0daae453f166b6c79452470379f87fea43b99437f7b72b065b21a8506 Oct 02 01:56:29 crc kubenswrapper[4885]: I1002 01:56:29.435427 4885 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-gr9bj" Oct 02 01:56:29 crc kubenswrapper[4885]: I1002 01:56:29.445151 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-pj866" Oct 02 01:56:29 crc kubenswrapper[4885]: I1002 01:56:29.742369 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-pj866"] Oct 02 01:56:29 crc kubenswrapper[4885]: I1002 01:56:29.963878 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-vh2n2" event={"ID":"18c66206-ad33-47d2-a37a-b3b50c2bf984","Type":"ContainerStarted","Data":"1dc5b2550f79b4e957b6d735594ec577772a861d4759ea279af35a046886e252"} Oct 02 01:56:29 crc kubenswrapper[4885]: I1002 01:56:29.965160 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-rbccc" event={"ID":"e50cfe49-c47f-4f2a-9799-0c79123fac70","Type":"ContainerStarted","Data":"2f48a6d0daae453f166b6c79452470379f87fea43b99437f7b72b065b21a8506"} Oct 02 01:56:29 crc kubenswrapper[4885]: I1002 01:56:29.966932 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-pj866" event={"ID":"0b6eafa9-42ce-4e5b-9ded-180a669d26e2","Type":"ContainerStarted","Data":"a1e3ff22b157ef9ad26ebb27229870da52873c97c450d4143a83f9b2509523c0"} Oct 02 01:56:37 crc kubenswrapper[4885]: I1002 01:56:37.010700 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-rbccc" event={"ID":"e50cfe49-c47f-4f2a-9799-0c79123fac70","Type":"ContainerStarted","Data":"ee26b6d49940eca1c1eefcaeba25f7adf7637422de272830394184c8f02cfc21"} Oct 02 01:56:37 crc kubenswrapper[4885]: I1002 01:56:37.015791 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-vh2n2" event={"ID":"18c66206-ad33-47d2-a37a-b3b50c2bf984","Type":"ContainerStarted","Data":"b7177510a58f5efb8d90544ea984365c86aa2e58ed4589b17375507ee5cd0bf1"} Oct 02 01:56:37 crc kubenswrapper[4885]: I1002 01:56:37.016015 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-vh2n2" Oct 02 01:56:37 crc kubenswrapper[4885]: I1002 01:56:37.033613 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-rbccc" podStartSLOduration=2.329492583 podStartE2EDuration="9.033589806s" podCreationTimestamp="2025-10-02 01:56:28 +0000 UTC" firstStartedPulling="2025-10-02 01:56:29.08957835 +0000 UTC m=+577.901325769" lastFinishedPulling="2025-10-02 01:56:35.793675563 +0000 UTC m=+584.605422992" observedRunningTime="2025-10-02 01:56:37.030600705 +0000 UTC m=+585.842348134" watchObservedRunningTime="2025-10-02 01:56:37.033589806 +0000 UTC m=+585.845337235" Oct 02 01:56:37 crc kubenswrapper[4885]: I1002 01:56:37.057466 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-vh2n2" podStartSLOduration=2.318578271 podStartE2EDuration="9.057440647s" podCreationTimestamp="2025-10-02 01:56:28 +0000 UTC" firstStartedPulling="2025-10-02 01:56:29.04806102 +0000 UTC m=+577.859808429" lastFinishedPulling="2025-10-02 01:56:35.786923376 +0000 UTC m=+584.598670805" observedRunningTime="2025-10-02 01:56:37.054247879 +0000 UTC m=+585.865995308" watchObservedRunningTime="2025-10-02 01:56:37.057440647 +0000 UTC m=+585.869188086" Oct 02 01:56:39 crc kubenswrapper[4885]: I1002 01:56:39.017695 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-dngcm"] Oct 02 01:56:39 crc kubenswrapper[4885]: I1002 01:56:39.018707 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" podUID="9e4679dd-f870-41e9-a32b-360f5d2ee81b" containerName="ovn-acl-logging" containerID="cri-o://ef403e519d197fb37b3e6b76a14a5128ba3f30a7a5e8bf0a420cf6a39b89575f" gracePeriod=30 Oct 02 01:56:39 crc kubenswrapper[4885]: I1002 01:56:39.018680 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" podUID="9e4679dd-f870-41e9-a32b-360f5d2ee81b" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://a7642033b0b5995d332ccf7992d6d7e6a5e9c9369ae2d1071ccb2f536938c8f4" gracePeriod=30 Oct 02 01:56:39 crc kubenswrapper[4885]: I1002 01:56:39.018694 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" podUID="9e4679dd-f870-41e9-a32b-360f5d2ee81b" containerName="kube-rbac-proxy-node" containerID="cri-o://4310909ad9c7b609ed6814e04a525a6afae747775598e7275080413dec8b23a3" gracePeriod=30 Oct 02 01:56:39 crc kubenswrapper[4885]: I1002 01:56:39.018895 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" podUID="9e4679dd-f870-41e9-a32b-360f5d2ee81b" containerName="nbdb" containerID="cri-o://c9f99f9e26a04fb42b01a0783a5c4e360514818ce9102edc7fa08626d50b9a34" gracePeriod=30 Oct 02 01:56:39 crc kubenswrapper[4885]: I1002 01:56:39.018969 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" podUID="9e4679dd-f870-41e9-a32b-360f5d2ee81b" containerName="northd" containerID="cri-o://a6fa72b030d309656b73121e4ac2a3924a4e449aad66683dd9de5ee5f22d5a4e" gracePeriod=30 Oct 02 01:56:39 crc kubenswrapper[4885]: I1002 01:56:39.019049 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" podUID="9e4679dd-f870-41e9-a32b-360f5d2ee81b" containerName="sbdb" containerID="cri-o://fafdd1094bd02786d745c984989852b0c690db38655f92b419733a2004f417c0" gracePeriod=30 Oct 02 01:56:39 crc kubenswrapper[4885]: I1002 01:56:39.018618 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" podUID="9e4679dd-f870-41e9-a32b-360f5d2ee81b" containerName="ovn-controller" containerID="cri-o://8482ba457683ffa03100ec1f07ea9d3f918be5e490b5ba05c2ba965d03033191" gracePeriod=30 Oct 02 01:56:39 crc kubenswrapper[4885]: I1002 01:56:39.042696 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-pj866" event={"ID":"0b6eafa9-42ce-4e5b-9ded-180a669d26e2","Type":"ContainerStarted","Data":"615692b031b52eabd11b31f86ead09edd5df947e783cef1da0463c36de2644db"} Oct 02 01:56:39 crc kubenswrapper[4885]: I1002 01:56:39.072747 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" podUID="9e4679dd-f870-41e9-a32b-360f5d2ee81b" containerName="ovnkube-controller" containerID="cri-o://41121c480bb0f68a539497e7d1018e6582a1cf812f819b4c7327ad8827b316b7" gracePeriod=30 Oct 02 01:56:39 crc kubenswrapper[4885]: I1002 01:56:39.076337 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-pj866" podStartSLOduration=2.235876814 podStartE2EDuration="11.076308053s" podCreationTimestamp="2025-10-02 01:56:28 +0000 UTC" firstStartedPulling="2025-10-02 01:56:29.756010379 +0000 UTC m=+578.567757778" lastFinishedPulling="2025-10-02 01:56:38.596441578 +0000 UTC m=+587.408189017" observedRunningTime="2025-10-02 01:56:39.073378603 +0000 UTC m=+587.885126012" watchObservedRunningTime="2025-10-02 01:56:39.076308053 +0000 UTC m=+587.888055492" Oct 02 01:56:39 crc kubenswrapper[4885]: I1002 01:56:39.921803 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dngcm_9e4679dd-f870-41e9-a32b-360f5d2ee81b/ovnkube-controller/3.log" Oct 02 01:56:39 crc kubenswrapper[4885]: I1002 01:56:39.925460 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dngcm_9e4679dd-f870-41e9-a32b-360f5d2ee81b/ovn-acl-logging/0.log" Oct 02 01:56:39 crc kubenswrapper[4885]: I1002 01:56:39.926297 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dngcm_9e4679dd-f870-41e9-a32b-360f5d2ee81b/ovn-controller/0.log" Oct 02 01:56:39 crc kubenswrapper[4885]: I1002 01:56:39.926901 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" Oct 02 01:56:39 crc kubenswrapper[4885]: I1002 01:56:39.996809 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-x69mt"] Oct 02 01:56:39 crc kubenswrapper[4885]: E1002 01:56:39.997070 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e4679dd-f870-41e9-a32b-360f5d2ee81b" containerName="kubecfg-setup" Oct 02 01:56:39 crc kubenswrapper[4885]: I1002 01:56:39.997089 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e4679dd-f870-41e9-a32b-360f5d2ee81b" containerName="kubecfg-setup" Oct 02 01:56:39 crc kubenswrapper[4885]: E1002 01:56:39.997104 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e4679dd-f870-41e9-a32b-360f5d2ee81b" containerName="ovnkube-controller" Oct 02 01:56:39 crc kubenswrapper[4885]: I1002 01:56:39.997114 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e4679dd-f870-41e9-a32b-360f5d2ee81b" containerName="ovnkube-controller" Oct 02 01:56:39 crc kubenswrapper[4885]: E1002 01:56:39.997132 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e4679dd-f870-41e9-a32b-360f5d2ee81b" containerName="nbdb" Oct 02 01:56:39 crc kubenswrapper[4885]: I1002 01:56:39.997142 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e4679dd-f870-41e9-a32b-360f5d2ee81b" containerName="nbdb" Oct 02 01:56:39 crc kubenswrapper[4885]: E1002 01:56:39.997163 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e4679dd-f870-41e9-a32b-360f5d2ee81b" containerName="ovn-acl-logging" Oct 02 01:56:39 crc kubenswrapper[4885]: I1002 01:56:39.997174 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e4679dd-f870-41e9-a32b-360f5d2ee81b" containerName="ovn-acl-logging" Oct 02 01:56:39 crc kubenswrapper[4885]: E1002 01:56:39.997191 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e4679dd-f870-41e9-a32b-360f5d2ee81b" containerName="ovnkube-controller" Oct 02 01:56:39 crc kubenswrapper[4885]: I1002 01:56:39.997201 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e4679dd-f870-41e9-a32b-360f5d2ee81b" containerName="ovnkube-controller" Oct 02 01:56:39 crc kubenswrapper[4885]: E1002 01:56:39.997215 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e4679dd-f870-41e9-a32b-360f5d2ee81b" containerName="ovnkube-controller" Oct 02 01:56:39 crc kubenswrapper[4885]: I1002 01:56:39.997226 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e4679dd-f870-41e9-a32b-360f5d2ee81b" containerName="ovnkube-controller" Oct 02 01:56:39 crc kubenswrapper[4885]: E1002 01:56:39.997238 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e4679dd-f870-41e9-a32b-360f5d2ee81b" containerName="kube-rbac-proxy-node" Oct 02 01:56:39 crc kubenswrapper[4885]: I1002 01:56:39.997248 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e4679dd-f870-41e9-a32b-360f5d2ee81b" containerName="kube-rbac-proxy-node" Oct 02 01:56:39 crc kubenswrapper[4885]: E1002 01:56:39.997267 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e4679dd-f870-41e9-a32b-360f5d2ee81b" containerName="ovn-controller" Oct 02 01:56:39 crc kubenswrapper[4885]: I1002 01:56:39.997297 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e4679dd-f870-41e9-a32b-360f5d2ee81b" containerName="ovn-controller" Oct 02 01:56:39 crc kubenswrapper[4885]: E1002 01:56:39.997314 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e4679dd-f870-41e9-a32b-360f5d2ee81b" containerName="kube-rbac-proxy-ovn-metrics" Oct 02 01:56:39 crc kubenswrapper[4885]: I1002 01:56:39.997324 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e4679dd-f870-41e9-a32b-360f5d2ee81b" containerName="kube-rbac-proxy-ovn-metrics" Oct 02 01:56:39 crc kubenswrapper[4885]: E1002 01:56:39.997336 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e4679dd-f870-41e9-a32b-360f5d2ee81b" containerName="ovnkube-controller" Oct 02 01:56:39 crc kubenswrapper[4885]: I1002 01:56:39.997346 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e4679dd-f870-41e9-a32b-360f5d2ee81b" containerName="ovnkube-controller" Oct 02 01:56:39 crc kubenswrapper[4885]: E1002 01:56:39.997358 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e4679dd-f870-41e9-a32b-360f5d2ee81b" containerName="northd" Oct 02 01:56:39 crc kubenswrapper[4885]: I1002 01:56:39.997367 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e4679dd-f870-41e9-a32b-360f5d2ee81b" containerName="northd" Oct 02 01:56:39 crc kubenswrapper[4885]: E1002 01:56:39.997378 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e4679dd-f870-41e9-a32b-360f5d2ee81b" containerName="sbdb" Oct 02 01:56:39 crc kubenswrapper[4885]: I1002 01:56:39.997388 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e4679dd-f870-41e9-a32b-360f5d2ee81b" containerName="sbdb" Oct 02 01:56:39 crc kubenswrapper[4885]: I1002 01:56:39.997551 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e4679dd-f870-41e9-a32b-360f5d2ee81b" containerName="sbdb" Oct 02 01:56:39 crc kubenswrapper[4885]: I1002 01:56:39.997569 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e4679dd-f870-41e9-a32b-360f5d2ee81b" containerName="ovnkube-controller" Oct 02 01:56:39 crc kubenswrapper[4885]: I1002 01:56:39.997580 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e4679dd-f870-41e9-a32b-360f5d2ee81b" containerName="northd" Oct 02 01:56:39 crc kubenswrapper[4885]: I1002 01:56:39.997591 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e4679dd-f870-41e9-a32b-360f5d2ee81b" containerName="kube-rbac-proxy-ovn-metrics" Oct 02 01:56:39 crc kubenswrapper[4885]: I1002 01:56:39.997603 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e4679dd-f870-41e9-a32b-360f5d2ee81b" containerName="ovnkube-controller" Oct 02 01:56:39 crc kubenswrapper[4885]: I1002 01:56:39.997618 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e4679dd-f870-41e9-a32b-360f5d2ee81b" containerName="kube-rbac-proxy-node" Oct 02 01:56:39 crc kubenswrapper[4885]: I1002 01:56:39.997634 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e4679dd-f870-41e9-a32b-360f5d2ee81b" containerName="ovnkube-controller" Oct 02 01:56:39 crc kubenswrapper[4885]: I1002 01:56:39.997653 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e4679dd-f870-41e9-a32b-360f5d2ee81b" containerName="ovn-controller" Oct 02 01:56:39 crc kubenswrapper[4885]: I1002 01:56:39.997667 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e4679dd-f870-41e9-a32b-360f5d2ee81b" containerName="ovn-acl-logging" Oct 02 01:56:39 crc kubenswrapper[4885]: I1002 01:56:39.997682 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e4679dd-f870-41e9-a32b-360f5d2ee81b" containerName="nbdb" Oct 02 01:56:39 crc kubenswrapper[4885]: E1002 01:56:39.997820 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e4679dd-f870-41e9-a32b-360f5d2ee81b" containerName="ovnkube-controller" Oct 02 01:56:39 crc kubenswrapper[4885]: I1002 01:56:39.997858 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e4679dd-f870-41e9-a32b-360f5d2ee81b" containerName="ovnkube-controller" Oct 02 01:56:39 crc kubenswrapper[4885]: I1002 01:56:39.997999 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e4679dd-f870-41e9-a32b-360f5d2ee81b" containerName="ovnkube-controller" Oct 02 01:56:39 crc kubenswrapper[4885]: I1002 01:56:39.998014 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e4679dd-f870-41e9-a32b-360f5d2ee81b" containerName="ovnkube-controller" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.001018 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-x69mt" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.066327 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/9e4679dd-f870-41e9-a32b-360f5d2ee81b-host-cni-netd\") pod \"9e4679dd-f870-41e9-a32b-360f5d2ee81b\" (UID: \"9e4679dd-f870-41e9-a32b-360f5d2ee81b\") " Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.066401 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9e4679dd-f870-41e9-a32b-360f5d2ee81b-etc-openvswitch\") pod \"9e4679dd-f870-41e9-a32b-360f5d2ee81b\" (UID: \"9e4679dd-f870-41e9-a32b-360f5d2ee81b\") " Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.066465 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9e4679dd-f870-41e9-a32b-360f5d2ee81b-env-overrides\") pod \"9e4679dd-f870-41e9-a32b-360f5d2ee81b\" (UID: \"9e4679dd-f870-41e9-a32b-360f5d2ee81b\") " Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.066487 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9e4679dd-f870-41e9-a32b-360f5d2ee81b-host-var-lib-cni-networks-ovn-kubernetes\") pod \"9e4679dd-f870-41e9-a32b-360f5d2ee81b\" (UID: \"9e4679dd-f870-41e9-a32b-360f5d2ee81b\") " Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.066505 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/9e4679dd-f870-41e9-a32b-360f5d2ee81b-log-socket\") pod \"9e4679dd-f870-41e9-a32b-360f5d2ee81b\" (UID: \"9e4679dd-f870-41e9-a32b-360f5d2ee81b\") " Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.066549 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9e4679dd-f870-41e9-a32b-360f5d2ee81b-var-lib-openvswitch\") pod \"9e4679dd-f870-41e9-a32b-360f5d2ee81b\" (UID: \"9e4679dd-f870-41e9-a32b-360f5d2ee81b\") " Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.066587 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4b7jr\" (UniqueName: \"kubernetes.io/projected/9e4679dd-f870-41e9-a32b-360f5d2ee81b-kube-api-access-4b7jr\") pod \"9e4679dd-f870-41e9-a32b-360f5d2ee81b\" (UID: \"9e4679dd-f870-41e9-a32b-360f5d2ee81b\") " Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.066609 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/9e4679dd-f870-41e9-a32b-360f5d2ee81b-host-slash\") pod \"9e4679dd-f870-41e9-a32b-360f5d2ee81b\" (UID: \"9e4679dd-f870-41e9-a32b-360f5d2ee81b\") " Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.066641 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/9e4679dd-f870-41e9-a32b-360f5d2ee81b-ovnkube-script-lib\") pod \"9e4679dd-f870-41e9-a32b-360f5d2ee81b\" (UID: \"9e4679dd-f870-41e9-a32b-360f5d2ee81b\") " Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.066668 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9e4679dd-f870-41e9-a32b-360f5d2ee81b-ovnkube-config\") pod \"9e4679dd-f870-41e9-a32b-360f5d2ee81b\" (UID: \"9e4679dd-f870-41e9-a32b-360f5d2ee81b\") " Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.066726 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9e4679dd-f870-41e9-a32b-360f5d2ee81b-ovn-node-metrics-cert\") pod \"9e4679dd-f870-41e9-a32b-360f5d2ee81b\" (UID: \"9e4679dd-f870-41e9-a32b-360f5d2ee81b\") " Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.066750 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9e4679dd-f870-41e9-a32b-360f5d2ee81b-host-run-ovn-kubernetes\") pod \"9e4679dd-f870-41e9-a32b-360f5d2ee81b\" (UID: \"9e4679dd-f870-41e9-a32b-360f5d2ee81b\") " Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.066772 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/9e4679dd-f870-41e9-a32b-360f5d2ee81b-run-ovn\") pod \"9e4679dd-f870-41e9-a32b-360f5d2ee81b\" (UID: \"9e4679dd-f870-41e9-a32b-360f5d2ee81b\") " Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.066803 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9e4679dd-f870-41e9-a32b-360f5d2ee81b-host-cni-bin\") pod \"9e4679dd-f870-41e9-a32b-360f5d2ee81b\" (UID: \"9e4679dd-f870-41e9-a32b-360f5d2ee81b\") " Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.066833 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/9e4679dd-f870-41e9-a32b-360f5d2ee81b-systemd-units\") pod \"9e4679dd-f870-41e9-a32b-360f5d2ee81b\" (UID: \"9e4679dd-f870-41e9-a32b-360f5d2ee81b\") " Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.066856 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/9e4679dd-f870-41e9-a32b-360f5d2ee81b-run-systemd\") pod \"9e4679dd-f870-41e9-a32b-360f5d2ee81b\" (UID: \"9e4679dd-f870-41e9-a32b-360f5d2ee81b\") " Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.066877 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9e4679dd-f870-41e9-a32b-360f5d2ee81b-run-openvswitch\") pod \"9e4679dd-f870-41e9-a32b-360f5d2ee81b\" (UID: \"9e4679dd-f870-41e9-a32b-360f5d2ee81b\") " Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.066900 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/9e4679dd-f870-41e9-a32b-360f5d2ee81b-node-log\") pod \"9e4679dd-f870-41e9-a32b-360f5d2ee81b\" (UID: \"9e4679dd-f870-41e9-a32b-360f5d2ee81b\") " Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.066925 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/9e4679dd-f870-41e9-a32b-360f5d2ee81b-host-kubelet\") pod \"9e4679dd-f870-41e9-a32b-360f5d2ee81b\" (UID: \"9e4679dd-f870-41e9-a32b-360f5d2ee81b\") " Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.066949 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9e4679dd-f870-41e9-a32b-360f5d2ee81b-host-run-netns\") pod \"9e4679dd-f870-41e9-a32b-360f5d2ee81b\" (UID: \"9e4679dd-f870-41e9-a32b-360f5d2ee81b\") " Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.067174 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9e4679dd-f870-41e9-a32b-360f5d2ee81b-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "9e4679dd-f870-41e9-a32b-360f5d2ee81b" (UID: "9e4679dd-f870-41e9-a32b-360f5d2ee81b"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.067224 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9e4679dd-f870-41e9-a32b-360f5d2ee81b-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "9e4679dd-f870-41e9-a32b-360f5d2ee81b" (UID: "9e4679dd-f870-41e9-a32b-360f5d2ee81b"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.067249 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9e4679dd-f870-41e9-a32b-360f5d2ee81b-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "9e4679dd-f870-41e9-a32b-360f5d2ee81b" (UID: "9e4679dd-f870-41e9-a32b-360f5d2ee81b"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.068468 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e4679dd-f870-41e9-a32b-360f5d2ee81b-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "9e4679dd-f870-41e9-a32b-360f5d2ee81b" (UID: "9e4679dd-f870-41e9-a32b-360f5d2ee81b"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.068550 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9e4679dd-f870-41e9-a32b-360f5d2ee81b-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "9e4679dd-f870-41e9-a32b-360f5d2ee81b" (UID: "9e4679dd-f870-41e9-a32b-360f5d2ee81b"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.068583 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9e4679dd-f870-41e9-a32b-360f5d2ee81b-log-socket" (OuterVolumeSpecName: "log-socket") pod "9e4679dd-f870-41e9-a32b-360f5d2ee81b" (UID: "9e4679dd-f870-41e9-a32b-360f5d2ee81b"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.068606 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9e4679dd-f870-41e9-a32b-360f5d2ee81b-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "9e4679dd-f870-41e9-a32b-360f5d2ee81b" (UID: "9e4679dd-f870-41e9-a32b-360f5d2ee81b"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.069522 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9e4679dd-f870-41e9-a32b-360f5d2ee81b-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "9e4679dd-f870-41e9-a32b-360f5d2ee81b" (UID: "9e4679dd-f870-41e9-a32b-360f5d2ee81b"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.069904 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9e4679dd-f870-41e9-a32b-360f5d2ee81b-node-log" (OuterVolumeSpecName: "node-log") pod "9e4679dd-f870-41e9-a32b-360f5d2ee81b" (UID: "9e4679dd-f870-41e9-a32b-360f5d2ee81b"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.069998 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9e4679dd-f870-41e9-a32b-360f5d2ee81b-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "9e4679dd-f870-41e9-a32b-360f5d2ee81b" (UID: "9e4679dd-f870-41e9-a32b-360f5d2ee81b"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.070050 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9e4679dd-f870-41e9-a32b-360f5d2ee81b-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "9e4679dd-f870-41e9-a32b-360f5d2ee81b" (UID: "9e4679dd-f870-41e9-a32b-360f5d2ee81b"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.070234 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9e4679dd-f870-41e9-a32b-360f5d2ee81b-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "9e4679dd-f870-41e9-a32b-360f5d2ee81b" (UID: "9e4679dd-f870-41e9-a32b-360f5d2ee81b"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.070221 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9e4679dd-f870-41e9-a32b-360f5d2ee81b-host-slash" (OuterVolumeSpecName: "host-slash") pod "9e4679dd-f870-41e9-a32b-360f5d2ee81b" (UID: "9e4679dd-f870-41e9-a32b-360f5d2ee81b"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.070320 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9e4679dd-f870-41e9-a32b-360f5d2ee81b-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "9e4679dd-f870-41e9-a32b-360f5d2ee81b" (UID: "9e4679dd-f870-41e9-a32b-360f5d2ee81b"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.070333 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9e4679dd-f870-41e9-a32b-360f5d2ee81b-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "9e4679dd-f870-41e9-a32b-360f5d2ee81b" (UID: "9e4679dd-f870-41e9-a32b-360f5d2ee81b"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.070321 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e4679dd-f870-41e9-a32b-360f5d2ee81b-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "9e4679dd-f870-41e9-a32b-360f5d2ee81b" (UID: "9e4679dd-f870-41e9-a32b-360f5d2ee81b"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.070772 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e4679dd-f870-41e9-a32b-360f5d2ee81b-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "9e4679dd-f870-41e9-a32b-360f5d2ee81b" (UID: "9e4679dd-f870-41e9-a32b-360f5d2ee81b"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.071552 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-vhgbd_ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f/kube-multus/2.log" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.072348 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-vhgbd_ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f/kube-multus/1.log" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.072403 4885 generic.go:334] "Generic (PLEG): container finished" podID="ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f" containerID="7476bbb9bcb8915493b4060254e1303bd96e825ec40425de522dce0a357fb6e6" exitCode=2 Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.072407 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-vhgbd" event={"ID":"ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f","Type":"ContainerDied","Data":"7476bbb9bcb8915493b4060254e1303bd96e825ec40425de522dce0a357fb6e6"} Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.072541 4885 scope.go:117] "RemoveContainer" containerID="df9792d80c807ed98c8442822d01f619327361c40425abfe26f11e5c27ba2584" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.072888 4885 scope.go:117] "RemoveContainer" containerID="7476bbb9bcb8915493b4060254e1303bd96e825ec40425de522dce0a357fb6e6" Oct 02 01:56:40 crc kubenswrapper[4885]: E1002 01:56:40.073098 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-vhgbd_openshift-multus(ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f)\"" pod="openshift-multus/multus-vhgbd" podUID="ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.097334 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dngcm_9e4679dd-f870-41e9-a32b-360f5d2ee81b/ovnkube-controller/3.log" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.097549 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e4679dd-f870-41e9-a32b-360f5d2ee81b-kube-api-access-4b7jr" (OuterVolumeSpecName: "kube-api-access-4b7jr") pod "9e4679dd-f870-41e9-a32b-360f5d2ee81b" (UID: "9e4679dd-f870-41e9-a32b-360f5d2ee81b"). InnerVolumeSpecName "kube-api-access-4b7jr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.097941 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e4679dd-f870-41e9-a32b-360f5d2ee81b-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "9e4679dd-f870-41e9-a32b-360f5d2ee81b" (UID: "9e4679dd-f870-41e9-a32b-360f5d2ee81b"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.099038 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9e4679dd-f870-41e9-a32b-360f5d2ee81b-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "9e4679dd-f870-41e9-a32b-360f5d2ee81b" (UID: "9e4679dd-f870-41e9-a32b-360f5d2ee81b"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.105696 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dngcm_9e4679dd-f870-41e9-a32b-360f5d2ee81b/ovn-acl-logging/0.log" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.106225 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dngcm_9e4679dd-f870-41e9-a32b-360f5d2ee81b/ovn-controller/0.log" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.106710 4885 generic.go:334] "Generic (PLEG): container finished" podID="9e4679dd-f870-41e9-a32b-360f5d2ee81b" containerID="41121c480bb0f68a539497e7d1018e6582a1cf812f819b4c7327ad8827b316b7" exitCode=0 Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.106737 4885 generic.go:334] "Generic (PLEG): container finished" podID="9e4679dd-f870-41e9-a32b-360f5d2ee81b" containerID="fafdd1094bd02786d745c984989852b0c690db38655f92b419733a2004f417c0" exitCode=0 Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.106745 4885 generic.go:334] "Generic (PLEG): container finished" podID="9e4679dd-f870-41e9-a32b-360f5d2ee81b" containerID="c9f99f9e26a04fb42b01a0783a5c4e360514818ce9102edc7fa08626d50b9a34" exitCode=0 Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.106752 4885 generic.go:334] "Generic (PLEG): container finished" podID="9e4679dd-f870-41e9-a32b-360f5d2ee81b" containerID="a6fa72b030d309656b73121e4ac2a3924a4e449aad66683dd9de5ee5f22d5a4e" exitCode=0 Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.106759 4885 generic.go:334] "Generic (PLEG): container finished" podID="9e4679dd-f870-41e9-a32b-360f5d2ee81b" containerID="a7642033b0b5995d332ccf7992d6d7e6a5e9c9369ae2d1071ccb2f536938c8f4" exitCode=0 Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.106765 4885 generic.go:334] "Generic (PLEG): container finished" podID="9e4679dd-f870-41e9-a32b-360f5d2ee81b" containerID="4310909ad9c7b609ed6814e04a525a6afae747775598e7275080413dec8b23a3" exitCode=0 Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.106771 4885 generic.go:334] "Generic (PLEG): container finished" podID="9e4679dd-f870-41e9-a32b-360f5d2ee81b" containerID="ef403e519d197fb37b3e6b76a14a5128ba3f30a7a5e8bf0a420cf6a39b89575f" exitCode=143 Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.106779 4885 generic.go:334] "Generic (PLEG): container finished" podID="9e4679dd-f870-41e9-a32b-360f5d2ee81b" containerID="8482ba457683ffa03100ec1f07ea9d3f918be5e490b5ba05c2ba965d03033191" exitCode=143 Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.106774 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" event={"ID":"9e4679dd-f870-41e9-a32b-360f5d2ee81b","Type":"ContainerDied","Data":"41121c480bb0f68a539497e7d1018e6582a1cf812f819b4c7327ad8827b316b7"} Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.106814 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" event={"ID":"9e4679dd-f870-41e9-a32b-360f5d2ee81b","Type":"ContainerDied","Data":"fafdd1094bd02786d745c984989852b0c690db38655f92b419733a2004f417c0"} Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.106830 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" event={"ID":"9e4679dd-f870-41e9-a32b-360f5d2ee81b","Type":"ContainerDied","Data":"c9f99f9e26a04fb42b01a0783a5c4e360514818ce9102edc7fa08626d50b9a34"} Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.106844 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" event={"ID":"9e4679dd-f870-41e9-a32b-360f5d2ee81b","Type":"ContainerDied","Data":"a6fa72b030d309656b73121e4ac2a3924a4e449aad66683dd9de5ee5f22d5a4e"} Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.106858 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" event={"ID":"9e4679dd-f870-41e9-a32b-360f5d2ee81b","Type":"ContainerDied","Data":"a7642033b0b5995d332ccf7992d6d7e6a5e9c9369ae2d1071ccb2f536938c8f4"} Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.106872 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" event={"ID":"9e4679dd-f870-41e9-a32b-360f5d2ee81b","Type":"ContainerDied","Data":"4310909ad9c7b609ed6814e04a525a6afae747775598e7275080413dec8b23a3"} Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.106880 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.106887 4885 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"41121c480bb0f68a539497e7d1018e6582a1cf812f819b4c7327ad8827b316b7"} Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.107043 4885 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2ef80e838accc59ad7a579c66f4fcd6881d23fa86fc5922d4b0085d211d5d94b"} Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.107053 4885 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fafdd1094bd02786d745c984989852b0c690db38655f92b419733a2004f417c0"} Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.107059 4885 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c9f99f9e26a04fb42b01a0783a5c4e360514818ce9102edc7fa08626d50b9a34"} Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.107065 4885 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a6fa72b030d309656b73121e4ac2a3924a4e449aad66683dd9de5ee5f22d5a4e"} Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.107070 4885 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a7642033b0b5995d332ccf7992d6d7e6a5e9c9369ae2d1071ccb2f536938c8f4"} Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.107075 4885 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4310909ad9c7b609ed6814e04a525a6afae747775598e7275080413dec8b23a3"} Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.107081 4885 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ef403e519d197fb37b3e6b76a14a5128ba3f30a7a5e8bf0a420cf6a39b89575f"} Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.107085 4885 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8482ba457683ffa03100ec1f07ea9d3f918be5e490b5ba05c2ba965d03033191"} Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.107091 4885 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0125e907d061a97286026ee823c4eed303734ed03ab4864f742592b23a1aa2f8"} Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.107101 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" event={"ID":"9e4679dd-f870-41e9-a32b-360f5d2ee81b","Type":"ContainerDied","Data":"ef403e519d197fb37b3e6b76a14a5128ba3f30a7a5e8bf0a420cf6a39b89575f"} Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.107116 4885 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"41121c480bb0f68a539497e7d1018e6582a1cf812f819b4c7327ad8827b316b7"} Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.107125 4885 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2ef80e838accc59ad7a579c66f4fcd6881d23fa86fc5922d4b0085d211d5d94b"} Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.107131 4885 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fafdd1094bd02786d745c984989852b0c690db38655f92b419733a2004f417c0"} Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.107135 4885 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c9f99f9e26a04fb42b01a0783a5c4e360514818ce9102edc7fa08626d50b9a34"} Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.107141 4885 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a6fa72b030d309656b73121e4ac2a3924a4e449aad66683dd9de5ee5f22d5a4e"} Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.107146 4885 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a7642033b0b5995d332ccf7992d6d7e6a5e9c9369ae2d1071ccb2f536938c8f4"} Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.107152 4885 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4310909ad9c7b609ed6814e04a525a6afae747775598e7275080413dec8b23a3"} Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.107159 4885 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ef403e519d197fb37b3e6b76a14a5128ba3f30a7a5e8bf0a420cf6a39b89575f"} Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.107163 4885 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8482ba457683ffa03100ec1f07ea9d3f918be5e490b5ba05c2ba965d03033191"} Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.107182 4885 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0125e907d061a97286026ee823c4eed303734ed03ab4864f742592b23a1aa2f8"} Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.107190 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" event={"ID":"9e4679dd-f870-41e9-a32b-360f5d2ee81b","Type":"ContainerDied","Data":"8482ba457683ffa03100ec1f07ea9d3f918be5e490b5ba05c2ba965d03033191"} Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.107198 4885 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"41121c480bb0f68a539497e7d1018e6582a1cf812f819b4c7327ad8827b316b7"} Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.107204 4885 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2ef80e838accc59ad7a579c66f4fcd6881d23fa86fc5922d4b0085d211d5d94b"} Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.107209 4885 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fafdd1094bd02786d745c984989852b0c690db38655f92b419733a2004f417c0"} Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.107215 4885 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c9f99f9e26a04fb42b01a0783a5c4e360514818ce9102edc7fa08626d50b9a34"} Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.107221 4885 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a6fa72b030d309656b73121e4ac2a3924a4e449aad66683dd9de5ee5f22d5a4e"} Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.107225 4885 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a7642033b0b5995d332ccf7992d6d7e6a5e9c9369ae2d1071ccb2f536938c8f4"} Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.107231 4885 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4310909ad9c7b609ed6814e04a525a6afae747775598e7275080413dec8b23a3"} Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.107236 4885 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ef403e519d197fb37b3e6b76a14a5128ba3f30a7a5e8bf0a420cf6a39b89575f"} Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.107242 4885 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8482ba457683ffa03100ec1f07ea9d3f918be5e490b5ba05c2ba965d03033191"} Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.107247 4885 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0125e907d061a97286026ee823c4eed303734ed03ab4864f742592b23a1aa2f8"} Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.107255 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dngcm" event={"ID":"9e4679dd-f870-41e9-a32b-360f5d2ee81b","Type":"ContainerDied","Data":"64efe9cde0ed394d9c75903a23e9b663a113d1fa0e7c74a8fb3c6217708805d4"} Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.107326 4885 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"41121c480bb0f68a539497e7d1018e6582a1cf812f819b4c7327ad8827b316b7"} Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.107335 4885 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2ef80e838accc59ad7a579c66f4fcd6881d23fa86fc5922d4b0085d211d5d94b"} Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.107340 4885 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fafdd1094bd02786d745c984989852b0c690db38655f92b419733a2004f417c0"} Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.107346 4885 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c9f99f9e26a04fb42b01a0783a5c4e360514818ce9102edc7fa08626d50b9a34"} Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.107374 4885 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a6fa72b030d309656b73121e4ac2a3924a4e449aad66683dd9de5ee5f22d5a4e"} Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.107401 4885 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a7642033b0b5995d332ccf7992d6d7e6a5e9c9369ae2d1071ccb2f536938c8f4"} Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.107407 4885 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4310909ad9c7b609ed6814e04a525a6afae747775598e7275080413dec8b23a3"} Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.107413 4885 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ef403e519d197fb37b3e6b76a14a5128ba3f30a7a5e8bf0a420cf6a39b89575f"} Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.107418 4885 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8482ba457683ffa03100ec1f07ea9d3f918be5e490b5ba05c2ba965d03033191"} Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.107424 4885 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0125e907d061a97286026ee823c4eed303734ed03ab4864f742592b23a1aa2f8"} Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.139472 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-dngcm"] Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.148843 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-dngcm"] Oct 02 01:56:40 crc kubenswrapper[4885]: E1002 01:56:40.152517 4885 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9e4679dd_f870_41e9_a32b_360f5d2ee81b.slice/crio-64efe9cde0ed394d9c75903a23e9b663a113d1fa0e7c74a8fb3c6217708805d4\": RecentStats: unable to find data in memory cache]" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.167998 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/c984321f-b4d7-49da-a4ee-2a61bcc84daa-run-systemd\") pod \"ovnkube-node-x69mt\" (UID: \"c984321f-b4d7-49da-a4ee-2a61bcc84daa\") " pod="openshift-ovn-kubernetes/ovnkube-node-x69mt" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.168089 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/c984321f-b4d7-49da-a4ee-2a61bcc84daa-systemd-units\") pod \"ovnkube-node-x69mt\" (UID: \"c984321f-b4d7-49da-a4ee-2a61bcc84daa\") " pod="openshift-ovn-kubernetes/ovnkube-node-x69mt" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.168134 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/c984321f-b4d7-49da-a4ee-2a61bcc84daa-ovnkube-script-lib\") pod \"ovnkube-node-x69mt\" (UID: \"c984321f-b4d7-49da-a4ee-2a61bcc84daa\") " pod="openshift-ovn-kubernetes/ovnkube-node-x69mt" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.168174 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c984321f-b4d7-49da-a4ee-2a61bcc84daa-etc-openvswitch\") pod \"ovnkube-node-x69mt\" (UID: \"c984321f-b4d7-49da-a4ee-2a61bcc84daa\") " pod="openshift-ovn-kubernetes/ovnkube-node-x69mt" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.168225 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/c984321f-b4d7-49da-a4ee-2a61bcc84daa-host-cni-netd\") pod \"ovnkube-node-x69mt\" (UID: \"c984321f-b4d7-49da-a4ee-2a61bcc84daa\") " pod="openshift-ovn-kubernetes/ovnkube-node-x69mt" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.168341 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hs7fp\" (UniqueName: \"kubernetes.io/projected/c984321f-b4d7-49da-a4ee-2a61bcc84daa-kube-api-access-hs7fp\") pod \"ovnkube-node-x69mt\" (UID: \"c984321f-b4d7-49da-a4ee-2a61bcc84daa\") " pod="openshift-ovn-kubernetes/ovnkube-node-x69mt" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.168380 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c984321f-b4d7-49da-a4ee-2a61bcc84daa-host-run-netns\") pod \"ovnkube-node-x69mt\" (UID: \"c984321f-b4d7-49da-a4ee-2a61bcc84daa\") " pod="openshift-ovn-kubernetes/ovnkube-node-x69mt" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.168415 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c984321f-b4d7-49da-a4ee-2a61bcc84daa-ovn-node-metrics-cert\") pod \"ovnkube-node-x69mt\" (UID: \"c984321f-b4d7-49da-a4ee-2a61bcc84daa\") " pod="openshift-ovn-kubernetes/ovnkube-node-x69mt" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.168481 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c984321f-b4d7-49da-a4ee-2a61bcc84daa-var-lib-openvswitch\") pod \"ovnkube-node-x69mt\" (UID: \"c984321f-b4d7-49da-a4ee-2a61bcc84daa\") " pod="openshift-ovn-kubernetes/ovnkube-node-x69mt" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.168562 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/c984321f-b4d7-49da-a4ee-2a61bcc84daa-host-kubelet\") pod \"ovnkube-node-x69mt\" (UID: \"c984321f-b4d7-49da-a4ee-2a61bcc84daa\") " pod="openshift-ovn-kubernetes/ovnkube-node-x69mt" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.168639 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/c984321f-b4d7-49da-a4ee-2a61bcc84daa-log-socket\") pod \"ovnkube-node-x69mt\" (UID: \"c984321f-b4d7-49da-a4ee-2a61bcc84daa\") " pod="openshift-ovn-kubernetes/ovnkube-node-x69mt" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.168682 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/c984321f-b4d7-49da-a4ee-2a61bcc84daa-env-overrides\") pod \"ovnkube-node-x69mt\" (UID: \"c984321f-b4d7-49da-a4ee-2a61bcc84daa\") " pod="openshift-ovn-kubernetes/ovnkube-node-x69mt" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.168746 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/c984321f-b4d7-49da-a4ee-2a61bcc84daa-ovnkube-config\") pod \"ovnkube-node-x69mt\" (UID: \"c984321f-b4d7-49da-a4ee-2a61bcc84daa\") " pod="openshift-ovn-kubernetes/ovnkube-node-x69mt" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.168771 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/c984321f-b4d7-49da-a4ee-2a61bcc84daa-run-ovn\") pod \"ovnkube-node-x69mt\" (UID: \"c984321f-b4d7-49da-a4ee-2a61bcc84daa\") " pod="openshift-ovn-kubernetes/ovnkube-node-x69mt" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.168798 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c984321f-b4d7-49da-a4ee-2a61bcc84daa-host-run-ovn-kubernetes\") pod \"ovnkube-node-x69mt\" (UID: \"c984321f-b4d7-49da-a4ee-2a61bcc84daa\") " pod="openshift-ovn-kubernetes/ovnkube-node-x69mt" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.168835 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c984321f-b4d7-49da-a4ee-2a61bcc84daa-run-openvswitch\") pod \"ovnkube-node-x69mt\" (UID: \"c984321f-b4d7-49da-a4ee-2a61bcc84daa\") " pod="openshift-ovn-kubernetes/ovnkube-node-x69mt" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.168854 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c984321f-b4d7-49da-a4ee-2a61bcc84daa-host-cni-bin\") pod \"ovnkube-node-x69mt\" (UID: \"c984321f-b4d7-49da-a4ee-2a61bcc84daa\") " pod="openshift-ovn-kubernetes/ovnkube-node-x69mt" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.168972 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/c984321f-b4d7-49da-a4ee-2a61bcc84daa-node-log\") pod \"ovnkube-node-x69mt\" (UID: \"c984321f-b4d7-49da-a4ee-2a61bcc84daa\") " pod="openshift-ovn-kubernetes/ovnkube-node-x69mt" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.169066 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c984321f-b4d7-49da-a4ee-2a61bcc84daa-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-x69mt\" (UID: \"c984321f-b4d7-49da-a4ee-2a61bcc84daa\") " pod="openshift-ovn-kubernetes/ovnkube-node-x69mt" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.169097 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/c984321f-b4d7-49da-a4ee-2a61bcc84daa-host-slash\") pod \"ovnkube-node-x69mt\" (UID: \"c984321f-b4d7-49da-a4ee-2a61bcc84daa\") " pod="openshift-ovn-kubernetes/ovnkube-node-x69mt" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.169314 4885 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9e4679dd-f870-41e9-a32b-360f5d2ee81b-host-cni-bin\") on node \"crc\" DevicePath \"\"" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.169336 4885 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/9e4679dd-f870-41e9-a32b-360f5d2ee81b-systemd-units\") on node \"crc\" DevicePath \"\"" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.169348 4885 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/9e4679dd-f870-41e9-a32b-360f5d2ee81b-run-systemd\") on node \"crc\" DevicePath \"\"" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.169363 4885 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9e4679dd-f870-41e9-a32b-360f5d2ee81b-run-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.169374 4885 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/9e4679dd-f870-41e9-a32b-360f5d2ee81b-node-log\") on node \"crc\" DevicePath \"\"" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.169383 4885 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/9e4679dd-f870-41e9-a32b-360f5d2ee81b-host-kubelet\") on node \"crc\" DevicePath \"\"" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.169395 4885 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9e4679dd-f870-41e9-a32b-360f5d2ee81b-host-run-netns\") on node \"crc\" DevicePath \"\"" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.169405 4885 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/9e4679dd-f870-41e9-a32b-360f5d2ee81b-host-cni-netd\") on node \"crc\" DevicePath \"\"" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.169417 4885 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9e4679dd-f870-41e9-a32b-360f5d2ee81b-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.169427 4885 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9e4679dd-f870-41e9-a32b-360f5d2ee81b-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.169438 4885 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9e4679dd-f870-41e9-a32b-360f5d2ee81b-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.169448 4885 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/9e4679dd-f870-41e9-a32b-360f5d2ee81b-log-socket\") on node \"crc\" DevicePath \"\"" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.169459 4885 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9e4679dd-f870-41e9-a32b-360f5d2ee81b-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.169475 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4b7jr\" (UniqueName: \"kubernetes.io/projected/9e4679dd-f870-41e9-a32b-360f5d2ee81b-kube-api-access-4b7jr\") on node \"crc\" DevicePath \"\"" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.169484 4885 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/9e4679dd-f870-41e9-a32b-360f5d2ee81b-host-slash\") on node \"crc\" DevicePath \"\"" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.169494 4885 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/9e4679dd-f870-41e9-a32b-360f5d2ee81b-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.169505 4885 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9e4679dd-f870-41e9-a32b-360f5d2ee81b-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.169516 4885 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9e4679dd-f870-41e9-a32b-360f5d2ee81b-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.169529 4885 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9e4679dd-f870-41e9-a32b-360f5d2ee81b-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.169539 4885 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/9e4679dd-f870-41e9-a32b-360f5d2ee81b-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.225607 4885 scope.go:117] "RemoveContainer" containerID="41121c480bb0f68a539497e7d1018e6582a1cf812f819b4c7327ad8827b316b7" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.244714 4885 scope.go:117] "RemoveContainer" containerID="2ef80e838accc59ad7a579c66f4fcd6881d23fa86fc5922d4b0085d211d5d94b" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.261170 4885 scope.go:117] "RemoveContainer" containerID="fafdd1094bd02786d745c984989852b0c690db38655f92b419733a2004f417c0" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.271175 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/c984321f-b4d7-49da-a4ee-2a61bcc84daa-run-systemd\") pod \"ovnkube-node-x69mt\" (UID: \"c984321f-b4d7-49da-a4ee-2a61bcc84daa\") " pod="openshift-ovn-kubernetes/ovnkube-node-x69mt" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.271268 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/c984321f-b4d7-49da-a4ee-2a61bcc84daa-systemd-units\") pod \"ovnkube-node-x69mt\" (UID: \"c984321f-b4d7-49da-a4ee-2a61bcc84daa\") " pod="openshift-ovn-kubernetes/ovnkube-node-x69mt" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.271332 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/c984321f-b4d7-49da-a4ee-2a61bcc84daa-ovnkube-script-lib\") pod \"ovnkube-node-x69mt\" (UID: \"c984321f-b4d7-49da-a4ee-2a61bcc84daa\") " pod="openshift-ovn-kubernetes/ovnkube-node-x69mt" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.271370 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c984321f-b4d7-49da-a4ee-2a61bcc84daa-etc-openvswitch\") pod \"ovnkube-node-x69mt\" (UID: \"c984321f-b4d7-49da-a4ee-2a61bcc84daa\") " pod="openshift-ovn-kubernetes/ovnkube-node-x69mt" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.271397 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/c984321f-b4d7-49da-a4ee-2a61bcc84daa-run-systemd\") pod \"ovnkube-node-x69mt\" (UID: \"c984321f-b4d7-49da-a4ee-2a61bcc84daa\") " pod="openshift-ovn-kubernetes/ovnkube-node-x69mt" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.271426 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/c984321f-b4d7-49da-a4ee-2a61bcc84daa-host-cni-netd\") pod \"ovnkube-node-x69mt\" (UID: \"c984321f-b4d7-49da-a4ee-2a61bcc84daa\") " pod="openshift-ovn-kubernetes/ovnkube-node-x69mt" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.271433 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/c984321f-b4d7-49da-a4ee-2a61bcc84daa-systemd-units\") pod \"ovnkube-node-x69mt\" (UID: \"c984321f-b4d7-49da-a4ee-2a61bcc84daa\") " pod="openshift-ovn-kubernetes/ovnkube-node-x69mt" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.271782 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/c984321f-b4d7-49da-a4ee-2a61bcc84daa-host-cni-netd\") pod \"ovnkube-node-x69mt\" (UID: \"c984321f-b4d7-49da-a4ee-2a61bcc84daa\") " pod="openshift-ovn-kubernetes/ovnkube-node-x69mt" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.271768 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c984321f-b4d7-49da-a4ee-2a61bcc84daa-etc-openvswitch\") pod \"ovnkube-node-x69mt\" (UID: \"c984321f-b4d7-49da-a4ee-2a61bcc84daa\") " pod="openshift-ovn-kubernetes/ovnkube-node-x69mt" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.271842 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hs7fp\" (UniqueName: \"kubernetes.io/projected/c984321f-b4d7-49da-a4ee-2a61bcc84daa-kube-api-access-hs7fp\") pod \"ovnkube-node-x69mt\" (UID: \"c984321f-b4d7-49da-a4ee-2a61bcc84daa\") " pod="openshift-ovn-kubernetes/ovnkube-node-x69mt" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.271885 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c984321f-b4d7-49da-a4ee-2a61bcc84daa-host-run-netns\") pod \"ovnkube-node-x69mt\" (UID: \"c984321f-b4d7-49da-a4ee-2a61bcc84daa\") " pod="openshift-ovn-kubernetes/ovnkube-node-x69mt" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.271935 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c984321f-b4d7-49da-a4ee-2a61bcc84daa-host-run-netns\") pod \"ovnkube-node-x69mt\" (UID: \"c984321f-b4d7-49da-a4ee-2a61bcc84daa\") " pod="openshift-ovn-kubernetes/ovnkube-node-x69mt" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.271943 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c984321f-b4d7-49da-a4ee-2a61bcc84daa-ovn-node-metrics-cert\") pod \"ovnkube-node-x69mt\" (UID: \"c984321f-b4d7-49da-a4ee-2a61bcc84daa\") " pod="openshift-ovn-kubernetes/ovnkube-node-x69mt" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.272001 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/c984321f-b4d7-49da-a4ee-2a61bcc84daa-host-kubelet\") pod \"ovnkube-node-x69mt\" (UID: \"c984321f-b4d7-49da-a4ee-2a61bcc84daa\") " pod="openshift-ovn-kubernetes/ovnkube-node-x69mt" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.272027 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c984321f-b4d7-49da-a4ee-2a61bcc84daa-var-lib-openvswitch\") pod \"ovnkube-node-x69mt\" (UID: \"c984321f-b4d7-49da-a4ee-2a61bcc84daa\") " pod="openshift-ovn-kubernetes/ovnkube-node-x69mt" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.272085 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/c984321f-b4d7-49da-a4ee-2a61bcc84daa-log-socket\") pod \"ovnkube-node-x69mt\" (UID: \"c984321f-b4d7-49da-a4ee-2a61bcc84daa\") " pod="openshift-ovn-kubernetes/ovnkube-node-x69mt" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.272124 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/c984321f-b4d7-49da-a4ee-2a61bcc84daa-env-overrides\") pod \"ovnkube-node-x69mt\" (UID: \"c984321f-b4d7-49da-a4ee-2a61bcc84daa\") " pod="openshift-ovn-kubernetes/ovnkube-node-x69mt" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.272194 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/c984321f-b4d7-49da-a4ee-2a61bcc84daa-log-socket\") pod \"ovnkube-node-x69mt\" (UID: \"c984321f-b4d7-49da-a4ee-2a61bcc84daa\") " pod="openshift-ovn-kubernetes/ovnkube-node-x69mt" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.272209 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/c984321f-b4d7-49da-a4ee-2a61bcc84daa-host-kubelet\") pod \"ovnkube-node-x69mt\" (UID: \"c984321f-b4d7-49da-a4ee-2a61bcc84daa\") " pod="openshift-ovn-kubernetes/ovnkube-node-x69mt" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.272455 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/c984321f-b4d7-49da-a4ee-2a61bcc84daa-ovnkube-config\") pod \"ovnkube-node-x69mt\" (UID: \"c984321f-b4d7-49da-a4ee-2a61bcc84daa\") " pod="openshift-ovn-kubernetes/ovnkube-node-x69mt" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.272908 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/c984321f-b4d7-49da-a4ee-2a61bcc84daa-ovnkube-script-lib\") pod \"ovnkube-node-x69mt\" (UID: \"c984321f-b4d7-49da-a4ee-2a61bcc84daa\") " pod="openshift-ovn-kubernetes/ovnkube-node-x69mt" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.273140 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c984321f-b4d7-49da-a4ee-2a61bcc84daa-var-lib-openvswitch\") pod \"ovnkube-node-x69mt\" (UID: \"c984321f-b4d7-49da-a4ee-2a61bcc84daa\") " pod="openshift-ovn-kubernetes/ovnkube-node-x69mt" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.273241 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/c984321f-b4d7-49da-a4ee-2a61bcc84daa-env-overrides\") pod \"ovnkube-node-x69mt\" (UID: \"c984321f-b4d7-49da-a4ee-2a61bcc84daa\") " pod="openshift-ovn-kubernetes/ovnkube-node-x69mt" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.273327 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/c984321f-b4d7-49da-a4ee-2a61bcc84daa-run-ovn\") pod \"ovnkube-node-x69mt\" (UID: \"c984321f-b4d7-49da-a4ee-2a61bcc84daa\") " pod="openshift-ovn-kubernetes/ovnkube-node-x69mt" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.273357 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c984321f-b4d7-49da-a4ee-2a61bcc84daa-host-run-ovn-kubernetes\") pod \"ovnkube-node-x69mt\" (UID: \"c984321f-b4d7-49da-a4ee-2a61bcc84daa\") " pod="openshift-ovn-kubernetes/ovnkube-node-x69mt" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.273388 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/c984321f-b4d7-49da-a4ee-2a61bcc84daa-run-ovn\") pod \"ovnkube-node-x69mt\" (UID: \"c984321f-b4d7-49da-a4ee-2a61bcc84daa\") " pod="openshift-ovn-kubernetes/ovnkube-node-x69mt" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.273359 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/c984321f-b4d7-49da-a4ee-2a61bcc84daa-ovnkube-config\") pod \"ovnkube-node-x69mt\" (UID: \"c984321f-b4d7-49da-a4ee-2a61bcc84daa\") " pod="openshift-ovn-kubernetes/ovnkube-node-x69mt" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.273461 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c984321f-b4d7-49da-a4ee-2a61bcc84daa-host-cni-bin\") pod \"ovnkube-node-x69mt\" (UID: \"c984321f-b4d7-49da-a4ee-2a61bcc84daa\") " pod="openshift-ovn-kubernetes/ovnkube-node-x69mt" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.273516 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c984321f-b4d7-49da-a4ee-2a61bcc84daa-run-openvswitch\") pod \"ovnkube-node-x69mt\" (UID: \"c984321f-b4d7-49da-a4ee-2a61bcc84daa\") " pod="openshift-ovn-kubernetes/ovnkube-node-x69mt" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.273554 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c984321f-b4d7-49da-a4ee-2a61bcc84daa-host-cni-bin\") pod \"ovnkube-node-x69mt\" (UID: \"c984321f-b4d7-49da-a4ee-2a61bcc84daa\") " pod="openshift-ovn-kubernetes/ovnkube-node-x69mt" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.273596 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c984321f-b4d7-49da-a4ee-2a61bcc84daa-run-openvswitch\") pod \"ovnkube-node-x69mt\" (UID: \"c984321f-b4d7-49da-a4ee-2a61bcc84daa\") " pod="openshift-ovn-kubernetes/ovnkube-node-x69mt" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.273601 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/c984321f-b4d7-49da-a4ee-2a61bcc84daa-node-log\") pod \"ovnkube-node-x69mt\" (UID: \"c984321f-b4d7-49da-a4ee-2a61bcc84daa\") " pod="openshift-ovn-kubernetes/ovnkube-node-x69mt" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.273647 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c984321f-b4d7-49da-a4ee-2a61bcc84daa-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-x69mt\" (UID: \"c984321f-b4d7-49da-a4ee-2a61bcc84daa\") " pod="openshift-ovn-kubernetes/ovnkube-node-x69mt" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.273682 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/c984321f-b4d7-49da-a4ee-2a61bcc84daa-host-slash\") pod \"ovnkube-node-x69mt\" (UID: \"c984321f-b4d7-49da-a4ee-2a61bcc84daa\") " pod="openshift-ovn-kubernetes/ovnkube-node-x69mt" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.273691 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c984321f-b4d7-49da-a4ee-2a61bcc84daa-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-x69mt\" (UID: \"c984321f-b4d7-49da-a4ee-2a61bcc84daa\") " pod="openshift-ovn-kubernetes/ovnkube-node-x69mt" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.273446 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c984321f-b4d7-49da-a4ee-2a61bcc84daa-host-run-ovn-kubernetes\") pod \"ovnkube-node-x69mt\" (UID: \"c984321f-b4d7-49da-a4ee-2a61bcc84daa\") " pod="openshift-ovn-kubernetes/ovnkube-node-x69mt" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.273819 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/c984321f-b4d7-49da-a4ee-2a61bcc84daa-host-slash\") pod \"ovnkube-node-x69mt\" (UID: \"c984321f-b4d7-49da-a4ee-2a61bcc84daa\") " pod="openshift-ovn-kubernetes/ovnkube-node-x69mt" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.274226 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/c984321f-b4d7-49da-a4ee-2a61bcc84daa-node-log\") pod \"ovnkube-node-x69mt\" (UID: \"c984321f-b4d7-49da-a4ee-2a61bcc84daa\") " pod="openshift-ovn-kubernetes/ovnkube-node-x69mt" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.280571 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c984321f-b4d7-49da-a4ee-2a61bcc84daa-ovn-node-metrics-cert\") pod \"ovnkube-node-x69mt\" (UID: \"c984321f-b4d7-49da-a4ee-2a61bcc84daa\") " pod="openshift-ovn-kubernetes/ovnkube-node-x69mt" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.281766 4885 scope.go:117] "RemoveContainer" containerID="c9f99f9e26a04fb42b01a0783a5c4e360514818ce9102edc7fa08626d50b9a34" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.297469 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hs7fp\" (UniqueName: \"kubernetes.io/projected/c984321f-b4d7-49da-a4ee-2a61bcc84daa-kube-api-access-hs7fp\") pod \"ovnkube-node-x69mt\" (UID: \"c984321f-b4d7-49da-a4ee-2a61bcc84daa\") " pod="openshift-ovn-kubernetes/ovnkube-node-x69mt" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.304815 4885 scope.go:117] "RemoveContainer" containerID="a6fa72b030d309656b73121e4ac2a3924a4e449aad66683dd9de5ee5f22d5a4e" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.324771 4885 scope.go:117] "RemoveContainer" containerID="a7642033b0b5995d332ccf7992d6d7e6a5e9c9369ae2d1071ccb2f536938c8f4" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.331246 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-x69mt" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.357497 4885 scope.go:117] "RemoveContainer" containerID="4310909ad9c7b609ed6814e04a525a6afae747775598e7275080413dec8b23a3" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.381574 4885 scope.go:117] "RemoveContainer" containerID="ef403e519d197fb37b3e6b76a14a5128ba3f30a7a5e8bf0a420cf6a39b89575f" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.486839 4885 scope.go:117] "RemoveContainer" containerID="8482ba457683ffa03100ec1f07ea9d3f918be5e490b5ba05c2ba965d03033191" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.505544 4885 scope.go:117] "RemoveContainer" containerID="0125e907d061a97286026ee823c4eed303734ed03ab4864f742592b23a1aa2f8" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.526373 4885 scope.go:117] "RemoveContainer" containerID="41121c480bb0f68a539497e7d1018e6582a1cf812f819b4c7327ad8827b316b7" Oct 02 01:56:40 crc kubenswrapper[4885]: E1002 01:56:40.526945 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"41121c480bb0f68a539497e7d1018e6582a1cf812f819b4c7327ad8827b316b7\": container with ID starting with 41121c480bb0f68a539497e7d1018e6582a1cf812f819b4c7327ad8827b316b7 not found: ID does not exist" containerID="41121c480bb0f68a539497e7d1018e6582a1cf812f819b4c7327ad8827b316b7" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.527006 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"41121c480bb0f68a539497e7d1018e6582a1cf812f819b4c7327ad8827b316b7"} err="failed to get container status \"41121c480bb0f68a539497e7d1018e6582a1cf812f819b4c7327ad8827b316b7\": rpc error: code = NotFound desc = could not find container \"41121c480bb0f68a539497e7d1018e6582a1cf812f819b4c7327ad8827b316b7\": container with ID starting with 41121c480bb0f68a539497e7d1018e6582a1cf812f819b4c7327ad8827b316b7 not found: ID does not exist" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.527052 4885 scope.go:117] "RemoveContainer" containerID="2ef80e838accc59ad7a579c66f4fcd6881d23fa86fc5922d4b0085d211d5d94b" Oct 02 01:56:40 crc kubenswrapper[4885]: E1002 01:56:40.527545 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2ef80e838accc59ad7a579c66f4fcd6881d23fa86fc5922d4b0085d211d5d94b\": container with ID starting with 2ef80e838accc59ad7a579c66f4fcd6881d23fa86fc5922d4b0085d211d5d94b not found: ID does not exist" containerID="2ef80e838accc59ad7a579c66f4fcd6881d23fa86fc5922d4b0085d211d5d94b" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.527595 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2ef80e838accc59ad7a579c66f4fcd6881d23fa86fc5922d4b0085d211d5d94b"} err="failed to get container status \"2ef80e838accc59ad7a579c66f4fcd6881d23fa86fc5922d4b0085d211d5d94b\": rpc error: code = NotFound desc = could not find container \"2ef80e838accc59ad7a579c66f4fcd6881d23fa86fc5922d4b0085d211d5d94b\": container with ID starting with 2ef80e838accc59ad7a579c66f4fcd6881d23fa86fc5922d4b0085d211d5d94b not found: ID does not exist" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.527636 4885 scope.go:117] "RemoveContainer" containerID="fafdd1094bd02786d745c984989852b0c690db38655f92b419733a2004f417c0" Oct 02 01:56:40 crc kubenswrapper[4885]: E1002 01:56:40.528069 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fafdd1094bd02786d745c984989852b0c690db38655f92b419733a2004f417c0\": container with ID starting with fafdd1094bd02786d745c984989852b0c690db38655f92b419733a2004f417c0 not found: ID does not exist" containerID="fafdd1094bd02786d745c984989852b0c690db38655f92b419733a2004f417c0" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.528117 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fafdd1094bd02786d745c984989852b0c690db38655f92b419733a2004f417c0"} err="failed to get container status \"fafdd1094bd02786d745c984989852b0c690db38655f92b419733a2004f417c0\": rpc error: code = NotFound desc = could not find container \"fafdd1094bd02786d745c984989852b0c690db38655f92b419733a2004f417c0\": container with ID starting with fafdd1094bd02786d745c984989852b0c690db38655f92b419733a2004f417c0 not found: ID does not exist" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.528155 4885 scope.go:117] "RemoveContainer" containerID="c9f99f9e26a04fb42b01a0783a5c4e360514818ce9102edc7fa08626d50b9a34" Oct 02 01:56:40 crc kubenswrapper[4885]: E1002 01:56:40.528806 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c9f99f9e26a04fb42b01a0783a5c4e360514818ce9102edc7fa08626d50b9a34\": container with ID starting with c9f99f9e26a04fb42b01a0783a5c4e360514818ce9102edc7fa08626d50b9a34 not found: ID does not exist" containerID="c9f99f9e26a04fb42b01a0783a5c4e360514818ce9102edc7fa08626d50b9a34" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.528849 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c9f99f9e26a04fb42b01a0783a5c4e360514818ce9102edc7fa08626d50b9a34"} err="failed to get container status \"c9f99f9e26a04fb42b01a0783a5c4e360514818ce9102edc7fa08626d50b9a34\": rpc error: code = NotFound desc = could not find container \"c9f99f9e26a04fb42b01a0783a5c4e360514818ce9102edc7fa08626d50b9a34\": container with ID starting with c9f99f9e26a04fb42b01a0783a5c4e360514818ce9102edc7fa08626d50b9a34 not found: ID does not exist" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.528885 4885 scope.go:117] "RemoveContainer" containerID="a6fa72b030d309656b73121e4ac2a3924a4e449aad66683dd9de5ee5f22d5a4e" Oct 02 01:56:40 crc kubenswrapper[4885]: E1002 01:56:40.529677 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a6fa72b030d309656b73121e4ac2a3924a4e449aad66683dd9de5ee5f22d5a4e\": container with ID starting with a6fa72b030d309656b73121e4ac2a3924a4e449aad66683dd9de5ee5f22d5a4e not found: ID does not exist" containerID="a6fa72b030d309656b73121e4ac2a3924a4e449aad66683dd9de5ee5f22d5a4e" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.529740 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a6fa72b030d309656b73121e4ac2a3924a4e449aad66683dd9de5ee5f22d5a4e"} err="failed to get container status \"a6fa72b030d309656b73121e4ac2a3924a4e449aad66683dd9de5ee5f22d5a4e\": rpc error: code = NotFound desc = could not find container \"a6fa72b030d309656b73121e4ac2a3924a4e449aad66683dd9de5ee5f22d5a4e\": container with ID starting with a6fa72b030d309656b73121e4ac2a3924a4e449aad66683dd9de5ee5f22d5a4e not found: ID does not exist" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.529782 4885 scope.go:117] "RemoveContainer" containerID="a7642033b0b5995d332ccf7992d6d7e6a5e9c9369ae2d1071ccb2f536938c8f4" Oct 02 01:56:40 crc kubenswrapper[4885]: E1002 01:56:40.530386 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a7642033b0b5995d332ccf7992d6d7e6a5e9c9369ae2d1071ccb2f536938c8f4\": container with ID starting with a7642033b0b5995d332ccf7992d6d7e6a5e9c9369ae2d1071ccb2f536938c8f4 not found: ID does not exist" containerID="a7642033b0b5995d332ccf7992d6d7e6a5e9c9369ae2d1071ccb2f536938c8f4" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.530466 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a7642033b0b5995d332ccf7992d6d7e6a5e9c9369ae2d1071ccb2f536938c8f4"} err="failed to get container status \"a7642033b0b5995d332ccf7992d6d7e6a5e9c9369ae2d1071ccb2f536938c8f4\": rpc error: code = NotFound desc = could not find container \"a7642033b0b5995d332ccf7992d6d7e6a5e9c9369ae2d1071ccb2f536938c8f4\": container with ID starting with a7642033b0b5995d332ccf7992d6d7e6a5e9c9369ae2d1071ccb2f536938c8f4 not found: ID does not exist" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.530520 4885 scope.go:117] "RemoveContainer" containerID="4310909ad9c7b609ed6814e04a525a6afae747775598e7275080413dec8b23a3" Oct 02 01:56:40 crc kubenswrapper[4885]: E1002 01:56:40.531366 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4310909ad9c7b609ed6814e04a525a6afae747775598e7275080413dec8b23a3\": container with ID starting with 4310909ad9c7b609ed6814e04a525a6afae747775598e7275080413dec8b23a3 not found: ID does not exist" containerID="4310909ad9c7b609ed6814e04a525a6afae747775598e7275080413dec8b23a3" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.531416 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4310909ad9c7b609ed6814e04a525a6afae747775598e7275080413dec8b23a3"} err="failed to get container status \"4310909ad9c7b609ed6814e04a525a6afae747775598e7275080413dec8b23a3\": rpc error: code = NotFound desc = could not find container \"4310909ad9c7b609ed6814e04a525a6afae747775598e7275080413dec8b23a3\": container with ID starting with 4310909ad9c7b609ed6814e04a525a6afae747775598e7275080413dec8b23a3 not found: ID does not exist" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.531452 4885 scope.go:117] "RemoveContainer" containerID="ef403e519d197fb37b3e6b76a14a5128ba3f30a7a5e8bf0a420cf6a39b89575f" Oct 02 01:56:40 crc kubenswrapper[4885]: E1002 01:56:40.531948 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ef403e519d197fb37b3e6b76a14a5128ba3f30a7a5e8bf0a420cf6a39b89575f\": container with ID starting with ef403e519d197fb37b3e6b76a14a5128ba3f30a7a5e8bf0a420cf6a39b89575f not found: ID does not exist" containerID="ef403e519d197fb37b3e6b76a14a5128ba3f30a7a5e8bf0a420cf6a39b89575f" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.531995 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef403e519d197fb37b3e6b76a14a5128ba3f30a7a5e8bf0a420cf6a39b89575f"} err="failed to get container status \"ef403e519d197fb37b3e6b76a14a5128ba3f30a7a5e8bf0a420cf6a39b89575f\": rpc error: code = NotFound desc = could not find container \"ef403e519d197fb37b3e6b76a14a5128ba3f30a7a5e8bf0a420cf6a39b89575f\": container with ID starting with ef403e519d197fb37b3e6b76a14a5128ba3f30a7a5e8bf0a420cf6a39b89575f not found: ID does not exist" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.532024 4885 scope.go:117] "RemoveContainer" containerID="8482ba457683ffa03100ec1f07ea9d3f918be5e490b5ba05c2ba965d03033191" Oct 02 01:56:40 crc kubenswrapper[4885]: E1002 01:56:40.532693 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8482ba457683ffa03100ec1f07ea9d3f918be5e490b5ba05c2ba965d03033191\": container with ID starting with 8482ba457683ffa03100ec1f07ea9d3f918be5e490b5ba05c2ba965d03033191 not found: ID does not exist" containerID="8482ba457683ffa03100ec1f07ea9d3f918be5e490b5ba05c2ba965d03033191" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.532742 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8482ba457683ffa03100ec1f07ea9d3f918be5e490b5ba05c2ba965d03033191"} err="failed to get container status \"8482ba457683ffa03100ec1f07ea9d3f918be5e490b5ba05c2ba965d03033191\": rpc error: code = NotFound desc = could not find container \"8482ba457683ffa03100ec1f07ea9d3f918be5e490b5ba05c2ba965d03033191\": container with ID starting with 8482ba457683ffa03100ec1f07ea9d3f918be5e490b5ba05c2ba965d03033191 not found: ID does not exist" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.532771 4885 scope.go:117] "RemoveContainer" containerID="0125e907d061a97286026ee823c4eed303734ed03ab4864f742592b23a1aa2f8" Oct 02 01:56:40 crc kubenswrapper[4885]: E1002 01:56:40.533661 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0125e907d061a97286026ee823c4eed303734ed03ab4864f742592b23a1aa2f8\": container with ID starting with 0125e907d061a97286026ee823c4eed303734ed03ab4864f742592b23a1aa2f8 not found: ID does not exist" containerID="0125e907d061a97286026ee823c4eed303734ed03ab4864f742592b23a1aa2f8" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.533707 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0125e907d061a97286026ee823c4eed303734ed03ab4864f742592b23a1aa2f8"} err="failed to get container status \"0125e907d061a97286026ee823c4eed303734ed03ab4864f742592b23a1aa2f8\": rpc error: code = NotFound desc = could not find container \"0125e907d061a97286026ee823c4eed303734ed03ab4864f742592b23a1aa2f8\": container with ID starting with 0125e907d061a97286026ee823c4eed303734ed03ab4864f742592b23a1aa2f8 not found: ID does not exist" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.533734 4885 scope.go:117] "RemoveContainer" containerID="41121c480bb0f68a539497e7d1018e6582a1cf812f819b4c7327ad8827b316b7" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.534433 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"41121c480bb0f68a539497e7d1018e6582a1cf812f819b4c7327ad8827b316b7"} err="failed to get container status \"41121c480bb0f68a539497e7d1018e6582a1cf812f819b4c7327ad8827b316b7\": rpc error: code = NotFound desc = could not find container \"41121c480bb0f68a539497e7d1018e6582a1cf812f819b4c7327ad8827b316b7\": container with ID starting with 41121c480bb0f68a539497e7d1018e6582a1cf812f819b4c7327ad8827b316b7 not found: ID does not exist" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.534497 4885 scope.go:117] "RemoveContainer" containerID="2ef80e838accc59ad7a579c66f4fcd6881d23fa86fc5922d4b0085d211d5d94b" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.534927 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2ef80e838accc59ad7a579c66f4fcd6881d23fa86fc5922d4b0085d211d5d94b"} err="failed to get container status \"2ef80e838accc59ad7a579c66f4fcd6881d23fa86fc5922d4b0085d211d5d94b\": rpc error: code = NotFound desc = could not find container \"2ef80e838accc59ad7a579c66f4fcd6881d23fa86fc5922d4b0085d211d5d94b\": container with ID starting with 2ef80e838accc59ad7a579c66f4fcd6881d23fa86fc5922d4b0085d211d5d94b not found: ID does not exist" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.534968 4885 scope.go:117] "RemoveContainer" containerID="fafdd1094bd02786d745c984989852b0c690db38655f92b419733a2004f417c0" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.535587 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fafdd1094bd02786d745c984989852b0c690db38655f92b419733a2004f417c0"} err="failed to get container status \"fafdd1094bd02786d745c984989852b0c690db38655f92b419733a2004f417c0\": rpc error: code = NotFound desc = could not find container \"fafdd1094bd02786d745c984989852b0c690db38655f92b419733a2004f417c0\": container with ID starting with fafdd1094bd02786d745c984989852b0c690db38655f92b419733a2004f417c0 not found: ID does not exist" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.535625 4885 scope.go:117] "RemoveContainer" containerID="c9f99f9e26a04fb42b01a0783a5c4e360514818ce9102edc7fa08626d50b9a34" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.535943 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c9f99f9e26a04fb42b01a0783a5c4e360514818ce9102edc7fa08626d50b9a34"} err="failed to get container status \"c9f99f9e26a04fb42b01a0783a5c4e360514818ce9102edc7fa08626d50b9a34\": rpc error: code = NotFound desc = could not find container \"c9f99f9e26a04fb42b01a0783a5c4e360514818ce9102edc7fa08626d50b9a34\": container with ID starting with c9f99f9e26a04fb42b01a0783a5c4e360514818ce9102edc7fa08626d50b9a34 not found: ID does not exist" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.535996 4885 scope.go:117] "RemoveContainer" containerID="a6fa72b030d309656b73121e4ac2a3924a4e449aad66683dd9de5ee5f22d5a4e" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.536724 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a6fa72b030d309656b73121e4ac2a3924a4e449aad66683dd9de5ee5f22d5a4e"} err="failed to get container status \"a6fa72b030d309656b73121e4ac2a3924a4e449aad66683dd9de5ee5f22d5a4e\": rpc error: code = NotFound desc = could not find container \"a6fa72b030d309656b73121e4ac2a3924a4e449aad66683dd9de5ee5f22d5a4e\": container with ID starting with a6fa72b030d309656b73121e4ac2a3924a4e449aad66683dd9de5ee5f22d5a4e not found: ID does not exist" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.536765 4885 scope.go:117] "RemoveContainer" containerID="a7642033b0b5995d332ccf7992d6d7e6a5e9c9369ae2d1071ccb2f536938c8f4" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.537092 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a7642033b0b5995d332ccf7992d6d7e6a5e9c9369ae2d1071ccb2f536938c8f4"} err="failed to get container status \"a7642033b0b5995d332ccf7992d6d7e6a5e9c9369ae2d1071ccb2f536938c8f4\": rpc error: code = NotFound desc = could not find container \"a7642033b0b5995d332ccf7992d6d7e6a5e9c9369ae2d1071ccb2f536938c8f4\": container with ID starting with a7642033b0b5995d332ccf7992d6d7e6a5e9c9369ae2d1071ccb2f536938c8f4 not found: ID does not exist" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.537141 4885 scope.go:117] "RemoveContainer" containerID="4310909ad9c7b609ed6814e04a525a6afae747775598e7275080413dec8b23a3" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.538704 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4310909ad9c7b609ed6814e04a525a6afae747775598e7275080413dec8b23a3"} err="failed to get container status \"4310909ad9c7b609ed6814e04a525a6afae747775598e7275080413dec8b23a3\": rpc error: code = NotFound desc = could not find container \"4310909ad9c7b609ed6814e04a525a6afae747775598e7275080413dec8b23a3\": container with ID starting with 4310909ad9c7b609ed6814e04a525a6afae747775598e7275080413dec8b23a3 not found: ID does not exist" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.538748 4885 scope.go:117] "RemoveContainer" containerID="ef403e519d197fb37b3e6b76a14a5128ba3f30a7a5e8bf0a420cf6a39b89575f" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.539171 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef403e519d197fb37b3e6b76a14a5128ba3f30a7a5e8bf0a420cf6a39b89575f"} err="failed to get container status \"ef403e519d197fb37b3e6b76a14a5128ba3f30a7a5e8bf0a420cf6a39b89575f\": rpc error: code = NotFound desc = could not find container \"ef403e519d197fb37b3e6b76a14a5128ba3f30a7a5e8bf0a420cf6a39b89575f\": container with ID starting with ef403e519d197fb37b3e6b76a14a5128ba3f30a7a5e8bf0a420cf6a39b89575f not found: ID does not exist" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.539336 4885 scope.go:117] "RemoveContainer" containerID="8482ba457683ffa03100ec1f07ea9d3f918be5e490b5ba05c2ba965d03033191" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.540194 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8482ba457683ffa03100ec1f07ea9d3f918be5e490b5ba05c2ba965d03033191"} err="failed to get container status \"8482ba457683ffa03100ec1f07ea9d3f918be5e490b5ba05c2ba965d03033191\": rpc error: code = NotFound desc = could not find container \"8482ba457683ffa03100ec1f07ea9d3f918be5e490b5ba05c2ba965d03033191\": container with ID starting with 8482ba457683ffa03100ec1f07ea9d3f918be5e490b5ba05c2ba965d03033191 not found: ID does not exist" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.540239 4885 scope.go:117] "RemoveContainer" containerID="0125e907d061a97286026ee823c4eed303734ed03ab4864f742592b23a1aa2f8" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.540714 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0125e907d061a97286026ee823c4eed303734ed03ab4864f742592b23a1aa2f8"} err="failed to get container status \"0125e907d061a97286026ee823c4eed303734ed03ab4864f742592b23a1aa2f8\": rpc error: code = NotFound desc = could not find container \"0125e907d061a97286026ee823c4eed303734ed03ab4864f742592b23a1aa2f8\": container with ID starting with 0125e907d061a97286026ee823c4eed303734ed03ab4864f742592b23a1aa2f8 not found: ID does not exist" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.540755 4885 scope.go:117] "RemoveContainer" containerID="41121c480bb0f68a539497e7d1018e6582a1cf812f819b4c7327ad8827b316b7" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.541246 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"41121c480bb0f68a539497e7d1018e6582a1cf812f819b4c7327ad8827b316b7"} err="failed to get container status \"41121c480bb0f68a539497e7d1018e6582a1cf812f819b4c7327ad8827b316b7\": rpc error: code = NotFound desc = could not find container \"41121c480bb0f68a539497e7d1018e6582a1cf812f819b4c7327ad8827b316b7\": container with ID starting with 41121c480bb0f68a539497e7d1018e6582a1cf812f819b4c7327ad8827b316b7 not found: ID does not exist" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.541327 4885 scope.go:117] "RemoveContainer" containerID="2ef80e838accc59ad7a579c66f4fcd6881d23fa86fc5922d4b0085d211d5d94b" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.541824 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2ef80e838accc59ad7a579c66f4fcd6881d23fa86fc5922d4b0085d211d5d94b"} err="failed to get container status \"2ef80e838accc59ad7a579c66f4fcd6881d23fa86fc5922d4b0085d211d5d94b\": rpc error: code = NotFound desc = could not find container \"2ef80e838accc59ad7a579c66f4fcd6881d23fa86fc5922d4b0085d211d5d94b\": container with ID starting with 2ef80e838accc59ad7a579c66f4fcd6881d23fa86fc5922d4b0085d211d5d94b not found: ID does not exist" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.541868 4885 scope.go:117] "RemoveContainer" containerID="fafdd1094bd02786d745c984989852b0c690db38655f92b419733a2004f417c0" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.542331 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fafdd1094bd02786d745c984989852b0c690db38655f92b419733a2004f417c0"} err="failed to get container status \"fafdd1094bd02786d745c984989852b0c690db38655f92b419733a2004f417c0\": rpc error: code = NotFound desc = could not find container \"fafdd1094bd02786d745c984989852b0c690db38655f92b419733a2004f417c0\": container with ID starting with fafdd1094bd02786d745c984989852b0c690db38655f92b419733a2004f417c0 not found: ID does not exist" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.542869 4885 scope.go:117] "RemoveContainer" containerID="c9f99f9e26a04fb42b01a0783a5c4e360514818ce9102edc7fa08626d50b9a34" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.544780 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c9f99f9e26a04fb42b01a0783a5c4e360514818ce9102edc7fa08626d50b9a34"} err="failed to get container status \"c9f99f9e26a04fb42b01a0783a5c4e360514818ce9102edc7fa08626d50b9a34\": rpc error: code = NotFound desc = could not find container \"c9f99f9e26a04fb42b01a0783a5c4e360514818ce9102edc7fa08626d50b9a34\": container with ID starting with c9f99f9e26a04fb42b01a0783a5c4e360514818ce9102edc7fa08626d50b9a34 not found: ID does not exist" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.544846 4885 scope.go:117] "RemoveContainer" containerID="a6fa72b030d309656b73121e4ac2a3924a4e449aad66683dd9de5ee5f22d5a4e" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.545612 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a6fa72b030d309656b73121e4ac2a3924a4e449aad66683dd9de5ee5f22d5a4e"} err="failed to get container status \"a6fa72b030d309656b73121e4ac2a3924a4e449aad66683dd9de5ee5f22d5a4e\": rpc error: code = NotFound desc = could not find container \"a6fa72b030d309656b73121e4ac2a3924a4e449aad66683dd9de5ee5f22d5a4e\": container with ID starting with a6fa72b030d309656b73121e4ac2a3924a4e449aad66683dd9de5ee5f22d5a4e not found: ID does not exist" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.545652 4885 scope.go:117] "RemoveContainer" containerID="a7642033b0b5995d332ccf7992d6d7e6a5e9c9369ae2d1071ccb2f536938c8f4" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.546252 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a7642033b0b5995d332ccf7992d6d7e6a5e9c9369ae2d1071ccb2f536938c8f4"} err="failed to get container status \"a7642033b0b5995d332ccf7992d6d7e6a5e9c9369ae2d1071ccb2f536938c8f4\": rpc error: code = NotFound desc = could not find container \"a7642033b0b5995d332ccf7992d6d7e6a5e9c9369ae2d1071ccb2f536938c8f4\": container with ID starting with a7642033b0b5995d332ccf7992d6d7e6a5e9c9369ae2d1071ccb2f536938c8f4 not found: ID does not exist" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.546327 4885 scope.go:117] "RemoveContainer" containerID="4310909ad9c7b609ed6814e04a525a6afae747775598e7275080413dec8b23a3" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.549493 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4310909ad9c7b609ed6814e04a525a6afae747775598e7275080413dec8b23a3"} err="failed to get container status \"4310909ad9c7b609ed6814e04a525a6afae747775598e7275080413dec8b23a3\": rpc error: code = NotFound desc = could not find container \"4310909ad9c7b609ed6814e04a525a6afae747775598e7275080413dec8b23a3\": container with ID starting with 4310909ad9c7b609ed6814e04a525a6afae747775598e7275080413dec8b23a3 not found: ID does not exist" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.549551 4885 scope.go:117] "RemoveContainer" containerID="ef403e519d197fb37b3e6b76a14a5128ba3f30a7a5e8bf0a420cf6a39b89575f" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.562310 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef403e519d197fb37b3e6b76a14a5128ba3f30a7a5e8bf0a420cf6a39b89575f"} err="failed to get container status \"ef403e519d197fb37b3e6b76a14a5128ba3f30a7a5e8bf0a420cf6a39b89575f\": rpc error: code = NotFound desc = could not find container \"ef403e519d197fb37b3e6b76a14a5128ba3f30a7a5e8bf0a420cf6a39b89575f\": container with ID starting with ef403e519d197fb37b3e6b76a14a5128ba3f30a7a5e8bf0a420cf6a39b89575f not found: ID does not exist" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.562382 4885 scope.go:117] "RemoveContainer" containerID="8482ba457683ffa03100ec1f07ea9d3f918be5e490b5ba05c2ba965d03033191" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.565775 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8482ba457683ffa03100ec1f07ea9d3f918be5e490b5ba05c2ba965d03033191"} err="failed to get container status \"8482ba457683ffa03100ec1f07ea9d3f918be5e490b5ba05c2ba965d03033191\": rpc error: code = NotFound desc = could not find container \"8482ba457683ffa03100ec1f07ea9d3f918be5e490b5ba05c2ba965d03033191\": container with ID starting with 8482ba457683ffa03100ec1f07ea9d3f918be5e490b5ba05c2ba965d03033191 not found: ID does not exist" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.565814 4885 scope.go:117] "RemoveContainer" containerID="0125e907d061a97286026ee823c4eed303734ed03ab4864f742592b23a1aa2f8" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.566365 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0125e907d061a97286026ee823c4eed303734ed03ab4864f742592b23a1aa2f8"} err="failed to get container status \"0125e907d061a97286026ee823c4eed303734ed03ab4864f742592b23a1aa2f8\": rpc error: code = NotFound desc = could not find container \"0125e907d061a97286026ee823c4eed303734ed03ab4864f742592b23a1aa2f8\": container with ID starting with 0125e907d061a97286026ee823c4eed303734ed03ab4864f742592b23a1aa2f8 not found: ID does not exist" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.566402 4885 scope.go:117] "RemoveContainer" containerID="41121c480bb0f68a539497e7d1018e6582a1cf812f819b4c7327ad8827b316b7" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.566714 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"41121c480bb0f68a539497e7d1018e6582a1cf812f819b4c7327ad8827b316b7"} err="failed to get container status \"41121c480bb0f68a539497e7d1018e6582a1cf812f819b4c7327ad8827b316b7\": rpc error: code = NotFound desc = could not find container \"41121c480bb0f68a539497e7d1018e6582a1cf812f819b4c7327ad8827b316b7\": container with ID starting with 41121c480bb0f68a539497e7d1018e6582a1cf812f819b4c7327ad8827b316b7 not found: ID does not exist" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.566731 4885 scope.go:117] "RemoveContainer" containerID="2ef80e838accc59ad7a579c66f4fcd6881d23fa86fc5922d4b0085d211d5d94b" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.567026 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2ef80e838accc59ad7a579c66f4fcd6881d23fa86fc5922d4b0085d211d5d94b"} err="failed to get container status \"2ef80e838accc59ad7a579c66f4fcd6881d23fa86fc5922d4b0085d211d5d94b\": rpc error: code = NotFound desc = could not find container \"2ef80e838accc59ad7a579c66f4fcd6881d23fa86fc5922d4b0085d211d5d94b\": container with ID starting with 2ef80e838accc59ad7a579c66f4fcd6881d23fa86fc5922d4b0085d211d5d94b not found: ID does not exist" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.567039 4885 scope.go:117] "RemoveContainer" containerID="fafdd1094bd02786d745c984989852b0c690db38655f92b419733a2004f417c0" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.567310 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fafdd1094bd02786d745c984989852b0c690db38655f92b419733a2004f417c0"} err="failed to get container status \"fafdd1094bd02786d745c984989852b0c690db38655f92b419733a2004f417c0\": rpc error: code = NotFound desc = could not find container \"fafdd1094bd02786d745c984989852b0c690db38655f92b419733a2004f417c0\": container with ID starting with fafdd1094bd02786d745c984989852b0c690db38655f92b419733a2004f417c0 not found: ID does not exist" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.567324 4885 scope.go:117] "RemoveContainer" containerID="c9f99f9e26a04fb42b01a0783a5c4e360514818ce9102edc7fa08626d50b9a34" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.567591 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c9f99f9e26a04fb42b01a0783a5c4e360514818ce9102edc7fa08626d50b9a34"} err="failed to get container status \"c9f99f9e26a04fb42b01a0783a5c4e360514818ce9102edc7fa08626d50b9a34\": rpc error: code = NotFound desc = could not find container \"c9f99f9e26a04fb42b01a0783a5c4e360514818ce9102edc7fa08626d50b9a34\": container with ID starting with c9f99f9e26a04fb42b01a0783a5c4e360514818ce9102edc7fa08626d50b9a34 not found: ID does not exist" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.567605 4885 scope.go:117] "RemoveContainer" containerID="a6fa72b030d309656b73121e4ac2a3924a4e449aad66683dd9de5ee5f22d5a4e" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.569315 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a6fa72b030d309656b73121e4ac2a3924a4e449aad66683dd9de5ee5f22d5a4e"} err="failed to get container status \"a6fa72b030d309656b73121e4ac2a3924a4e449aad66683dd9de5ee5f22d5a4e\": rpc error: code = NotFound desc = could not find container \"a6fa72b030d309656b73121e4ac2a3924a4e449aad66683dd9de5ee5f22d5a4e\": container with ID starting with a6fa72b030d309656b73121e4ac2a3924a4e449aad66683dd9de5ee5f22d5a4e not found: ID does not exist" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.569333 4885 scope.go:117] "RemoveContainer" containerID="a7642033b0b5995d332ccf7992d6d7e6a5e9c9369ae2d1071ccb2f536938c8f4" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.569648 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a7642033b0b5995d332ccf7992d6d7e6a5e9c9369ae2d1071ccb2f536938c8f4"} err="failed to get container status \"a7642033b0b5995d332ccf7992d6d7e6a5e9c9369ae2d1071ccb2f536938c8f4\": rpc error: code = NotFound desc = could not find container \"a7642033b0b5995d332ccf7992d6d7e6a5e9c9369ae2d1071ccb2f536938c8f4\": container with ID starting with a7642033b0b5995d332ccf7992d6d7e6a5e9c9369ae2d1071ccb2f536938c8f4 not found: ID does not exist" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.569666 4885 scope.go:117] "RemoveContainer" containerID="4310909ad9c7b609ed6814e04a525a6afae747775598e7275080413dec8b23a3" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.569912 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4310909ad9c7b609ed6814e04a525a6afae747775598e7275080413dec8b23a3"} err="failed to get container status \"4310909ad9c7b609ed6814e04a525a6afae747775598e7275080413dec8b23a3\": rpc error: code = NotFound desc = could not find container \"4310909ad9c7b609ed6814e04a525a6afae747775598e7275080413dec8b23a3\": container with ID starting with 4310909ad9c7b609ed6814e04a525a6afae747775598e7275080413dec8b23a3 not found: ID does not exist" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.569928 4885 scope.go:117] "RemoveContainer" containerID="ef403e519d197fb37b3e6b76a14a5128ba3f30a7a5e8bf0a420cf6a39b89575f" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.570154 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef403e519d197fb37b3e6b76a14a5128ba3f30a7a5e8bf0a420cf6a39b89575f"} err="failed to get container status \"ef403e519d197fb37b3e6b76a14a5128ba3f30a7a5e8bf0a420cf6a39b89575f\": rpc error: code = NotFound desc = could not find container \"ef403e519d197fb37b3e6b76a14a5128ba3f30a7a5e8bf0a420cf6a39b89575f\": container with ID starting with ef403e519d197fb37b3e6b76a14a5128ba3f30a7a5e8bf0a420cf6a39b89575f not found: ID does not exist" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.570171 4885 scope.go:117] "RemoveContainer" containerID="8482ba457683ffa03100ec1f07ea9d3f918be5e490b5ba05c2ba965d03033191" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.570449 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8482ba457683ffa03100ec1f07ea9d3f918be5e490b5ba05c2ba965d03033191"} err="failed to get container status \"8482ba457683ffa03100ec1f07ea9d3f918be5e490b5ba05c2ba965d03033191\": rpc error: code = NotFound desc = could not find container \"8482ba457683ffa03100ec1f07ea9d3f918be5e490b5ba05c2ba965d03033191\": container with ID starting with 8482ba457683ffa03100ec1f07ea9d3f918be5e490b5ba05c2ba965d03033191 not found: ID does not exist" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.570466 4885 scope.go:117] "RemoveContainer" containerID="0125e907d061a97286026ee823c4eed303734ed03ab4864f742592b23a1aa2f8" Oct 02 01:56:40 crc kubenswrapper[4885]: I1002 01:56:40.570783 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0125e907d061a97286026ee823c4eed303734ed03ab4864f742592b23a1aa2f8"} err="failed to get container status \"0125e907d061a97286026ee823c4eed303734ed03ab4864f742592b23a1aa2f8\": rpc error: code = NotFound desc = could not find container \"0125e907d061a97286026ee823c4eed303734ed03ab4864f742592b23a1aa2f8\": container with ID starting with 0125e907d061a97286026ee823c4eed303734ed03ab4864f742592b23a1aa2f8 not found: ID does not exist" Oct 02 01:56:41 crc kubenswrapper[4885]: I1002 01:56:41.118536 4885 generic.go:334] "Generic (PLEG): container finished" podID="c984321f-b4d7-49da-a4ee-2a61bcc84daa" containerID="f03e7d0f57466e3662d251ad3036ccd64a8672d7481d9b2733f780bcc95d9d04" exitCode=0 Oct 02 01:56:41 crc kubenswrapper[4885]: I1002 01:56:41.118699 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x69mt" event={"ID":"c984321f-b4d7-49da-a4ee-2a61bcc84daa","Type":"ContainerDied","Data":"f03e7d0f57466e3662d251ad3036ccd64a8672d7481d9b2733f780bcc95d9d04"} Oct 02 01:56:41 crc kubenswrapper[4885]: I1002 01:56:41.118802 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x69mt" event={"ID":"c984321f-b4d7-49da-a4ee-2a61bcc84daa","Type":"ContainerStarted","Data":"f159c1d9fde170e0e5afa30f44a6f41f60dbc23bb7ee7a53a845c07f16f5a02a"} Oct 02 01:56:41 crc kubenswrapper[4885]: I1002 01:56:41.121692 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-vhgbd_ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f/kube-multus/2.log" Oct 02 01:56:42 crc kubenswrapper[4885]: I1002 01:56:42.062959 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9e4679dd-f870-41e9-a32b-360f5d2ee81b" path="/var/lib/kubelet/pods/9e4679dd-f870-41e9-a32b-360f5d2ee81b/volumes" Oct 02 01:56:42 crc kubenswrapper[4885]: I1002 01:56:42.138889 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x69mt" event={"ID":"c984321f-b4d7-49da-a4ee-2a61bcc84daa","Type":"ContainerStarted","Data":"67c15f806fe3f3505e0d6ecb93cc2f03044a0d9336ed5f21cacb559ca8ca3460"} Oct 02 01:56:42 crc kubenswrapper[4885]: I1002 01:56:42.139219 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x69mt" event={"ID":"c984321f-b4d7-49da-a4ee-2a61bcc84daa","Type":"ContainerStarted","Data":"f1ac29755096e2b7d08a28ae7d33450cd67472b6dacb0ef40a28548e80151945"} Oct 02 01:56:42 crc kubenswrapper[4885]: I1002 01:56:42.139308 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x69mt" event={"ID":"c984321f-b4d7-49da-a4ee-2a61bcc84daa","Type":"ContainerStarted","Data":"cf33e2d7c66837e58b6caf027225badcbe4f6440a4f1057684b3d190105c1cdc"} Oct 02 01:56:43 crc kubenswrapper[4885]: I1002 01:56:43.150727 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x69mt" event={"ID":"c984321f-b4d7-49da-a4ee-2a61bcc84daa","Type":"ContainerStarted","Data":"0edf9a7b4fff1275702b97bdd40d39b707f42da8b0de697d7a8ce1a057836560"} Oct 02 01:56:43 crc kubenswrapper[4885]: I1002 01:56:43.150784 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x69mt" event={"ID":"c984321f-b4d7-49da-a4ee-2a61bcc84daa","Type":"ContainerStarted","Data":"a010aa76108e31313227583d6465b9cae0e2351529c9be3e7668bb47ff8da8c1"} Oct 02 01:56:43 crc kubenswrapper[4885]: I1002 01:56:43.150805 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x69mt" event={"ID":"c984321f-b4d7-49da-a4ee-2a61bcc84daa","Type":"ContainerStarted","Data":"7e1ff3906c814f3e3a1bd3e2764b31c3df8587e502a564471c6efdc601380a71"} Oct 02 01:56:43 crc kubenswrapper[4885]: I1002 01:56:43.265786 4885 patch_prober.go:28] interesting pod/machine-config-daemon-rttx8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 01:56:43 crc kubenswrapper[4885]: I1002 01:56:43.265896 4885 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 01:56:43 crc kubenswrapper[4885]: I1002 01:56:43.266007 4885 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" Oct 02 01:56:43 crc kubenswrapper[4885]: I1002 01:56:43.267140 4885 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a1c63629d5a2ca3023ebf7324c43c17375f971212dce224761d6f69e9b4ea0ce"} pod="openshift-machine-config-operator/machine-config-daemon-rttx8" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 01:56:43 crc kubenswrapper[4885]: I1002 01:56:43.267307 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" containerName="machine-config-daemon" containerID="cri-o://a1c63629d5a2ca3023ebf7324c43c17375f971212dce224761d6f69e9b4ea0ce" gracePeriod=600 Oct 02 01:56:43 crc kubenswrapper[4885]: I1002 01:56:43.820015 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-vh2n2" Oct 02 01:56:44 crc kubenswrapper[4885]: I1002 01:56:44.161706 4885 generic.go:334] "Generic (PLEG): container finished" podID="d169c8ba-000a-4218-8e3b-4ae53035b110" containerID="a1c63629d5a2ca3023ebf7324c43c17375f971212dce224761d6f69e9b4ea0ce" exitCode=0 Oct 02 01:56:44 crc kubenswrapper[4885]: I1002 01:56:44.161770 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" event={"ID":"d169c8ba-000a-4218-8e3b-4ae53035b110","Type":"ContainerDied","Data":"a1c63629d5a2ca3023ebf7324c43c17375f971212dce224761d6f69e9b4ea0ce"} Oct 02 01:56:44 crc kubenswrapper[4885]: I1002 01:56:44.161826 4885 scope.go:117] "RemoveContainer" containerID="edcfaf9e26fcac2c575da14a9bd0f19b12460248f9779d8e37fc41a2db9b606d" Oct 02 01:56:45 crc kubenswrapper[4885]: I1002 01:56:45.189691 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" event={"ID":"d169c8ba-000a-4218-8e3b-4ae53035b110","Type":"ContainerStarted","Data":"83cae60dd6556fc08f2eddf3fd26fd0abb56b3d0a4d62534298ea5f48fc39e18"} Oct 02 01:56:46 crc kubenswrapper[4885]: I1002 01:56:46.202021 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x69mt" event={"ID":"c984321f-b4d7-49da-a4ee-2a61bcc84daa","Type":"ContainerStarted","Data":"5389eafa1670cac22b9ecd1915150be199465516d8ca91fc29aa3e570caa9c53"} Oct 02 01:56:48 crc kubenswrapper[4885]: I1002 01:56:48.223721 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x69mt" event={"ID":"c984321f-b4d7-49da-a4ee-2a61bcc84daa","Type":"ContainerStarted","Data":"d869f93d67b02d804d19bb8507cc80bbaff788e74dec52a9fab9ecd54606e221"} Oct 02 01:56:48 crc kubenswrapper[4885]: I1002 01:56:48.224051 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-x69mt" Oct 02 01:56:48 crc kubenswrapper[4885]: I1002 01:56:48.224077 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-x69mt" Oct 02 01:56:48 crc kubenswrapper[4885]: I1002 01:56:48.224096 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-x69mt" Oct 02 01:56:48 crc kubenswrapper[4885]: I1002 01:56:48.256117 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-x69mt" Oct 02 01:56:48 crc kubenswrapper[4885]: I1002 01:56:48.272234 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-x69mt" podStartSLOduration=9.272215434 podStartE2EDuration="9.272215434s" podCreationTimestamp="2025-10-02 01:56:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:56:48.271444281 +0000 UTC m=+597.083191720" watchObservedRunningTime="2025-10-02 01:56:48.272215434 +0000 UTC m=+597.083962843" Oct 02 01:56:48 crc kubenswrapper[4885]: I1002 01:56:48.278365 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-x69mt" Oct 02 01:56:55 crc kubenswrapper[4885]: I1002 01:56:55.047327 4885 scope.go:117] "RemoveContainer" containerID="7476bbb9bcb8915493b4060254e1303bd96e825ec40425de522dce0a357fb6e6" Oct 02 01:56:55 crc kubenswrapper[4885]: E1002 01:56:55.049713 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-vhgbd_openshift-multus(ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f)\"" pod="openshift-multus/multus-vhgbd" podUID="ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f" Oct 02 01:57:10 crc kubenswrapper[4885]: I1002 01:57:10.047255 4885 scope.go:117] "RemoveContainer" containerID="7476bbb9bcb8915493b4060254e1303bd96e825ec40425de522dce0a357fb6e6" Oct 02 01:57:10 crc kubenswrapper[4885]: I1002 01:57:10.359587 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-x69mt" Oct 02 01:57:10 crc kubenswrapper[4885]: I1002 01:57:10.384850 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-vhgbd_ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f/kube-multus/2.log" Oct 02 01:57:10 crc kubenswrapper[4885]: I1002 01:57:10.384922 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-vhgbd" event={"ID":"ecfe7ad8-f68d-48fd-bc04-8f7a5f9d447f","Type":"ContainerStarted","Data":"faf0ec60ac006f0cd28f142fffe85f6fb14cfc1b7a0b46a30fca45b53b9c3529"} Oct 02 01:57:32 crc kubenswrapper[4885]: I1002 01:57:32.884519 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ctlhn5"] Oct 02 01:57:32 crc kubenswrapper[4885]: I1002 01:57:32.886777 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ctlhn5" Oct 02 01:57:32 crc kubenswrapper[4885]: I1002 01:57:32.889507 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 02 01:57:32 crc kubenswrapper[4885]: I1002 01:57:32.898402 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ctlhn5"] Oct 02 01:57:33 crc kubenswrapper[4885]: I1002 01:57:33.058148 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/533f5d4b-02ed-4235-8eed-715dd68a58fd-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ctlhn5\" (UID: \"533f5d4b-02ed-4235-8eed-715dd68a58fd\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ctlhn5" Oct 02 01:57:33 crc kubenswrapper[4885]: I1002 01:57:33.058236 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5ljmq\" (UniqueName: \"kubernetes.io/projected/533f5d4b-02ed-4235-8eed-715dd68a58fd-kube-api-access-5ljmq\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ctlhn5\" (UID: \"533f5d4b-02ed-4235-8eed-715dd68a58fd\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ctlhn5" Oct 02 01:57:33 crc kubenswrapper[4885]: I1002 01:57:33.058318 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/533f5d4b-02ed-4235-8eed-715dd68a58fd-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ctlhn5\" (UID: \"533f5d4b-02ed-4235-8eed-715dd68a58fd\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ctlhn5" Oct 02 01:57:33 crc kubenswrapper[4885]: I1002 01:57:33.159540 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5ljmq\" (UniqueName: \"kubernetes.io/projected/533f5d4b-02ed-4235-8eed-715dd68a58fd-kube-api-access-5ljmq\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ctlhn5\" (UID: \"533f5d4b-02ed-4235-8eed-715dd68a58fd\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ctlhn5" Oct 02 01:57:33 crc kubenswrapper[4885]: I1002 01:57:33.159629 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/533f5d4b-02ed-4235-8eed-715dd68a58fd-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ctlhn5\" (UID: \"533f5d4b-02ed-4235-8eed-715dd68a58fd\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ctlhn5" Oct 02 01:57:33 crc kubenswrapper[4885]: I1002 01:57:33.159759 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/533f5d4b-02ed-4235-8eed-715dd68a58fd-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ctlhn5\" (UID: \"533f5d4b-02ed-4235-8eed-715dd68a58fd\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ctlhn5" Oct 02 01:57:33 crc kubenswrapper[4885]: I1002 01:57:33.160535 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/533f5d4b-02ed-4235-8eed-715dd68a58fd-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ctlhn5\" (UID: \"533f5d4b-02ed-4235-8eed-715dd68a58fd\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ctlhn5" Oct 02 01:57:33 crc kubenswrapper[4885]: I1002 01:57:33.160650 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/533f5d4b-02ed-4235-8eed-715dd68a58fd-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ctlhn5\" (UID: \"533f5d4b-02ed-4235-8eed-715dd68a58fd\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ctlhn5" Oct 02 01:57:33 crc kubenswrapper[4885]: I1002 01:57:33.191814 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5ljmq\" (UniqueName: \"kubernetes.io/projected/533f5d4b-02ed-4235-8eed-715dd68a58fd-kube-api-access-5ljmq\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ctlhn5\" (UID: \"533f5d4b-02ed-4235-8eed-715dd68a58fd\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ctlhn5" Oct 02 01:57:33 crc kubenswrapper[4885]: I1002 01:57:33.209901 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ctlhn5" Oct 02 01:57:33 crc kubenswrapper[4885]: I1002 01:57:33.723478 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ctlhn5"] Oct 02 01:57:34 crc kubenswrapper[4885]: I1002 01:57:34.548687 4885 generic.go:334] "Generic (PLEG): container finished" podID="533f5d4b-02ed-4235-8eed-715dd68a58fd" containerID="e3273053c5cfdcfaa59e4775a066a76c14b597f6f96774818013f9e658cccbfb" exitCode=0 Oct 02 01:57:34 crc kubenswrapper[4885]: I1002 01:57:34.548758 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ctlhn5" event={"ID":"533f5d4b-02ed-4235-8eed-715dd68a58fd","Type":"ContainerDied","Data":"e3273053c5cfdcfaa59e4775a066a76c14b597f6f96774818013f9e658cccbfb"} Oct 02 01:57:34 crc kubenswrapper[4885]: I1002 01:57:34.548800 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ctlhn5" event={"ID":"533f5d4b-02ed-4235-8eed-715dd68a58fd","Type":"ContainerStarted","Data":"d37973266db7fce040cd3ca23c368457102c69e0460f4ae03ae732d404588911"} Oct 02 01:57:36 crc kubenswrapper[4885]: I1002 01:57:36.564754 4885 generic.go:334] "Generic (PLEG): container finished" podID="533f5d4b-02ed-4235-8eed-715dd68a58fd" containerID="a2510d42ffae8c12057369e825d565c485fe6725814a3492b8e34305632e66ff" exitCode=0 Oct 02 01:57:36 crc kubenswrapper[4885]: I1002 01:57:36.564839 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ctlhn5" event={"ID":"533f5d4b-02ed-4235-8eed-715dd68a58fd","Type":"ContainerDied","Data":"a2510d42ffae8c12057369e825d565c485fe6725814a3492b8e34305632e66ff"} Oct 02 01:57:37 crc kubenswrapper[4885]: I1002 01:57:37.573945 4885 generic.go:334] "Generic (PLEG): container finished" podID="533f5d4b-02ed-4235-8eed-715dd68a58fd" containerID="9d56e6e663d7f7110bad43dcdcd41424aef8009bcbd010c5f84a93ef5e4af389" exitCode=0 Oct 02 01:57:37 crc kubenswrapper[4885]: I1002 01:57:37.574033 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ctlhn5" event={"ID":"533f5d4b-02ed-4235-8eed-715dd68a58fd","Type":"ContainerDied","Data":"9d56e6e663d7f7110bad43dcdcd41424aef8009bcbd010c5f84a93ef5e4af389"} Oct 02 01:57:38 crc kubenswrapper[4885]: I1002 01:57:38.912081 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ctlhn5" Oct 02 01:57:39 crc kubenswrapper[4885]: I1002 01:57:39.050156 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/533f5d4b-02ed-4235-8eed-715dd68a58fd-util\") pod \"533f5d4b-02ed-4235-8eed-715dd68a58fd\" (UID: \"533f5d4b-02ed-4235-8eed-715dd68a58fd\") " Oct 02 01:57:39 crc kubenswrapper[4885]: I1002 01:57:39.050223 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5ljmq\" (UniqueName: \"kubernetes.io/projected/533f5d4b-02ed-4235-8eed-715dd68a58fd-kube-api-access-5ljmq\") pod \"533f5d4b-02ed-4235-8eed-715dd68a58fd\" (UID: \"533f5d4b-02ed-4235-8eed-715dd68a58fd\") " Oct 02 01:57:39 crc kubenswrapper[4885]: I1002 01:57:39.050343 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/533f5d4b-02ed-4235-8eed-715dd68a58fd-bundle\") pod \"533f5d4b-02ed-4235-8eed-715dd68a58fd\" (UID: \"533f5d4b-02ed-4235-8eed-715dd68a58fd\") " Oct 02 01:57:39 crc kubenswrapper[4885]: I1002 01:57:39.051347 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/533f5d4b-02ed-4235-8eed-715dd68a58fd-bundle" (OuterVolumeSpecName: "bundle") pod "533f5d4b-02ed-4235-8eed-715dd68a58fd" (UID: "533f5d4b-02ed-4235-8eed-715dd68a58fd"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 01:57:39 crc kubenswrapper[4885]: I1002 01:57:39.058829 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/533f5d4b-02ed-4235-8eed-715dd68a58fd-kube-api-access-5ljmq" (OuterVolumeSpecName: "kube-api-access-5ljmq") pod "533f5d4b-02ed-4235-8eed-715dd68a58fd" (UID: "533f5d4b-02ed-4235-8eed-715dd68a58fd"). InnerVolumeSpecName "kube-api-access-5ljmq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:57:39 crc kubenswrapper[4885]: I1002 01:57:39.081561 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/533f5d4b-02ed-4235-8eed-715dd68a58fd-util" (OuterVolumeSpecName: "util") pod "533f5d4b-02ed-4235-8eed-715dd68a58fd" (UID: "533f5d4b-02ed-4235-8eed-715dd68a58fd"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 01:57:39 crc kubenswrapper[4885]: I1002 01:57:39.152433 4885 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/533f5d4b-02ed-4235-8eed-715dd68a58fd-util\") on node \"crc\" DevicePath \"\"" Oct 02 01:57:39 crc kubenswrapper[4885]: I1002 01:57:39.152480 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5ljmq\" (UniqueName: \"kubernetes.io/projected/533f5d4b-02ed-4235-8eed-715dd68a58fd-kube-api-access-5ljmq\") on node \"crc\" DevicePath \"\"" Oct 02 01:57:39 crc kubenswrapper[4885]: I1002 01:57:39.152501 4885 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/533f5d4b-02ed-4235-8eed-715dd68a58fd-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 01:57:39 crc kubenswrapper[4885]: I1002 01:57:39.588359 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ctlhn5" event={"ID":"533f5d4b-02ed-4235-8eed-715dd68a58fd","Type":"ContainerDied","Data":"d37973266db7fce040cd3ca23c368457102c69e0460f4ae03ae732d404588911"} Oct 02 01:57:39 crc kubenswrapper[4885]: I1002 01:57:39.588687 4885 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d37973266db7fce040cd3ca23c368457102c69e0460f4ae03ae732d404588911" Oct 02 01:57:39 crc kubenswrapper[4885]: I1002 01:57:39.588423 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ctlhn5" Oct 02 01:57:41 crc kubenswrapper[4885]: I1002 01:57:41.506206 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-r4dhw"] Oct 02 01:57:41 crc kubenswrapper[4885]: E1002 01:57:41.506738 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="533f5d4b-02ed-4235-8eed-715dd68a58fd" containerName="util" Oct 02 01:57:41 crc kubenswrapper[4885]: I1002 01:57:41.506749 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="533f5d4b-02ed-4235-8eed-715dd68a58fd" containerName="util" Oct 02 01:57:41 crc kubenswrapper[4885]: E1002 01:57:41.506762 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="533f5d4b-02ed-4235-8eed-715dd68a58fd" containerName="pull" Oct 02 01:57:41 crc kubenswrapper[4885]: I1002 01:57:41.506767 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="533f5d4b-02ed-4235-8eed-715dd68a58fd" containerName="pull" Oct 02 01:57:41 crc kubenswrapper[4885]: E1002 01:57:41.506776 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="533f5d4b-02ed-4235-8eed-715dd68a58fd" containerName="extract" Oct 02 01:57:41 crc kubenswrapper[4885]: I1002 01:57:41.506782 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="533f5d4b-02ed-4235-8eed-715dd68a58fd" containerName="extract" Oct 02 01:57:41 crc kubenswrapper[4885]: I1002 01:57:41.506864 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="533f5d4b-02ed-4235-8eed-715dd68a58fd" containerName="extract" Oct 02 01:57:41 crc kubenswrapper[4885]: I1002 01:57:41.507176 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-r4dhw" Oct 02 01:57:41 crc kubenswrapper[4885]: I1002 01:57:41.508813 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Oct 02 01:57:41 crc kubenswrapper[4885]: I1002 01:57:41.509155 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Oct 02 01:57:41 crc kubenswrapper[4885]: I1002 01:57:41.510912 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-9pdlz" Oct 02 01:57:41 crc kubenswrapper[4885]: I1002 01:57:41.521821 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-r4dhw"] Oct 02 01:57:41 crc kubenswrapper[4885]: I1002 01:57:41.527532 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gsvjs\" (UniqueName: \"kubernetes.io/projected/6f1fd2a7-4a15-46b4-818f-c24f9fcd2bcd-kube-api-access-gsvjs\") pod \"nmstate-operator-858ddd8f98-r4dhw\" (UID: \"6f1fd2a7-4a15-46b4-818f-c24f9fcd2bcd\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-r4dhw" Oct 02 01:57:41 crc kubenswrapper[4885]: I1002 01:57:41.628891 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gsvjs\" (UniqueName: \"kubernetes.io/projected/6f1fd2a7-4a15-46b4-818f-c24f9fcd2bcd-kube-api-access-gsvjs\") pod \"nmstate-operator-858ddd8f98-r4dhw\" (UID: \"6f1fd2a7-4a15-46b4-818f-c24f9fcd2bcd\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-r4dhw" Oct 02 01:57:41 crc kubenswrapper[4885]: I1002 01:57:41.647790 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gsvjs\" (UniqueName: \"kubernetes.io/projected/6f1fd2a7-4a15-46b4-818f-c24f9fcd2bcd-kube-api-access-gsvjs\") pod \"nmstate-operator-858ddd8f98-r4dhw\" (UID: \"6f1fd2a7-4a15-46b4-818f-c24f9fcd2bcd\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-r4dhw" Oct 02 01:57:41 crc kubenswrapper[4885]: I1002 01:57:41.821629 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-r4dhw" Oct 02 01:57:41 crc kubenswrapper[4885]: I1002 01:57:41.997978 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-r4dhw"] Oct 02 01:57:42 crc kubenswrapper[4885]: I1002 01:57:42.607699 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-858ddd8f98-r4dhw" event={"ID":"6f1fd2a7-4a15-46b4-818f-c24f9fcd2bcd","Type":"ContainerStarted","Data":"363a0b8942efbcbd1a926b071e2efdc8b352d3fee3edbdf2b9608b30fd6bed1a"} Oct 02 01:57:47 crc kubenswrapper[4885]: I1002 01:57:47.638024 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-858ddd8f98-r4dhw" event={"ID":"6f1fd2a7-4a15-46b4-818f-c24f9fcd2bcd","Type":"ContainerStarted","Data":"bdbcb418e2c338bf4bd2852c0b3ce279f7254c4e66e87f828eb6af9a9e254805"} Oct 02 01:57:47 crc kubenswrapper[4885]: I1002 01:57:47.664510 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-858ddd8f98-r4dhw" podStartSLOduration=2.476372929 podStartE2EDuration="6.664492822s" podCreationTimestamp="2025-10-02 01:57:41 +0000 UTC" firstStartedPulling="2025-10-02 01:57:42.005382103 +0000 UTC m=+650.817129492" lastFinishedPulling="2025-10-02 01:57:46.193501976 +0000 UTC m=+655.005249385" observedRunningTime="2025-10-02 01:57:47.660839111 +0000 UTC m=+656.472586510" watchObservedRunningTime="2025-10-02 01:57:47.664492822 +0000 UTC m=+656.476240231" Oct 02 01:57:48 crc kubenswrapper[4885]: I1002 01:57:48.613611 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-sr7mv"] Oct 02 01:57:48 crc kubenswrapper[4885]: I1002 01:57:48.614559 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-sr7mv" Oct 02 01:57:48 crc kubenswrapper[4885]: I1002 01:57:48.622198 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-jk45f"] Oct 02 01:57:48 crc kubenswrapper[4885]: I1002 01:57:48.623087 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-jk45f" Oct 02 01:57:48 crc kubenswrapper[4885]: I1002 01:57:48.623711 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/f75137b4-4551-43cd-b20c-6044d13d27a1-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-jk45f\" (UID: \"f75137b4-4551-43cd-b20c-6044d13d27a1\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-jk45f" Oct 02 01:57:48 crc kubenswrapper[4885]: I1002 01:57:48.623744 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-txhrb\" (UniqueName: \"kubernetes.io/projected/f75137b4-4551-43cd-b20c-6044d13d27a1-kube-api-access-txhrb\") pod \"nmstate-webhook-6cdbc54649-jk45f\" (UID: \"f75137b4-4551-43cd-b20c-6044d13d27a1\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-jk45f" Oct 02 01:57:48 crc kubenswrapper[4885]: I1002 01:57:48.623887 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n5r4n\" (UniqueName: \"kubernetes.io/projected/c2a4785b-6378-4f85-ae1d-46e27903af36-kube-api-access-n5r4n\") pod \"nmstate-metrics-fdff9cb8d-sr7mv\" (UID: \"c2a4785b-6378-4f85-ae1d-46e27903af36\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-sr7mv" Oct 02 01:57:48 crc kubenswrapper[4885]: I1002 01:57:48.624500 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-d6tsn" Oct 02 01:57:48 crc kubenswrapper[4885]: I1002 01:57:48.624826 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Oct 02 01:57:48 crc kubenswrapper[4885]: I1002 01:57:48.634243 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-jk45f"] Oct 02 01:57:48 crc kubenswrapper[4885]: I1002 01:57:48.672428 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-sr7mv"] Oct 02 01:57:48 crc kubenswrapper[4885]: I1002 01:57:48.688125 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-4xzpg"] Oct 02 01:57:48 crc kubenswrapper[4885]: I1002 01:57:48.689785 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-4xzpg" Oct 02 01:57:48 crc kubenswrapper[4885]: I1002 01:57:48.725607 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/4376783f-66f6-41a4-a42e-2666f1cc9978-nmstate-lock\") pod \"nmstate-handler-4xzpg\" (UID: \"4376783f-66f6-41a4-a42e-2666f1cc9978\") " pod="openshift-nmstate/nmstate-handler-4xzpg" Oct 02 01:57:48 crc kubenswrapper[4885]: I1002 01:57:48.725667 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/4376783f-66f6-41a4-a42e-2666f1cc9978-ovs-socket\") pod \"nmstate-handler-4xzpg\" (UID: \"4376783f-66f6-41a4-a42e-2666f1cc9978\") " pod="openshift-nmstate/nmstate-handler-4xzpg" Oct 02 01:57:48 crc kubenswrapper[4885]: I1002 01:57:48.725683 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b6dtq\" (UniqueName: \"kubernetes.io/projected/4376783f-66f6-41a4-a42e-2666f1cc9978-kube-api-access-b6dtq\") pod \"nmstate-handler-4xzpg\" (UID: \"4376783f-66f6-41a4-a42e-2666f1cc9978\") " pod="openshift-nmstate/nmstate-handler-4xzpg" Oct 02 01:57:48 crc kubenswrapper[4885]: I1002 01:57:48.725712 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/f75137b4-4551-43cd-b20c-6044d13d27a1-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-jk45f\" (UID: \"f75137b4-4551-43cd-b20c-6044d13d27a1\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-jk45f" Oct 02 01:57:48 crc kubenswrapper[4885]: I1002 01:57:48.725727 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-txhrb\" (UniqueName: \"kubernetes.io/projected/f75137b4-4551-43cd-b20c-6044d13d27a1-kube-api-access-txhrb\") pod \"nmstate-webhook-6cdbc54649-jk45f\" (UID: \"f75137b4-4551-43cd-b20c-6044d13d27a1\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-jk45f" Oct 02 01:57:48 crc kubenswrapper[4885]: I1002 01:57:48.725743 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/4376783f-66f6-41a4-a42e-2666f1cc9978-dbus-socket\") pod \"nmstate-handler-4xzpg\" (UID: \"4376783f-66f6-41a4-a42e-2666f1cc9978\") " pod="openshift-nmstate/nmstate-handler-4xzpg" Oct 02 01:57:48 crc kubenswrapper[4885]: I1002 01:57:48.725773 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n5r4n\" (UniqueName: \"kubernetes.io/projected/c2a4785b-6378-4f85-ae1d-46e27903af36-kube-api-access-n5r4n\") pod \"nmstate-metrics-fdff9cb8d-sr7mv\" (UID: \"c2a4785b-6378-4f85-ae1d-46e27903af36\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-sr7mv" Oct 02 01:57:48 crc kubenswrapper[4885]: E1002 01:57:48.726531 4885 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Oct 02 01:57:48 crc kubenswrapper[4885]: E1002 01:57:48.726596 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f75137b4-4551-43cd-b20c-6044d13d27a1-tls-key-pair podName:f75137b4-4551-43cd-b20c-6044d13d27a1 nodeName:}" failed. No retries permitted until 2025-10-02 01:57:49.226577275 +0000 UTC m=+658.038324674 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/f75137b4-4551-43cd-b20c-6044d13d27a1-tls-key-pair") pod "nmstate-webhook-6cdbc54649-jk45f" (UID: "f75137b4-4551-43cd-b20c-6044d13d27a1") : secret "openshift-nmstate-webhook" not found Oct 02 01:57:48 crc kubenswrapper[4885]: I1002 01:57:48.746000 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n5r4n\" (UniqueName: \"kubernetes.io/projected/c2a4785b-6378-4f85-ae1d-46e27903af36-kube-api-access-n5r4n\") pod \"nmstate-metrics-fdff9cb8d-sr7mv\" (UID: \"c2a4785b-6378-4f85-ae1d-46e27903af36\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-sr7mv" Oct 02 01:57:48 crc kubenswrapper[4885]: I1002 01:57:48.746623 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-txhrb\" (UniqueName: \"kubernetes.io/projected/f75137b4-4551-43cd-b20c-6044d13d27a1-kube-api-access-txhrb\") pod \"nmstate-webhook-6cdbc54649-jk45f\" (UID: \"f75137b4-4551-43cd-b20c-6044d13d27a1\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-jk45f" Oct 02 01:57:48 crc kubenswrapper[4885]: I1002 01:57:48.792895 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-dvlk4"] Oct 02 01:57:48 crc kubenswrapper[4885]: I1002 01:57:48.793623 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-dvlk4" Oct 02 01:57:48 crc kubenswrapper[4885]: I1002 01:57:48.798744 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-dnx8q" Oct 02 01:57:48 crc kubenswrapper[4885]: I1002 01:57:48.799248 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Oct 02 01:57:48 crc kubenswrapper[4885]: I1002 01:57:48.800385 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Oct 02 01:57:48 crc kubenswrapper[4885]: I1002 01:57:48.811058 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-dvlk4"] Oct 02 01:57:48 crc kubenswrapper[4885]: I1002 01:57:48.826912 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/4376783f-66f6-41a4-a42e-2666f1cc9978-nmstate-lock\") pod \"nmstate-handler-4xzpg\" (UID: \"4376783f-66f6-41a4-a42e-2666f1cc9978\") " pod="openshift-nmstate/nmstate-handler-4xzpg" Oct 02 01:57:48 crc kubenswrapper[4885]: I1002 01:57:48.826972 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/4376783f-66f6-41a4-a42e-2666f1cc9978-ovs-socket\") pod \"nmstate-handler-4xzpg\" (UID: \"4376783f-66f6-41a4-a42e-2666f1cc9978\") " pod="openshift-nmstate/nmstate-handler-4xzpg" Oct 02 01:57:48 crc kubenswrapper[4885]: I1002 01:57:48.826995 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b6dtq\" (UniqueName: \"kubernetes.io/projected/4376783f-66f6-41a4-a42e-2666f1cc9978-kube-api-access-b6dtq\") pod \"nmstate-handler-4xzpg\" (UID: \"4376783f-66f6-41a4-a42e-2666f1cc9978\") " pod="openshift-nmstate/nmstate-handler-4xzpg" Oct 02 01:57:48 crc kubenswrapper[4885]: I1002 01:57:48.827036 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/4376783f-66f6-41a4-a42e-2666f1cc9978-dbus-socket\") pod \"nmstate-handler-4xzpg\" (UID: \"4376783f-66f6-41a4-a42e-2666f1cc9978\") " pod="openshift-nmstate/nmstate-handler-4xzpg" Oct 02 01:57:48 crc kubenswrapper[4885]: I1002 01:57:48.827301 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/4376783f-66f6-41a4-a42e-2666f1cc9978-dbus-socket\") pod \"nmstate-handler-4xzpg\" (UID: \"4376783f-66f6-41a4-a42e-2666f1cc9978\") " pod="openshift-nmstate/nmstate-handler-4xzpg" Oct 02 01:57:48 crc kubenswrapper[4885]: I1002 01:57:48.827343 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/4376783f-66f6-41a4-a42e-2666f1cc9978-ovs-socket\") pod \"nmstate-handler-4xzpg\" (UID: \"4376783f-66f6-41a4-a42e-2666f1cc9978\") " pod="openshift-nmstate/nmstate-handler-4xzpg" Oct 02 01:57:48 crc kubenswrapper[4885]: I1002 01:57:48.827368 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/4376783f-66f6-41a4-a42e-2666f1cc9978-nmstate-lock\") pod \"nmstate-handler-4xzpg\" (UID: \"4376783f-66f6-41a4-a42e-2666f1cc9978\") " pod="openshift-nmstate/nmstate-handler-4xzpg" Oct 02 01:57:48 crc kubenswrapper[4885]: I1002 01:57:48.849984 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b6dtq\" (UniqueName: \"kubernetes.io/projected/4376783f-66f6-41a4-a42e-2666f1cc9978-kube-api-access-b6dtq\") pod \"nmstate-handler-4xzpg\" (UID: \"4376783f-66f6-41a4-a42e-2666f1cc9978\") " pod="openshift-nmstate/nmstate-handler-4xzpg" Oct 02 01:57:48 crc kubenswrapper[4885]: I1002 01:57:48.928243 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/c7f8ff02-de9d-456d-ba5e-a794f568c5cf-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-dvlk4\" (UID: \"c7f8ff02-de9d-456d-ba5e-a794f568c5cf\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-dvlk4" Oct 02 01:57:48 crc kubenswrapper[4885]: I1002 01:57:48.928357 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4fwh8\" (UniqueName: \"kubernetes.io/projected/c7f8ff02-de9d-456d-ba5e-a794f568c5cf-kube-api-access-4fwh8\") pod \"nmstate-console-plugin-6b874cbd85-dvlk4\" (UID: \"c7f8ff02-de9d-456d-ba5e-a794f568c5cf\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-dvlk4" Oct 02 01:57:48 crc kubenswrapper[4885]: I1002 01:57:48.928470 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/c7f8ff02-de9d-456d-ba5e-a794f568c5cf-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-dvlk4\" (UID: \"c7f8ff02-de9d-456d-ba5e-a794f568c5cf\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-dvlk4" Oct 02 01:57:48 crc kubenswrapper[4885]: I1002 01:57:48.973180 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-sr7mv" Oct 02 01:57:48 crc kubenswrapper[4885]: I1002 01:57:48.980187 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-77b75945b8-kqms5"] Oct 02 01:57:48 crc kubenswrapper[4885]: I1002 01:57:48.981060 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-77b75945b8-kqms5" Oct 02 01:57:48 crc kubenswrapper[4885]: I1002 01:57:48.996717 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-77b75945b8-kqms5"] Oct 02 01:57:49 crc kubenswrapper[4885]: I1002 01:57:49.025730 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-4xzpg" Oct 02 01:57:49 crc kubenswrapper[4885]: I1002 01:57:49.030898 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/d45824eb-fa61-4a06-8b9c-d0fc6a7adb69-oauth-serving-cert\") pod \"console-77b75945b8-kqms5\" (UID: \"d45824eb-fa61-4a06-8b9c-d0fc6a7adb69\") " pod="openshift-console/console-77b75945b8-kqms5" Oct 02 01:57:49 crc kubenswrapper[4885]: I1002 01:57:49.030944 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/d45824eb-fa61-4a06-8b9c-d0fc6a7adb69-console-serving-cert\") pod \"console-77b75945b8-kqms5\" (UID: \"d45824eb-fa61-4a06-8b9c-d0fc6a7adb69\") " pod="openshift-console/console-77b75945b8-kqms5" Oct 02 01:57:49 crc kubenswrapper[4885]: I1002 01:57:49.031007 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d45824eb-fa61-4a06-8b9c-d0fc6a7adb69-trusted-ca-bundle\") pod \"console-77b75945b8-kqms5\" (UID: \"d45824eb-fa61-4a06-8b9c-d0fc6a7adb69\") " pod="openshift-console/console-77b75945b8-kqms5" Oct 02 01:57:49 crc kubenswrapper[4885]: I1002 01:57:49.031038 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d45824eb-fa61-4a06-8b9c-d0fc6a7adb69-service-ca\") pod \"console-77b75945b8-kqms5\" (UID: \"d45824eb-fa61-4a06-8b9c-d0fc6a7adb69\") " pod="openshift-console/console-77b75945b8-kqms5" Oct 02 01:57:49 crc kubenswrapper[4885]: I1002 01:57:49.031082 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/c7f8ff02-de9d-456d-ba5e-a794f568c5cf-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-dvlk4\" (UID: \"c7f8ff02-de9d-456d-ba5e-a794f568c5cf\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-dvlk4" Oct 02 01:57:49 crc kubenswrapper[4885]: I1002 01:57:49.031127 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/d45824eb-fa61-4a06-8b9c-d0fc6a7adb69-console-config\") pod \"console-77b75945b8-kqms5\" (UID: \"d45824eb-fa61-4a06-8b9c-d0fc6a7adb69\") " pod="openshift-console/console-77b75945b8-kqms5" Oct 02 01:57:49 crc kubenswrapper[4885]: I1002 01:57:49.031162 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/d45824eb-fa61-4a06-8b9c-d0fc6a7adb69-console-oauth-config\") pod \"console-77b75945b8-kqms5\" (UID: \"d45824eb-fa61-4a06-8b9c-d0fc6a7adb69\") " pod="openshift-console/console-77b75945b8-kqms5" Oct 02 01:57:49 crc kubenswrapper[4885]: I1002 01:57:49.031206 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/c7f8ff02-de9d-456d-ba5e-a794f568c5cf-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-dvlk4\" (UID: \"c7f8ff02-de9d-456d-ba5e-a794f568c5cf\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-dvlk4" Oct 02 01:57:49 crc kubenswrapper[4885]: I1002 01:57:49.031239 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4fwh8\" (UniqueName: \"kubernetes.io/projected/c7f8ff02-de9d-456d-ba5e-a794f568c5cf-kube-api-access-4fwh8\") pod \"nmstate-console-plugin-6b874cbd85-dvlk4\" (UID: \"c7f8ff02-de9d-456d-ba5e-a794f568c5cf\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-dvlk4" Oct 02 01:57:49 crc kubenswrapper[4885]: I1002 01:57:49.031282 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vprdw\" (UniqueName: \"kubernetes.io/projected/d45824eb-fa61-4a06-8b9c-d0fc6a7adb69-kube-api-access-vprdw\") pod \"console-77b75945b8-kqms5\" (UID: \"d45824eb-fa61-4a06-8b9c-d0fc6a7adb69\") " pod="openshift-console/console-77b75945b8-kqms5" Oct 02 01:57:49 crc kubenswrapper[4885]: I1002 01:57:49.032303 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/c7f8ff02-de9d-456d-ba5e-a794f568c5cf-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-dvlk4\" (UID: \"c7f8ff02-de9d-456d-ba5e-a794f568c5cf\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-dvlk4" Oct 02 01:57:49 crc kubenswrapper[4885]: I1002 01:57:49.035210 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/c7f8ff02-de9d-456d-ba5e-a794f568c5cf-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-dvlk4\" (UID: \"c7f8ff02-de9d-456d-ba5e-a794f568c5cf\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-dvlk4" Oct 02 01:57:49 crc kubenswrapper[4885]: I1002 01:57:49.046380 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4fwh8\" (UniqueName: \"kubernetes.io/projected/c7f8ff02-de9d-456d-ba5e-a794f568c5cf-kube-api-access-4fwh8\") pod \"nmstate-console-plugin-6b874cbd85-dvlk4\" (UID: \"c7f8ff02-de9d-456d-ba5e-a794f568c5cf\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-dvlk4" Oct 02 01:57:49 crc kubenswrapper[4885]: I1002 01:57:49.109907 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-dvlk4" Oct 02 01:57:49 crc kubenswrapper[4885]: I1002 01:57:49.132111 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vprdw\" (UniqueName: \"kubernetes.io/projected/d45824eb-fa61-4a06-8b9c-d0fc6a7adb69-kube-api-access-vprdw\") pod \"console-77b75945b8-kqms5\" (UID: \"d45824eb-fa61-4a06-8b9c-d0fc6a7adb69\") " pod="openshift-console/console-77b75945b8-kqms5" Oct 02 01:57:49 crc kubenswrapper[4885]: I1002 01:57:49.132391 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/d45824eb-fa61-4a06-8b9c-d0fc6a7adb69-oauth-serving-cert\") pod \"console-77b75945b8-kqms5\" (UID: \"d45824eb-fa61-4a06-8b9c-d0fc6a7adb69\") " pod="openshift-console/console-77b75945b8-kqms5" Oct 02 01:57:49 crc kubenswrapper[4885]: I1002 01:57:49.132408 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/d45824eb-fa61-4a06-8b9c-d0fc6a7adb69-console-serving-cert\") pod \"console-77b75945b8-kqms5\" (UID: \"d45824eb-fa61-4a06-8b9c-d0fc6a7adb69\") " pod="openshift-console/console-77b75945b8-kqms5" Oct 02 01:57:49 crc kubenswrapper[4885]: I1002 01:57:49.132441 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d45824eb-fa61-4a06-8b9c-d0fc6a7adb69-trusted-ca-bundle\") pod \"console-77b75945b8-kqms5\" (UID: \"d45824eb-fa61-4a06-8b9c-d0fc6a7adb69\") " pod="openshift-console/console-77b75945b8-kqms5" Oct 02 01:57:49 crc kubenswrapper[4885]: I1002 01:57:49.132459 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d45824eb-fa61-4a06-8b9c-d0fc6a7adb69-service-ca\") pod \"console-77b75945b8-kqms5\" (UID: \"d45824eb-fa61-4a06-8b9c-d0fc6a7adb69\") " pod="openshift-console/console-77b75945b8-kqms5" Oct 02 01:57:49 crc kubenswrapper[4885]: I1002 01:57:49.132486 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/d45824eb-fa61-4a06-8b9c-d0fc6a7adb69-console-config\") pod \"console-77b75945b8-kqms5\" (UID: \"d45824eb-fa61-4a06-8b9c-d0fc6a7adb69\") " pod="openshift-console/console-77b75945b8-kqms5" Oct 02 01:57:49 crc kubenswrapper[4885]: I1002 01:57:49.132504 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/d45824eb-fa61-4a06-8b9c-d0fc6a7adb69-console-oauth-config\") pod \"console-77b75945b8-kqms5\" (UID: \"d45824eb-fa61-4a06-8b9c-d0fc6a7adb69\") " pod="openshift-console/console-77b75945b8-kqms5" Oct 02 01:57:49 crc kubenswrapper[4885]: I1002 01:57:49.133825 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/d45824eb-fa61-4a06-8b9c-d0fc6a7adb69-console-config\") pod \"console-77b75945b8-kqms5\" (UID: \"d45824eb-fa61-4a06-8b9c-d0fc6a7adb69\") " pod="openshift-console/console-77b75945b8-kqms5" Oct 02 01:57:49 crc kubenswrapper[4885]: I1002 01:57:49.134073 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d45824eb-fa61-4a06-8b9c-d0fc6a7adb69-service-ca\") pod \"console-77b75945b8-kqms5\" (UID: \"d45824eb-fa61-4a06-8b9c-d0fc6a7adb69\") " pod="openshift-console/console-77b75945b8-kqms5" Oct 02 01:57:49 crc kubenswrapper[4885]: I1002 01:57:49.134823 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/d45824eb-fa61-4a06-8b9c-d0fc6a7adb69-oauth-serving-cert\") pod \"console-77b75945b8-kqms5\" (UID: \"d45824eb-fa61-4a06-8b9c-d0fc6a7adb69\") " pod="openshift-console/console-77b75945b8-kqms5" Oct 02 01:57:49 crc kubenswrapper[4885]: I1002 01:57:49.135944 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d45824eb-fa61-4a06-8b9c-d0fc6a7adb69-trusted-ca-bundle\") pod \"console-77b75945b8-kqms5\" (UID: \"d45824eb-fa61-4a06-8b9c-d0fc6a7adb69\") " pod="openshift-console/console-77b75945b8-kqms5" Oct 02 01:57:49 crc kubenswrapper[4885]: I1002 01:57:49.137415 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/d45824eb-fa61-4a06-8b9c-d0fc6a7adb69-console-oauth-config\") pod \"console-77b75945b8-kqms5\" (UID: \"d45824eb-fa61-4a06-8b9c-d0fc6a7adb69\") " pod="openshift-console/console-77b75945b8-kqms5" Oct 02 01:57:49 crc kubenswrapper[4885]: I1002 01:57:49.137692 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/d45824eb-fa61-4a06-8b9c-d0fc6a7adb69-console-serving-cert\") pod \"console-77b75945b8-kqms5\" (UID: \"d45824eb-fa61-4a06-8b9c-d0fc6a7adb69\") " pod="openshift-console/console-77b75945b8-kqms5" Oct 02 01:57:49 crc kubenswrapper[4885]: I1002 01:57:49.149140 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vprdw\" (UniqueName: \"kubernetes.io/projected/d45824eb-fa61-4a06-8b9c-d0fc6a7adb69-kube-api-access-vprdw\") pod \"console-77b75945b8-kqms5\" (UID: \"d45824eb-fa61-4a06-8b9c-d0fc6a7adb69\") " pod="openshift-console/console-77b75945b8-kqms5" Oct 02 01:57:49 crc kubenswrapper[4885]: I1002 01:57:49.186626 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-sr7mv"] Oct 02 01:57:49 crc kubenswrapper[4885]: I1002 01:57:49.235310 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/f75137b4-4551-43cd-b20c-6044d13d27a1-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-jk45f\" (UID: \"f75137b4-4551-43cd-b20c-6044d13d27a1\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-jk45f" Oct 02 01:57:49 crc kubenswrapper[4885]: I1002 01:57:49.238956 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/f75137b4-4551-43cd-b20c-6044d13d27a1-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-jk45f\" (UID: \"f75137b4-4551-43cd-b20c-6044d13d27a1\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-jk45f" Oct 02 01:57:49 crc kubenswrapper[4885]: I1002 01:57:49.304091 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-jk45f" Oct 02 01:57:49 crc kubenswrapper[4885]: I1002 01:57:49.336114 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-77b75945b8-kqms5" Oct 02 01:57:49 crc kubenswrapper[4885]: I1002 01:57:49.488337 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-dvlk4"] Oct 02 01:57:49 crc kubenswrapper[4885]: W1002 01:57:49.496382 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc7f8ff02_de9d_456d_ba5e_a794f568c5cf.slice/crio-4e74be0b8a64cd346b7b69596d431029a7d0aacf300bb73bb28f31f5afb19362 WatchSource:0}: Error finding container 4e74be0b8a64cd346b7b69596d431029a7d0aacf300bb73bb28f31f5afb19362: Status 404 returned error can't find the container with id 4e74be0b8a64cd346b7b69596d431029a7d0aacf300bb73bb28f31f5afb19362 Oct 02 01:57:49 crc kubenswrapper[4885]: I1002 01:57:49.525554 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-jk45f"] Oct 02 01:57:49 crc kubenswrapper[4885]: W1002 01:57:49.529450 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf75137b4_4551_43cd_b20c_6044d13d27a1.slice/crio-6ce438fddc0bc3f39eeacf966d111d34550d1ce94d876afb9f13d5747c6a9824 WatchSource:0}: Error finding container 6ce438fddc0bc3f39eeacf966d111d34550d1ce94d876afb9f13d5747c6a9824: Status 404 returned error can't find the container with id 6ce438fddc0bc3f39eeacf966d111d34550d1ce94d876afb9f13d5747c6a9824 Oct 02 01:57:49 crc kubenswrapper[4885]: I1002 01:57:49.565657 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-77b75945b8-kqms5"] Oct 02 01:57:49 crc kubenswrapper[4885]: W1002 01:57:49.571899 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd45824eb_fa61_4a06_8b9c_d0fc6a7adb69.slice/crio-d05b16b26edb6b80e130c9ebc575e31492e5d97fea1e868434c96d616656ce6e WatchSource:0}: Error finding container d05b16b26edb6b80e130c9ebc575e31492e5d97fea1e868434c96d616656ce6e: Status 404 returned error can't find the container with id d05b16b26edb6b80e130c9ebc575e31492e5d97fea1e868434c96d616656ce6e Oct 02 01:57:49 crc kubenswrapper[4885]: I1002 01:57:49.678724 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-sr7mv" event={"ID":"c2a4785b-6378-4f85-ae1d-46e27903af36","Type":"ContainerStarted","Data":"2a0a52f871cf99feb64e524a859f8fd8bd1531474d7bf5001d52ebf8057914f3"} Oct 02 01:57:49 crc kubenswrapper[4885]: I1002 01:57:49.679741 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-77b75945b8-kqms5" event={"ID":"d45824eb-fa61-4a06-8b9c-d0fc6a7adb69","Type":"ContainerStarted","Data":"d05b16b26edb6b80e130c9ebc575e31492e5d97fea1e868434c96d616656ce6e"} Oct 02 01:57:49 crc kubenswrapper[4885]: I1002 01:57:49.680537 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-jk45f" event={"ID":"f75137b4-4551-43cd-b20c-6044d13d27a1","Type":"ContainerStarted","Data":"6ce438fddc0bc3f39eeacf966d111d34550d1ce94d876afb9f13d5747c6a9824"} Oct 02 01:57:49 crc kubenswrapper[4885]: I1002 01:57:49.681920 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-4xzpg" event={"ID":"4376783f-66f6-41a4-a42e-2666f1cc9978","Type":"ContainerStarted","Data":"28d49f99d7f0226f6ffed78113d832ffb115255954bbcdf1b7a770875c2bb6d0"} Oct 02 01:57:49 crc kubenswrapper[4885]: I1002 01:57:49.683804 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-dvlk4" event={"ID":"c7f8ff02-de9d-456d-ba5e-a794f568c5cf","Type":"ContainerStarted","Data":"4e74be0b8a64cd346b7b69596d431029a7d0aacf300bb73bb28f31f5afb19362"} Oct 02 01:57:50 crc kubenswrapper[4885]: I1002 01:57:50.696331 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-77b75945b8-kqms5" event={"ID":"d45824eb-fa61-4a06-8b9c-d0fc6a7adb69","Type":"ContainerStarted","Data":"ad7e493ce823a4673396b3af348f0fa921837a0675623f0b0813f979fa5888ff"} Oct 02 01:57:50 crc kubenswrapper[4885]: I1002 01:57:50.730172 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-77b75945b8-kqms5" podStartSLOduration=2.730147188 podStartE2EDuration="2.730147188s" podCreationTimestamp="2025-10-02 01:57:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:57:50.727498057 +0000 UTC m=+659.539245496" watchObservedRunningTime="2025-10-02 01:57:50.730147188 +0000 UTC m=+659.541894617" Oct 02 01:57:55 crc kubenswrapper[4885]: I1002 01:57:55.728414 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-sr7mv" event={"ID":"c2a4785b-6378-4f85-ae1d-46e27903af36","Type":"ContainerStarted","Data":"5308d9072312e09099369470c793057113a6ab7e07b02ecdfd09ba549b14c2a1"} Oct 02 01:57:55 crc kubenswrapper[4885]: I1002 01:57:55.730045 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-jk45f" event={"ID":"f75137b4-4551-43cd-b20c-6044d13d27a1","Type":"ContainerStarted","Data":"41b7dfac871f6fe485973f9fde74e01afa5a59d498a1e2ef7bc515c6f9166797"} Oct 02 01:57:55 crc kubenswrapper[4885]: I1002 01:57:55.730242 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-jk45f" Oct 02 01:57:55 crc kubenswrapper[4885]: I1002 01:57:55.731538 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-dvlk4" event={"ID":"c7f8ff02-de9d-456d-ba5e-a794f568c5cf","Type":"ContainerStarted","Data":"199a7e479418d8d58fe6311a2a2f04f258cf64c86ff1a561be3991ae9ea5c29c"} Oct 02 01:57:55 crc kubenswrapper[4885]: I1002 01:57:55.732958 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-4xzpg" event={"ID":"4376783f-66f6-41a4-a42e-2666f1cc9978","Type":"ContainerStarted","Data":"2633b2e6bda18f21fba7afb7c80c2f21dd94c11b55569ffde33fdec449437635"} Oct 02 01:57:55 crc kubenswrapper[4885]: I1002 01:57:55.733102 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-4xzpg" Oct 02 01:57:55 crc kubenswrapper[4885]: I1002 01:57:55.750756 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-jk45f" podStartSLOduration=2.170085397 podStartE2EDuration="7.750740202s" podCreationTimestamp="2025-10-02 01:57:48 +0000 UTC" firstStartedPulling="2025-10-02 01:57:49.531408987 +0000 UTC m=+658.343156386" lastFinishedPulling="2025-10-02 01:57:55.112063762 +0000 UTC m=+663.923811191" observedRunningTime="2025-10-02 01:57:55.749202404 +0000 UTC m=+664.560949813" watchObservedRunningTime="2025-10-02 01:57:55.750740202 +0000 UTC m=+664.562487621" Oct 02 01:57:55 crc kubenswrapper[4885]: I1002 01:57:55.773992 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-4xzpg" podStartSLOduration=1.585331074 podStartE2EDuration="7.773967876s" podCreationTimestamp="2025-10-02 01:57:48 +0000 UTC" firstStartedPulling="2025-10-02 01:57:49.052937072 +0000 UTC m=+657.864684471" lastFinishedPulling="2025-10-02 01:57:55.241573824 +0000 UTC m=+664.053321273" observedRunningTime="2025-10-02 01:57:55.771920314 +0000 UTC m=+664.583667723" watchObservedRunningTime="2025-10-02 01:57:55.773967876 +0000 UTC m=+664.585715315" Oct 02 01:57:55 crc kubenswrapper[4885]: I1002 01:57:55.790451 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-dvlk4" podStartSLOduration=2.103081601 podStartE2EDuration="7.790430786s" podCreationTimestamp="2025-10-02 01:57:48 +0000 UTC" firstStartedPulling="2025-10-02 01:57:49.49925009 +0000 UTC m=+658.310997489" lastFinishedPulling="2025-10-02 01:57:55.186599235 +0000 UTC m=+663.998346674" observedRunningTime="2025-10-02 01:57:55.789700494 +0000 UTC m=+664.601447903" watchObservedRunningTime="2025-10-02 01:57:55.790430786 +0000 UTC m=+664.602178195" Oct 02 01:57:59 crc kubenswrapper[4885]: I1002 01:57:59.336985 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-77b75945b8-kqms5" Oct 02 01:57:59 crc kubenswrapper[4885]: I1002 01:57:59.337712 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-77b75945b8-kqms5" Oct 02 01:57:59 crc kubenswrapper[4885]: I1002 01:57:59.345135 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-77b75945b8-kqms5" Oct 02 01:57:59 crc kubenswrapper[4885]: I1002 01:57:59.787634 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-77b75945b8-kqms5" Oct 02 01:57:59 crc kubenswrapper[4885]: I1002 01:57:59.882452 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-kckm8"] Oct 02 01:58:01 crc kubenswrapper[4885]: I1002 01:58:01.797753 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-sr7mv" event={"ID":"c2a4785b-6378-4f85-ae1d-46e27903af36","Type":"ContainerStarted","Data":"50f460d9e17c0078d7b0a90e6d21796e47c62b37afdda748042febd0336d9c6d"} Oct 02 01:58:01 crc kubenswrapper[4885]: I1002 01:58:01.833332 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-sr7mv" podStartSLOduration=2.345593683 podStartE2EDuration="13.833301153s" podCreationTimestamp="2025-10-02 01:57:48 +0000 UTC" firstStartedPulling="2025-10-02 01:57:49.196119588 +0000 UTC m=+658.007866987" lastFinishedPulling="2025-10-02 01:58:00.683827028 +0000 UTC m=+669.495574457" observedRunningTime="2025-10-02 01:58:01.826163007 +0000 UTC m=+670.637910456" watchObservedRunningTime="2025-10-02 01:58:01.833301153 +0000 UTC m=+670.645048592" Oct 02 01:58:04 crc kubenswrapper[4885]: I1002 01:58:04.067288 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-4xzpg" Oct 02 01:58:09 crc kubenswrapper[4885]: I1002 01:58:09.313690 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-jk45f" Oct 02 01:58:24 crc kubenswrapper[4885]: I1002 01:58:24.929203 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-kckm8" podUID="a86af5ef-01e3-47e4-b324-208537e5ec71" containerName="console" containerID="cri-o://d39fd1cf35e39882b5385bdcc10cce5b20cc3842cc876b09766cfecd0dde162f" gracePeriod=15 Oct 02 01:58:25 crc kubenswrapper[4885]: I1002 01:58:25.351641 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-kckm8_a86af5ef-01e3-47e4-b324-208537e5ec71/console/0.log" Oct 02 01:58:25 crc kubenswrapper[4885]: I1002 01:58:25.351889 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-kckm8" Oct 02 01:58:25 crc kubenswrapper[4885]: I1002 01:58:25.502933 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/a86af5ef-01e3-47e4-b324-208537e5ec71-console-config\") pod \"a86af5ef-01e3-47e4-b324-208537e5ec71\" (UID: \"a86af5ef-01e3-47e4-b324-208537e5ec71\") " Oct 02 01:58:25 crc kubenswrapper[4885]: I1002 01:58:25.503012 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/a86af5ef-01e3-47e4-b324-208537e5ec71-console-oauth-config\") pod \"a86af5ef-01e3-47e4-b324-208537e5ec71\" (UID: \"a86af5ef-01e3-47e4-b324-208537e5ec71\") " Oct 02 01:58:25 crc kubenswrapper[4885]: I1002 01:58:25.503041 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/a86af5ef-01e3-47e4-b324-208537e5ec71-console-serving-cert\") pod \"a86af5ef-01e3-47e4-b324-208537e5ec71\" (UID: \"a86af5ef-01e3-47e4-b324-208537e5ec71\") " Oct 02 01:58:25 crc kubenswrapper[4885]: I1002 01:58:25.503079 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a86af5ef-01e3-47e4-b324-208537e5ec71-service-ca\") pod \"a86af5ef-01e3-47e4-b324-208537e5ec71\" (UID: \"a86af5ef-01e3-47e4-b324-208537e5ec71\") " Oct 02 01:58:25 crc kubenswrapper[4885]: I1002 01:58:25.503103 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a86af5ef-01e3-47e4-b324-208537e5ec71-trusted-ca-bundle\") pod \"a86af5ef-01e3-47e4-b324-208537e5ec71\" (UID: \"a86af5ef-01e3-47e4-b324-208537e5ec71\") " Oct 02 01:58:25 crc kubenswrapper[4885]: I1002 01:58:25.503136 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qbmxg\" (UniqueName: \"kubernetes.io/projected/a86af5ef-01e3-47e4-b324-208537e5ec71-kube-api-access-qbmxg\") pod \"a86af5ef-01e3-47e4-b324-208537e5ec71\" (UID: \"a86af5ef-01e3-47e4-b324-208537e5ec71\") " Oct 02 01:58:25 crc kubenswrapper[4885]: I1002 01:58:25.503152 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/a86af5ef-01e3-47e4-b324-208537e5ec71-oauth-serving-cert\") pod \"a86af5ef-01e3-47e4-b324-208537e5ec71\" (UID: \"a86af5ef-01e3-47e4-b324-208537e5ec71\") " Oct 02 01:58:25 crc kubenswrapper[4885]: I1002 01:58:25.504106 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a86af5ef-01e3-47e4-b324-208537e5ec71-service-ca" (OuterVolumeSpecName: "service-ca") pod "a86af5ef-01e3-47e4-b324-208537e5ec71" (UID: "a86af5ef-01e3-47e4-b324-208537e5ec71"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:58:25 crc kubenswrapper[4885]: I1002 01:58:25.504124 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a86af5ef-01e3-47e4-b324-208537e5ec71-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "a86af5ef-01e3-47e4-b324-208537e5ec71" (UID: "a86af5ef-01e3-47e4-b324-208537e5ec71"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:58:25 crc kubenswrapper[4885]: I1002 01:58:25.504244 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a86af5ef-01e3-47e4-b324-208537e5ec71-console-config" (OuterVolumeSpecName: "console-config") pod "a86af5ef-01e3-47e4-b324-208537e5ec71" (UID: "a86af5ef-01e3-47e4-b324-208537e5ec71"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:58:25 crc kubenswrapper[4885]: I1002 01:58:25.504340 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a86af5ef-01e3-47e4-b324-208537e5ec71-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "a86af5ef-01e3-47e4-b324-208537e5ec71" (UID: "a86af5ef-01e3-47e4-b324-208537e5ec71"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:58:25 crc kubenswrapper[4885]: I1002 01:58:25.511040 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a86af5ef-01e3-47e4-b324-208537e5ec71-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "a86af5ef-01e3-47e4-b324-208537e5ec71" (UID: "a86af5ef-01e3-47e4-b324-208537e5ec71"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:58:25 crc kubenswrapper[4885]: I1002 01:58:25.512197 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a86af5ef-01e3-47e4-b324-208537e5ec71-kube-api-access-qbmxg" (OuterVolumeSpecName: "kube-api-access-qbmxg") pod "a86af5ef-01e3-47e4-b324-208537e5ec71" (UID: "a86af5ef-01e3-47e4-b324-208537e5ec71"). InnerVolumeSpecName "kube-api-access-qbmxg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:58:25 crc kubenswrapper[4885]: I1002 01:58:25.512580 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a86af5ef-01e3-47e4-b324-208537e5ec71-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "a86af5ef-01e3-47e4-b324-208537e5ec71" (UID: "a86af5ef-01e3-47e4-b324-208537e5ec71"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:58:25 crc kubenswrapper[4885]: I1002 01:58:25.607517 4885 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a86af5ef-01e3-47e4-b324-208537e5ec71-service-ca\") on node \"crc\" DevicePath \"\"" Oct 02 01:58:25 crc kubenswrapper[4885]: I1002 01:58:25.607552 4885 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a86af5ef-01e3-47e4-b324-208537e5ec71-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 01:58:25 crc kubenswrapper[4885]: I1002 01:58:25.607567 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qbmxg\" (UniqueName: \"kubernetes.io/projected/a86af5ef-01e3-47e4-b324-208537e5ec71-kube-api-access-qbmxg\") on node \"crc\" DevicePath \"\"" Oct 02 01:58:25 crc kubenswrapper[4885]: I1002 01:58:25.607579 4885 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/a86af5ef-01e3-47e4-b324-208537e5ec71-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 01:58:25 crc kubenswrapper[4885]: I1002 01:58:25.607591 4885 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/a86af5ef-01e3-47e4-b324-208537e5ec71-console-config\") on node \"crc\" DevicePath \"\"" Oct 02 01:58:25 crc kubenswrapper[4885]: I1002 01:58:25.607603 4885 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/a86af5ef-01e3-47e4-b324-208537e5ec71-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 02 01:58:25 crc kubenswrapper[4885]: I1002 01:58:25.607615 4885 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/a86af5ef-01e3-47e4-b324-208537e5ec71-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 01:58:25 crc kubenswrapper[4885]: I1002 01:58:25.624089 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2ms2kw"] Oct 02 01:58:25 crc kubenswrapper[4885]: E1002 01:58:25.624383 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a86af5ef-01e3-47e4-b324-208537e5ec71" containerName="console" Oct 02 01:58:25 crc kubenswrapper[4885]: I1002 01:58:25.624397 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="a86af5ef-01e3-47e4-b324-208537e5ec71" containerName="console" Oct 02 01:58:25 crc kubenswrapper[4885]: I1002 01:58:25.624538 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="a86af5ef-01e3-47e4-b324-208537e5ec71" containerName="console" Oct 02 01:58:25 crc kubenswrapper[4885]: I1002 01:58:25.635529 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2ms2kw" Oct 02 01:58:25 crc kubenswrapper[4885]: I1002 01:58:25.639132 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 02 01:58:25 crc kubenswrapper[4885]: I1002 01:58:25.640825 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2ms2kw"] Oct 02 01:58:25 crc kubenswrapper[4885]: I1002 01:58:25.709170 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/65b0be27-bd89-4ddd-a578-7977b676f674-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2ms2kw\" (UID: \"65b0be27-bd89-4ddd-a578-7977b676f674\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2ms2kw" Oct 02 01:58:25 crc kubenswrapper[4885]: I1002 01:58:25.709412 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2f6kl\" (UniqueName: \"kubernetes.io/projected/65b0be27-bd89-4ddd-a578-7977b676f674-kube-api-access-2f6kl\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2ms2kw\" (UID: \"65b0be27-bd89-4ddd-a578-7977b676f674\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2ms2kw" Oct 02 01:58:25 crc kubenswrapper[4885]: I1002 01:58:25.709500 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/65b0be27-bd89-4ddd-a578-7977b676f674-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2ms2kw\" (UID: \"65b0be27-bd89-4ddd-a578-7977b676f674\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2ms2kw" Oct 02 01:58:25 crc kubenswrapper[4885]: I1002 01:58:25.810508 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/65b0be27-bd89-4ddd-a578-7977b676f674-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2ms2kw\" (UID: \"65b0be27-bd89-4ddd-a578-7977b676f674\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2ms2kw" Oct 02 01:58:25 crc kubenswrapper[4885]: I1002 01:58:25.810650 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2f6kl\" (UniqueName: \"kubernetes.io/projected/65b0be27-bd89-4ddd-a578-7977b676f674-kube-api-access-2f6kl\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2ms2kw\" (UID: \"65b0be27-bd89-4ddd-a578-7977b676f674\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2ms2kw" Oct 02 01:58:25 crc kubenswrapper[4885]: I1002 01:58:25.810711 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/65b0be27-bd89-4ddd-a578-7977b676f674-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2ms2kw\" (UID: \"65b0be27-bd89-4ddd-a578-7977b676f674\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2ms2kw" Oct 02 01:58:25 crc kubenswrapper[4885]: I1002 01:58:25.811484 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/65b0be27-bd89-4ddd-a578-7977b676f674-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2ms2kw\" (UID: \"65b0be27-bd89-4ddd-a578-7977b676f674\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2ms2kw" Oct 02 01:58:25 crc kubenswrapper[4885]: I1002 01:58:25.811565 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/65b0be27-bd89-4ddd-a578-7977b676f674-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2ms2kw\" (UID: \"65b0be27-bd89-4ddd-a578-7977b676f674\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2ms2kw" Oct 02 01:58:25 crc kubenswrapper[4885]: I1002 01:58:25.828466 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2f6kl\" (UniqueName: \"kubernetes.io/projected/65b0be27-bd89-4ddd-a578-7977b676f674-kube-api-access-2f6kl\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2ms2kw\" (UID: \"65b0be27-bd89-4ddd-a578-7977b676f674\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2ms2kw" Oct 02 01:58:25 crc kubenswrapper[4885]: I1002 01:58:25.965454 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2ms2kw" Oct 02 01:58:25 crc kubenswrapper[4885]: I1002 01:58:25.983375 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-kckm8_a86af5ef-01e3-47e4-b324-208537e5ec71/console/0.log" Oct 02 01:58:25 crc kubenswrapper[4885]: I1002 01:58:25.983450 4885 generic.go:334] "Generic (PLEG): container finished" podID="a86af5ef-01e3-47e4-b324-208537e5ec71" containerID="d39fd1cf35e39882b5385bdcc10cce5b20cc3842cc876b09766cfecd0dde162f" exitCode=2 Oct 02 01:58:25 crc kubenswrapper[4885]: I1002 01:58:25.983490 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-kckm8" event={"ID":"a86af5ef-01e3-47e4-b324-208537e5ec71","Type":"ContainerDied","Data":"d39fd1cf35e39882b5385bdcc10cce5b20cc3842cc876b09766cfecd0dde162f"} Oct 02 01:58:25 crc kubenswrapper[4885]: I1002 01:58:25.983538 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-kckm8" event={"ID":"a86af5ef-01e3-47e4-b324-208537e5ec71","Type":"ContainerDied","Data":"56f1c5035b5dae4325cc2f40ec856cf63cd4d17ebf0505ccdb38b814b3f7224b"} Oct 02 01:58:25 crc kubenswrapper[4885]: I1002 01:58:25.983562 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-kckm8" Oct 02 01:58:25 crc kubenswrapper[4885]: I1002 01:58:25.983567 4885 scope.go:117] "RemoveContainer" containerID="d39fd1cf35e39882b5385bdcc10cce5b20cc3842cc876b09766cfecd0dde162f" Oct 02 01:58:26 crc kubenswrapper[4885]: I1002 01:58:26.010589 4885 scope.go:117] "RemoveContainer" containerID="d39fd1cf35e39882b5385bdcc10cce5b20cc3842cc876b09766cfecd0dde162f" Oct 02 01:58:26 crc kubenswrapper[4885]: E1002 01:58:26.011477 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d39fd1cf35e39882b5385bdcc10cce5b20cc3842cc876b09766cfecd0dde162f\": container with ID starting with d39fd1cf35e39882b5385bdcc10cce5b20cc3842cc876b09766cfecd0dde162f not found: ID does not exist" containerID="d39fd1cf35e39882b5385bdcc10cce5b20cc3842cc876b09766cfecd0dde162f" Oct 02 01:58:26 crc kubenswrapper[4885]: I1002 01:58:26.011520 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d39fd1cf35e39882b5385bdcc10cce5b20cc3842cc876b09766cfecd0dde162f"} err="failed to get container status \"d39fd1cf35e39882b5385bdcc10cce5b20cc3842cc876b09766cfecd0dde162f\": rpc error: code = NotFound desc = could not find container \"d39fd1cf35e39882b5385bdcc10cce5b20cc3842cc876b09766cfecd0dde162f\": container with ID starting with d39fd1cf35e39882b5385bdcc10cce5b20cc3842cc876b09766cfecd0dde162f not found: ID does not exist" Oct 02 01:58:26 crc kubenswrapper[4885]: I1002 01:58:26.030466 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-kckm8"] Oct 02 01:58:26 crc kubenswrapper[4885]: I1002 01:58:26.037592 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-kckm8"] Oct 02 01:58:26 crc kubenswrapper[4885]: I1002 01:58:26.059541 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a86af5ef-01e3-47e4-b324-208537e5ec71" path="/var/lib/kubelet/pods/a86af5ef-01e3-47e4-b324-208537e5ec71/volumes" Oct 02 01:58:26 crc kubenswrapper[4885]: I1002 01:58:26.448035 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2ms2kw"] Oct 02 01:58:26 crc kubenswrapper[4885]: I1002 01:58:26.994868 4885 generic.go:334] "Generic (PLEG): container finished" podID="65b0be27-bd89-4ddd-a578-7977b676f674" containerID="a3f5a3d9edd17f61b8f7b0b273c3600e4fdd3f9e2c95de92a4fd44379baaf342" exitCode=0 Oct 02 01:58:26 crc kubenswrapper[4885]: I1002 01:58:26.994949 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2ms2kw" event={"ID":"65b0be27-bd89-4ddd-a578-7977b676f674","Type":"ContainerDied","Data":"a3f5a3d9edd17f61b8f7b0b273c3600e4fdd3f9e2c95de92a4fd44379baaf342"} Oct 02 01:58:26 crc kubenswrapper[4885]: I1002 01:58:26.995175 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2ms2kw" event={"ID":"65b0be27-bd89-4ddd-a578-7977b676f674","Type":"ContainerStarted","Data":"e79606cd079d7ebef71331460277cbe908bcdf9b7c132d4c9fd3cf250e7744e3"} Oct 02 01:58:29 crc kubenswrapper[4885]: I1002 01:58:29.017580 4885 generic.go:334] "Generic (PLEG): container finished" podID="65b0be27-bd89-4ddd-a578-7977b676f674" containerID="be68c82a3873e490166f9f0bef59f46c30c386cef2b4bd65dee296b27c3a1b83" exitCode=0 Oct 02 01:58:29 crc kubenswrapper[4885]: I1002 01:58:29.017648 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2ms2kw" event={"ID":"65b0be27-bd89-4ddd-a578-7977b676f674","Type":"ContainerDied","Data":"be68c82a3873e490166f9f0bef59f46c30c386cef2b4bd65dee296b27c3a1b83"} Oct 02 01:58:30 crc kubenswrapper[4885]: I1002 01:58:30.029693 4885 generic.go:334] "Generic (PLEG): container finished" podID="65b0be27-bd89-4ddd-a578-7977b676f674" containerID="6bf23d94b7fe6841a2ff890707507552e5e5cae77e51a31dd04dd5e5962d8bf4" exitCode=0 Oct 02 01:58:30 crc kubenswrapper[4885]: I1002 01:58:30.029754 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2ms2kw" event={"ID":"65b0be27-bd89-4ddd-a578-7977b676f674","Type":"ContainerDied","Data":"6bf23d94b7fe6841a2ff890707507552e5e5cae77e51a31dd04dd5e5962d8bf4"} Oct 02 01:58:31 crc kubenswrapper[4885]: I1002 01:58:31.334430 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2ms2kw" Oct 02 01:58:31 crc kubenswrapper[4885]: I1002 01:58:31.419166 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/65b0be27-bd89-4ddd-a578-7977b676f674-bundle\") pod \"65b0be27-bd89-4ddd-a578-7977b676f674\" (UID: \"65b0be27-bd89-4ddd-a578-7977b676f674\") " Oct 02 01:58:31 crc kubenswrapper[4885]: I1002 01:58:31.419342 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2f6kl\" (UniqueName: \"kubernetes.io/projected/65b0be27-bd89-4ddd-a578-7977b676f674-kube-api-access-2f6kl\") pod \"65b0be27-bd89-4ddd-a578-7977b676f674\" (UID: \"65b0be27-bd89-4ddd-a578-7977b676f674\") " Oct 02 01:58:31 crc kubenswrapper[4885]: I1002 01:58:31.419434 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/65b0be27-bd89-4ddd-a578-7977b676f674-util\") pod \"65b0be27-bd89-4ddd-a578-7977b676f674\" (UID: \"65b0be27-bd89-4ddd-a578-7977b676f674\") " Oct 02 01:58:31 crc kubenswrapper[4885]: I1002 01:58:31.420398 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/65b0be27-bd89-4ddd-a578-7977b676f674-bundle" (OuterVolumeSpecName: "bundle") pod "65b0be27-bd89-4ddd-a578-7977b676f674" (UID: "65b0be27-bd89-4ddd-a578-7977b676f674"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 01:58:31 crc kubenswrapper[4885]: I1002 01:58:31.428646 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/65b0be27-bd89-4ddd-a578-7977b676f674-kube-api-access-2f6kl" (OuterVolumeSpecName: "kube-api-access-2f6kl") pod "65b0be27-bd89-4ddd-a578-7977b676f674" (UID: "65b0be27-bd89-4ddd-a578-7977b676f674"). InnerVolumeSpecName "kube-api-access-2f6kl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:58:31 crc kubenswrapper[4885]: I1002 01:58:31.446818 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/65b0be27-bd89-4ddd-a578-7977b676f674-util" (OuterVolumeSpecName: "util") pod "65b0be27-bd89-4ddd-a578-7977b676f674" (UID: "65b0be27-bd89-4ddd-a578-7977b676f674"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 01:58:31 crc kubenswrapper[4885]: I1002 01:58:31.522044 4885 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/65b0be27-bd89-4ddd-a578-7977b676f674-util\") on node \"crc\" DevicePath \"\"" Oct 02 01:58:31 crc kubenswrapper[4885]: I1002 01:58:31.522108 4885 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/65b0be27-bd89-4ddd-a578-7977b676f674-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 01:58:31 crc kubenswrapper[4885]: I1002 01:58:31.522138 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2f6kl\" (UniqueName: \"kubernetes.io/projected/65b0be27-bd89-4ddd-a578-7977b676f674-kube-api-access-2f6kl\") on node \"crc\" DevicePath \"\"" Oct 02 01:58:32 crc kubenswrapper[4885]: I1002 01:58:32.051291 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2ms2kw" Oct 02 01:58:32 crc kubenswrapper[4885]: I1002 01:58:32.057678 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2ms2kw" event={"ID":"65b0be27-bd89-4ddd-a578-7977b676f674","Type":"ContainerDied","Data":"e79606cd079d7ebef71331460277cbe908bcdf9b7c132d4c9fd3cf250e7744e3"} Oct 02 01:58:32 crc kubenswrapper[4885]: I1002 01:58:32.057728 4885 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e79606cd079d7ebef71331460277cbe908bcdf9b7c132d4c9fd3cf250e7744e3" Oct 02 01:58:40 crc kubenswrapper[4885]: I1002 01:58:40.618124 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-6c54f448c5-tknpj"] Oct 02 01:58:40 crc kubenswrapper[4885]: E1002 01:58:40.618857 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65b0be27-bd89-4ddd-a578-7977b676f674" containerName="pull" Oct 02 01:58:40 crc kubenswrapper[4885]: I1002 01:58:40.618871 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="65b0be27-bd89-4ddd-a578-7977b676f674" containerName="pull" Oct 02 01:58:40 crc kubenswrapper[4885]: E1002 01:58:40.618893 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65b0be27-bd89-4ddd-a578-7977b676f674" containerName="util" Oct 02 01:58:40 crc kubenswrapper[4885]: I1002 01:58:40.618902 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="65b0be27-bd89-4ddd-a578-7977b676f674" containerName="util" Oct 02 01:58:40 crc kubenswrapper[4885]: E1002 01:58:40.618925 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65b0be27-bd89-4ddd-a578-7977b676f674" containerName="extract" Oct 02 01:58:40 crc kubenswrapper[4885]: I1002 01:58:40.618935 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="65b0be27-bd89-4ddd-a578-7977b676f674" containerName="extract" Oct 02 01:58:40 crc kubenswrapper[4885]: I1002 01:58:40.619050 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="65b0be27-bd89-4ddd-a578-7977b676f674" containerName="extract" Oct 02 01:58:40 crc kubenswrapper[4885]: I1002 01:58:40.619616 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-6c54f448c5-tknpj" Oct 02 01:58:40 crc kubenswrapper[4885]: I1002 01:58:40.623309 4885 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Oct 02 01:58:40 crc kubenswrapper[4885]: I1002 01:58:40.623558 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Oct 02 01:58:40 crc kubenswrapper[4885]: I1002 01:58:40.623677 4885 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Oct 02 01:58:40 crc kubenswrapper[4885]: I1002 01:58:40.624051 4885 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-crj4q" Oct 02 01:58:40 crc kubenswrapper[4885]: I1002 01:58:40.625073 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Oct 02 01:58:40 crc kubenswrapper[4885]: I1002 01:58:40.667873 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/7dc51fda-276d-4739-975d-e88c9fcec5d3-apiservice-cert\") pod \"metallb-operator-controller-manager-6c54f448c5-tknpj\" (UID: \"7dc51fda-276d-4739-975d-e88c9fcec5d3\") " pod="metallb-system/metallb-operator-controller-manager-6c54f448c5-tknpj" Oct 02 01:58:40 crc kubenswrapper[4885]: I1002 01:58:40.668236 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/7dc51fda-276d-4739-975d-e88c9fcec5d3-webhook-cert\") pod \"metallb-operator-controller-manager-6c54f448c5-tknpj\" (UID: \"7dc51fda-276d-4739-975d-e88c9fcec5d3\") " pod="metallb-system/metallb-operator-controller-manager-6c54f448c5-tknpj" Oct 02 01:58:40 crc kubenswrapper[4885]: I1002 01:58:40.668413 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bjx7c\" (UniqueName: \"kubernetes.io/projected/7dc51fda-276d-4739-975d-e88c9fcec5d3-kube-api-access-bjx7c\") pod \"metallb-operator-controller-manager-6c54f448c5-tknpj\" (UID: \"7dc51fda-276d-4739-975d-e88c9fcec5d3\") " pod="metallb-system/metallb-operator-controller-manager-6c54f448c5-tknpj" Oct 02 01:58:40 crc kubenswrapper[4885]: I1002 01:58:40.682389 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-6c54f448c5-tknpj"] Oct 02 01:58:40 crc kubenswrapper[4885]: I1002 01:58:40.751715 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-fd987dcb7-f2dx2"] Oct 02 01:58:40 crc kubenswrapper[4885]: I1002 01:58:40.753087 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-fd987dcb7-f2dx2" Oct 02 01:58:40 crc kubenswrapper[4885]: I1002 01:58:40.754978 4885 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-6hznj" Oct 02 01:58:40 crc kubenswrapper[4885]: I1002 01:58:40.755682 4885 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Oct 02 01:58:40 crc kubenswrapper[4885]: I1002 01:58:40.755906 4885 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Oct 02 01:58:40 crc kubenswrapper[4885]: I1002 01:58:40.768712 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-fd987dcb7-f2dx2"] Oct 02 01:58:40 crc kubenswrapper[4885]: I1002 01:58:40.769230 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/7dc51fda-276d-4739-975d-e88c9fcec5d3-apiservice-cert\") pod \"metallb-operator-controller-manager-6c54f448c5-tknpj\" (UID: \"7dc51fda-276d-4739-975d-e88c9fcec5d3\") " pod="metallb-system/metallb-operator-controller-manager-6c54f448c5-tknpj" Oct 02 01:58:40 crc kubenswrapper[4885]: I1002 01:58:40.769296 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/7dc51fda-276d-4739-975d-e88c9fcec5d3-webhook-cert\") pod \"metallb-operator-controller-manager-6c54f448c5-tknpj\" (UID: \"7dc51fda-276d-4739-975d-e88c9fcec5d3\") " pod="metallb-system/metallb-operator-controller-manager-6c54f448c5-tknpj" Oct 02 01:58:40 crc kubenswrapper[4885]: I1002 01:58:40.769326 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bjx7c\" (UniqueName: \"kubernetes.io/projected/7dc51fda-276d-4739-975d-e88c9fcec5d3-kube-api-access-bjx7c\") pod \"metallb-operator-controller-manager-6c54f448c5-tknpj\" (UID: \"7dc51fda-276d-4739-975d-e88c9fcec5d3\") " pod="metallb-system/metallb-operator-controller-manager-6c54f448c5-tknpj" Oct 02 01:58:40 crc kubenswrapper[4885]: I1002 01:58:40.780975 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/7dc51fda-276d-4739-975d-e88c9fcec5d3-webhook-cert\") pod \"metallb-operator-controller-manager-6c54f448c5-tknpj\" (UID: \"7dc51fda-276d-4739-975d-e88c9fcec5d3\") " pod="metallb-system/metallb-operator-controller-manager-6c54f448c5-tknpj" Oct 02 01:58:40 crc kubenswrapper[4885]: I1002 01:58:40.781462 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/7dc51fda-276d-4739-975d-e88c9fcec5d3-apiservice-cert\") pod \"metallb-operator-controller-manager-6c54f448c5-tknpj\" (UID: \"7dc51fda-276d-4739-975d-e88c9fcec5d3\") " pod="metallb-system/metallb-operator-controller-manager-6c54f448c5-tknpj" Oct 02 01:58:40 crc kubenswrapper[4885]: I1002 01:58:40.794095 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bjx7c\" (UniqueName: \"kubernetes.io/projected/7dc51fda-276d-4739-975d-e88c9fcec5d3-kube-api-access-bjx7c\") pod \"metallb-operator-controller-manager-6c54f448c5-tknpj\" (UID: \"7dc51fda-276d-4739-975d-e88c9fcec5d3\") " pod="metallb-system/metallb-operator-controller-manager-6c54f448c5-tknpj" Oct 02 01:58:40 crc kubenswrapper[4885]: I1002 01:58:40.870843 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pd658\" (UniqueName: \"kubernetes.io/projected/e0315b09-692d-4c5f-8861-0863f90338fa-kube-api-access-pd658\") pod \"metallb-operator-webhook-server-fd987dcb7-f2dx2\" (UID: \"e0315b09-692d-4c5f-8861-0863f90338fa\") " pod="metallb-system/metallb-operator-webhook-server-fd987dcb7-f2dx2" Oct 02 01:58:40 crc kubenswrapper[4885]: I1002 01:58:40.870905 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e0315b09-692d-4c5f-8861-0863f90338fa-webhook-cert\") pod \"metallb-operator-webhook-server-fd987dcb7-f2dx2\" (UID: \"e0315b09-692d-4c5f-8861-0863f90338fa\") " pod="metallb-system/metallb-operator-webhook-server-fd987dcb7-f2dx2" Oct 02 01:58:40 crc kubenswrapper[4885]: I1002 01:58:40.871043 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e0315b09-692d-4c5f-8861-0863f90338fa-apiservice-cert\") pod \"metallb-operator-webhook-server-fd987dcb7-f2dx2\" (UID: \"e0315b09-692d-4c5f-8861-0863f90338fa\") " pod="metallb-system/metallb-operator-webhook-server-fd987dcb7-f2dx2" Oct 02 01:58:40 crc kubenswrapper[4885]: I1002 01:58:40.935568 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-6c54f448c5-tknpj" Oct 02 01:58:40 crc kubenswrapper[4885]: I1002 01:58:40.972263 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e0315b09-692d-4c5f-8861-0863f90338fa-apiservice-cert\") pod \"metallb-operator-webhook-server-fd987dcb7-f2dx2\" (UID: \"e0315b09-692d-4c5f-8861-0863f90338fa\") " pod="metallb-system/metallb-operator-webhook-server-fd987dcb7-f2dx2" Oct 02 01:58:40 crc kubenswrapper[4885]: I1002 01:58:40.972354 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pd658\" (UniqueName: \"kubernetes.io/projected/e0315b09-692d-4c5f-8861-0863f90338fa-kube-api-access-pd658\") pod \"metallb-operator-webhook-server-fd987dcb7-f2dx2\" (UID: \"e0315b09-692d-4c5f-8861-0863f90338fa\") " pod="metallb-system/metallb-operator-webhook-server-fd987dcb7-f2dx2" Oct 02 01:58:40 crc kubenswrapper[4885]: I1002 01:58:40.972392 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e0315b09-692d-4c5f-8861-0863f90338fa-webhook-cert\") pod \"metallb-operator-webhook-server-fd987dcb7-f2dx2\" (UID: \"e0315b09-692d-4c5f-8861-0863f90338fa\") " pod="metallb-system/metallb-operator-webhook-server-fd987dcb7-f2dx2" Oct 02 01:58:40 crc kubenswrapper[4885]: I1002 01:58:40.975493 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e0315b09-692d-4c5f-8861-0863f90338fa-webhook-cert\") pod \"metallb-operator-webhook-server-fd987dcb7-f2dx2\" (UID: \"e0315b09-692d-4c5f-8861-0863f90338fa\") " pod="metallb-system/metallb-operator-webhook-server-fd987dcb7-f2dx2" Oct 02 01:58:40 crc kubenswrapper[4885]: I1002 01:58:40.975561 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e0315b09-692d-4c5f-8861-0863f90338fa-apiservice-cert\") pod \"metallb-operator-webhook-server-fd987dcb7-f2dx2\" (UID: \"e0315b09-692d-4c5f-8861-0863f90338fa\") " pod="metallb-system/metallb-operator-webhook-server-fd987dcb7-f2dx2" Oct 02 01:58:40 crc kubenswrapper[4885]: I1002 01:58:40.988951 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pd658\" (UniqueName: \"kubernetes.io/projected/e0315b09-692d-4c5f-8861-0863f90338fa-kube-api-access-pd658\") pod \"metallb-operator-webhook-server-fd987dcb7-f2dx2\" (UID: \"e0315b09-692d-4c5f-8861-0863f90338fa\") " pod="metallb-system/metallb-operator-webhook-server-fd987dcb7-f2dx2" Oct 02 01:58:41 crc kubenswrapper[4885]: I1002 01:58:41.065346 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-fd987dcb7-f2dx2" Oct 02 01:58:41 crc kubenswrapper[4885]: I1002 01:58:41.315291 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-fd987dcb7-f2dx2"] Oct 02 01:58:41 crc kubenswrapper[4885]: I1002 01:58:41.440782 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-6c54f448c5-tknpj"] Oct 02 01:58:41 crc kubenswrapper[4885]: W1002 01:58:41.446041 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7dc51fda_276d_4739_975d_e88c9fcec5d3.slice/crio-02ea157358c85deeeae1e4f493c3a0b27e87a5e6c48b97df72f7e75f49baea9b WatchSource:0}: Error finding container 02ea157358c85deeeae1e4f493c3a0b27e87a5e6c48b97df72f7e75f49baea9b: Status 404 returned error can't find the container with id 02ea157358c85deeeae1e4f493c3a0b27e87a5e6c48b97df72f7e75f49baea9b Oct 02 01:58:42 crc kubenswrapper[4885]: I1002 01:58:42.103984 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-fd987dcb7-f2dx2" event={"ID":"e0315b09-692d-4c5f-8861-0863f90338fa","Type":"ContainerStarted","Data":"14c35e110763b63e75dab021b406fde000bc61d401135a20c3ddcce388b17bd4"} Oct 02 01:58:42 crc kubenswrapper[4885]: I1002 01:58:42.105225 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-6c54f448c5-tknpj" event={"ID":"7dc51fda-276d-4739-975d-e88c9fcec5d3","Type":"ContainerStarted","Data":"02ea157358c85deeeae1e4f493c3a0b27e87a5e6c48b97df72f7e75f49baea9b"} Oct 02 01:58:47 crc kubenswrapper[4885]: I1002 01:58:47.133088 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-6c54f448c5-tknpj" event={"ID":"7dc51fda-276d-4739-975d-e88c9fcec5d3","Type":"ContainerStarted","Data":"36485228b99fc6d0acb999f7bf4567ba658fab1a0e07587f933f453db2b1902a"} Oct 02 01:58:47 crc kubenswrapper[4885]: I1002 01:58:47.134107 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-6c54f448c5-tknpj" Oct 02 01:58:47 crc kubenswrapper[4885]: I1002 01:58:47.135121 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-fd987dcb7-f2dx2" event={"ID":"e0315b09-692d-4c5f-8861-0863f90338fa","Type":"ContainerStarted","Data":"a8f00202b5341d6df3605d23e6fa0737330b2c5b1d8a18b4f160bd219a7baa0a"} Oct 02 01:58:47 crc kubenswrapper[4885]: I1002 01:58:47.135341 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-fd987dcb7-f2dx2" Oct 02 01:58:47 crc kubenswrapper[4885]: I1002 01:58:47.166852 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-6c54f448c5-tknpj" podStartSLOduration=2.194052963 podStartE2EDuration="7.166826526s" podCreationTimestamp="2025-10-02 01:58:40 +0000 UTC" firstStartedPulling="2025-10-02 01:58:41.44907344 +0000 UTC m=+710.260820849" lastFinishedPulling="2025-10-02 01:58:46.421847013 +0000 UTC m=+715.233594412" observedRunningTime="2025-10-02 01:58:47.159056661 +0000 UTC m=+715.970804080" watchObservedRunningTime="2025-10-02 01:58:47.166826526 +0000 UTC m=+715.978573945" Oct 02 01:58:47 crc kubenswrapper[4885]: I1002 01:58:47.180561 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-fd987dcb7-f2dx2" podStartSLOduration=2.0516162 podStartE2EDuration="7.180538423s" podCreationTimestamp="2025-10-02 01:58:40 +0000 UTC" firstStartedPulling="2025-10-02 01:58:41.327714536 +0000 UTC m=+710.139461935" lastFinishedPulling="2025-10-02 01:58:46.456636759 +0000 UTC m=+715.268384158" observedRunningTime="2025-10-02 01:58:47.178761789 +0000 UTC m=+715.990509198" watchObservedRunningTime="2025-10-02 01:58:47.180538423 +0000 UTC m=+715.992285832" Oct 02 01:59:01 crc kubenswrapper[4885]: I1002 01:59:01.073215 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-fd987dcb7-f2dx2" Oct 02 01:59:13 crc kubenswrapper[4885]: I1002 01:59:13.265840 4885 patch_prober.go:28] interesting pod/machine-config-daemon-rttx8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 01:59:13 crc kubenswrapper[4885]: I1002 01:59:13.266515 4885 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 01:59:19 crc kubenswrapper[4885]: I1002 01:59:19.656631 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-vs6d8"] Oct 02 01:59:19 crc kubenswrapper[4885]: I1002 01:59:19.658019 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-vs6d8" podUID="075528cd-eab2-47bb-8aa0-3d39002fb3d1" containerName="controller-manager" containerID="cri-o://0334e77510540e98449dacc7eddb4b6b9513daf0f37d0d7d39fc07f8b59b004c" gracePeriod=30 Oct 02 01:59:19 crc kubenswrapper[4885]: I1002 01:59:19.740248 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-g4vj4"] Oct 02 01:59:19 crc kubenswrapper[4885]: I1002 01:59:19.740676 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-g4vj4" podUID="45e90d9b-8520-4b1f-8f16-170787f4c8b6" containerName="route-controller-manager" containerID="cri-o://84fc3ecbcb85114c98bdfaf8649d09eab5466a1e07ec102855f9e750ec6196ba" gracePeriod=30 Oct 02 01:59:20 crc kubenswrapper[4885]: I1002 01:59:20.010832 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-vs6d8" Oct 02 01:59:20 crc kubenswrapper[4885]: I1002 01:59:20.073752 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-g4vj4" Oct 02 01:59:20 crc kubenswrapper[4885]: I1002 01:59:20.152766 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/075528cd-eab2-47bb-8aa0-3d39002fb3d1-serving-cert\") pod \"075528cd-eab2-47bb-8aa0-3d39002fb3d1\" (UID: \"075528cd-eab2-47bb-8aa0-3d39002fb3d1\") " Oct 02 01:59:20 crc kubenswrapper[4885]: I1002 01:59:20.152819 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/075528cd-eab2-47bb-8aa0-3d39002fb3d1-client-ca\") pod \"075528cd-eab2-47bb-8aa0-3d39002fb3d1\" (UID: \"075528cd-eab2-47bb-8aa0-3d39002fb3d1\") " Oct 02 01:59:20 crc kubenswrapper[4885]: I1002 01:59:20.152854 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/075528cd-eab2-47bb-8aa0-3d39002fb3d1-proxy-ca-bundles\") pod \"075528cd-eab2-47bb-8aa0-3d39002fb3d1\" (UID: \"075528cd-eab2-47bb-8aa0-3d39002fb3d1\") " Oct 02 01:59:20 crc kubenswrapper[4885]: I1002 01:59:20.152897 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vw9xq\" (UniqueName: \"kubernetes.io/projected/075528cd-eab2-47bb-8aa0-3d39002fb3d1-kube-api-access-vw9xq\") pod \"075528cd-eab2-47bb-8aa0-3d39002fb3d1\" (UID: \"075528cd-eab2-47bb-8aa0-3d39002fb3d1\") " Oct 02 01:59:20 crc kubenswrapper[4885]: I1002 01:59:20.152937 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/075528cd-eab2-47bb-8aa0-3d39002fb3d1-config\") pod \"075528cd-eab2-47bb-8aa0-3d39002fb3d1\" (UID: \"075528cd-eab2-47bb-8aa0-3d39002fb3d1\") " Oct 02 01:59:20 crc kubenswrapper[4885]: I1002 01:59:20.154245 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/075528cd-eab2-47bb-8aa0-3d39002fb3d1-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "075528cd-eab2-47bb-8aa0-3d39002fb3d1" (UID: "075528cd-eab2-47bb-8aa0-3d39002fb3d1"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:59:20 crc kubenswrapper[4885]: I1002 01:59:20.154335 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/075528cd-eab2-47bb-8aa0-3d39002fb3d1-client-ca" (OuterVolumeSpecName: "client-ca") pod "075528cd-eab2-47bb-8aa0-3d39002fb3d1" (UID: "075528cd-eab2-47bb-8aa0-3d39002fb3d1"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:59:20 crc kubenswrapper[4885]: I1002 01:59:20.154370 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/075528cd-eab2-47bb-8aa0-3d39002fb3d1-config" (OuterVolumeSpecName: "config") pod "075528cd-eab2-47bb-8aa0-3d39002fb3d1" (UID: "075528cd-eab2-47bb-8aa0-3d39002fb3d1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:59:20 crc kubenswrapper[4885]: I1002 01:59:20.158421 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/075528cd-eab2-47bb-8aa0-3d39002fb3d1-kube-api-access-vw9xq" (OuterVolumeSpecName: "kube-api-access-vw9xq") pod "075528cd-eab2-47bb-8aa0-3d39002fb3d1" (UID: "075528cd-eab2-47bb-8aa0-3d39002fb3d1"). InnerVolumeSpecName "kube-api-access-vw9xq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:59:20 crc kubenswrapper[4885]: I1002 01:59:20.158460 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/075528cd-eab2-47bb-8aa0-3d39002fb3d1-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "075528cd-eab2-47bb-8aa0-3d39002fb3d1" (UID: "075528cd-eab2-47bb-8aa0-3d39002fb3d1"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:59:20 crc kubenswrapper[4885]: I1002 01:59:20.254049 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/45e90d9b-8520-4b1f-8f16-170787f4c8b6-client-ca\") pod \"45e90d9b-8520-4b1f-8f16-170787f4c8b6\" (UID: \"45e90d9b-8520-4b1f-8f16-170787f4c8b6\") " Oct 02 01:59:20 crc kubenswrapper[4885]: I1002 01:59:20.254207 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/45e90d9b-8520-4b1f-8f16-170787f4c8b6-serving-cert\") pod \"45e90d9b-8520-4b1f-8f16-170787f4c8b6\" (UID: \"45e90d9b-8520-4b1f-8f16-170787f4c8b6\") " Oct 02 01:59:20 crc kubenswrapper[4885]: I1002 01:59:20.254253 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mpmq4\" (UniqueName: \"kubernetes.io/projected/45e90d9b-8520-4b1f-8f16-170787f4c8b6-kube-api-access-mpmq4\") pod \"45e90d9b-8520-4b1f-8f16-170787f4c8b6\" (UID: \"45e90d9b-8520-4b1f-8f16-170787f4c8b6\") " Oct 02 01:59:20 crc kubenswrapper[4885]: I1002 01:59:20.254305 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/45e90d9b-8520-4b1f-8f16-170787f4c8b6-config\") pod \"45e90d9b-8520-4b1f-8f16-170787f4c8b6\" (UID: \"45e90d9b-8520-4b1f-8f16-170787f4c8b6\") " Oct 02 01:59:20 crc kubenswrapper[4885]: I1002 01:59:20.254565 4885 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/075528cd-eab2-47bb-8aa0-3d39002fb3d1-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 01:59:20 crc kubenswrapper[4885]: I1002 01:59:20.254581 4885 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/075528cd-eab2-47bb-8aa0-3d39002fb3d1-client-ca\") on node \"crc\" DevicePath \"\"" Oct 02 01:59:20 crc kubenswrapper[4885]: I1002 01:59:20.254592 4885 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/075528cd-eab2-47bb-8aa0-3d39002fb3d1-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 02 01:59:20 crc kubenswrapper[4885]: I1002 01:59:20.254604 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vw9xq\" (UniqueName: \"kubernetes.io/projected/075528cd-eab2-47bb-8aa0-3d39002fb3d1-kube-api-access-vw9xq\") on node \"crc\" DevicePath \"\"" Oct 02 01:59:20 crc kubenswrapper[4885]: I1002 01:59:20.254615 4885 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/075528cd-eab2-47bb-8aa0-3d39002fb3d1-config\") on node \"crc\" DevicePath \"\"" Oct 02 01:59:20 crc kubenswrapper[4885]: I1002 01:59:20.255213 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/45e90d9b-8520-4b1f-8f16-170787f4c8b6-client-ca" (OuterVolumeSpecName: "client-ca") pod "45e90d9b-8520-4b1f-8f16-170787f4c8b6" (UID: "45e90d9b-8520-4b1f-8f16-170787f4c8b6"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:59:20 crc kubenswrapper[4885]: I1002 01:59:20.255440 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/45e90d9b-8520-4b1f-8f16-170787f4c8b6-config" (OuterVolumeSpecName: "config") pod "45e90d9b-8520-4b1f-8f16-170787f4c8b6" (UID: "45e90d9b-8520-4b1f-8f16-170787f4c8b6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:59:20 crc kubenswrapper[4885]: I1002 01:59:20.257554 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/45e90d9b-8520-4b1f-8f16-170787f4c8b6-kube-api-access-mpmq4" (OuterVolumeSpecName: "kube-api-access-mpmq4") pod "45e90d9b-8520-4b1f-8f16-170787f4c8b6" (UID: "45e90d9b-8520-4b1f-8f16-170787f4c8b6"). InnerVolumeSpecName "kube-api-access-mpmq4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:59:20 crc kubenswrapper[4885]: I1002 01:59:20.258019 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/45e90d9b-8520-4b1f-8f16-170787f4c8b6-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "45e90d9b-8520-4b1f-8f16-170787f4c8b6" (UID: "45e90d9b-8520-4b1f-8f16-170787f4c8b6"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:59:20 crc kubenswrapper[4885]: I1002 01:59:20.356357 4885 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/45e90d9b-8520-4b1f-8f16-170787f4c8b6-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 01:59:20 crc kubenswrapper[4885]: I1002 01:59:20.356408 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mpmq4\" (UniqueName: \"kubernetes.io/projected/45e90d9b-8520-4b1f-8f16-170787f4c8b6-kube-api-access-mpmq4\") on node \"crc\" DevicePath \"\"" Oct 02 01:59:20 crc kubenswrapper[4885]: I1002 01:59:20.356431 4885 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/45e90d9b-8520-4b1f-8f16-170787f4c8b6-config\") on node \"crc\" DevicePath \"\"" Oct 02 01:59:20 crc kubenswrapper[4885]: I1002 01:59:20.356448 4885 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/45e90d9b-8520-4b1f-8f16-170787f4c8b6-client-ca\") on node \"crc\" DevicePath \"\"" Oct 02 01:59:20 crc kubenswrapper[4885]: I1002 01:59:20.361658 4885 generic.go:334] "Generic (PLEG): container finished" podID="45e90d9b-8520-4b1f-8f16-170787f4c8b6" containerID="84fc3ecbcb85114c98bdfaf8649d09eab5466a1e07ec102855f9e750ec6196ba" exitCode=0 Oct 02 01:59:20 crc kubenswrapper[4885]: I1002 01:59:20.361753 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-g4vj4" event={"ID":"45e90d9b-8520-4b1f-8f16-170787f4c8b6","Type":"ContainerDied","Data":"84fc3ecbcb85114c98bdfaf8649d09eab5466a1e07ec102855f9e750ec6196ba"} Oct 02 01:59:20 crc kubenswrapper[4885]: I1002 01:59:20.361792 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-g4vj4" event={"ID":"45e90d9b-8520-4b1f-8f16-170787f4c8b6","Type":"ContainerDied","Data":"7830569e1288c370ed713a70e5eee29f8266a7e5dc69fe9fc372ac47eb9e8061"} Oct 02 01:59:20 crc kubenswrapper[4885]: I1002 01:59:20.361821 4885 scope.go:117] "RemoveContainer" containerID="84fc3ecbcb85114c98bdfaf8649d09eab5466a1e07ec102855f9e750ec6196ba" Oct 02 01:59:20 crc kubenswrapper[4885]: I1002 01:59:20.361964 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-g4vj4" Oct 02 01:59:20 crc kubenswrapper[4885]: I1002 01:59:20.371228 4885 generic.go:334] "Generic (PLEG): container finished" podID="075528cd-eab2-47bb-8aa0-3d39002fb3d1" containerID="0334e77510540e98449dacc7eddb4b6b9513daf0f37d0d7d39fc07f8b59b004c" exitCode=0 Oct 02 01:59:20 crc kubenswrapper[4885]: I1002 01:59:20.371334 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-vs6d8" event={"ID":"075528cd-eab2-47bb-8aa0-3d39002fb3d1","Type":"ContainerDied","Data":"0334e77510540e98449dacc7eddb4b6b9513daf0f37d0d7d39fc07f8b59b004c"} Oct 02 01:59:20 crc kubenswrapper[4885]: I1002 01:59:20.371388 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-vs6d8" event={"ID":"075528cd-eab2-47bb-8aa0-3d39002fb3d1","Type":"ContainerDied","Data":"9386f1ad17039deb3ae8e6c18844ef75711e75e80749203854287098ebd5f47b"} Oct 02 01:59:20 crc kubenswrapper[4885]: I1002 01:59:20.371334 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-vs6d8" Oct 02 01:59:20 crc kubenswrapper[4885]: I1002 01:59:20.405721 4885 scope.go:117] "RemoveContainer" containerID="84fc3ecbcb85114c98bdfaf8649d09eab5466a1e07ec102855f9e750ec6196ba" Oct 02 01:59:20 crc kubenswrapper[4885]: E1002 01:59:20.407873 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"84fc3ecbcb85114c98bdfaf8649d09eab5466a1e07ec102855f9e750ec6196ba\": container with ID starting with 84fc3ecbcb85114c98bdfaf8649d09eab5466a1e07ec102855f9e750ec6196ba not found: ID does not exist" containerID="84fc3ecbcb85114c98bdfaf8649d09eab5466a1e07ec102855f9e750ec6196ba" Oct 02 01:59:20 crc kubenswrapper[4885]: I1002 01:59:20.407932 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"84fc3ecbcb85114c98bdfaf8649d09eab5466a1e07ec102855f9e750ec6196ba"} err="failed to get container status \"84fc3ecbcb85114c98bdfaf8649d09eab5466a1e07ec102855f9e750ec6196ba\": rpc error: code = NotFound desc = could not find container \"84fc3ecbcb85114c98bdfaf8649d09eab5466a1e07ec102855f9e750ec6196ba\": container with ID starting with 84fc3ecbcb85114c98bdfaf8649d09eab5466a1e07ec102855f9e750ec6196ba not found: ID does not exist" Oct 02 01:59:20 crc kubenswrapper[4885]: I1002 01:59:20.407964 4885 scope.go:117] "RemoveContainer" containerID="0334e77510540e98449dacc7eddb4b6b9513daf0f37d0d7d39fc07f8b59b004c" Oct 02 01:59:20 crc kubenswrapper[4885]: I1002 01:59:20.422144 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-g4vj4"] Oct 02 01:59:20 crc kubenswrapper[4885]: I1002 01:59:20.425349 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-g4vj4"] Oct 02 01:59:20 crc kubenswrapper[4885]: I1002 01:59:20.437463 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-vs6d8"] Oct 02 01:59:20 crc kubenswrapper[4885]: I1002 01:59:20.441241 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-vs6d8"] Oct 02 01:59:20 crc kubenswrapper[4885]: I1002 01:59:20.443423 4885 scope.go:117] "RemoveContainer" containerID="0334e77510540e98449dacc7eddb4b6b9513daf0f37d0d7d39fc07f8b59b004c" Oct 02 01:59:20 crc kubenswrapper[4885]: E1002 01:59:20.446502 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0334e77510540e98449dacc7eddb4b6b9513daf0f37d0d7d39fc07f8b59b004c\": container with ID starting with 0334e77510540e98449dacc7eddb4b6b9513daf0f37d0d7d39fc07f8b59b004c not found: ID does not exist" containerID="0334e77510540e98449dacc7eddb4b6b9513daf0f37d0d7d39fc07f8b59b004c" Oct 02 01:59:20 crc kubenswrapper[4885]: I1002 01:59:20.446542 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0334e77510540e98449dacc7eddb4b6b9513daf0f37d0d7d39fc07f8b59b004c"} err="failed to get container status \"0334e77510540e98449dacc7eddb4b6b9513daf0f37d0d7d39fc07f8b59b004c\": rpc error: code = NotFound desc = could not find container \"0334e77510540e98449dacc7eddb4b6b9513daf0f37d0d7d39fc07f8b59b004c\": container with ID starting with 0334e77510540e98449dacc7eddb4b6b9513daf0f37d0d7d39fc07f8b59b004c not found: ID does not exist" Oct 02 01:59:20 crc kubenswrapper[4885]: I1002 01:59:20.841971 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-f8bcd69bb-lzxjz"] Oct 02 01:59:20 crc kubenswrapper[4885]: E1002 01:59:20.843882 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45e90d9b-8520-4b1f-8f16-170787f4c8b6" containerName="route-controller-manager" Oct 02 01:59:20 crc kubenswrapper[4885]: I1002 01:59:20.844067 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="45e90d9b-8520-4b1f-8f16-170787f4c8b6" containerName="route-controller-manager" Oct 02 01:59:20 crc kubenswrapper[4885]: E1002 01:59:20.844235 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="075528cd-eab2-47bb-8aa0-3d39002fb3d1" containerName="controller-manager" Oct 02 01:59:20 crc kubenswrapper[4885]: I1002 01:59:20.844438 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="075528cd-eab2-47bb-8aa0-3d39002fb3d1" containerName="controller-manager" Oct 02 01:59:20 crc kubenswrapper[4885]: I1002 01:59:20.844823 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="45e90d9b-8520-4b1f-8f16-170787f4c8b6" containerName="route-controller-manager" Oct 02 01:59:20 crc kubenswrapper[4885]: I1002 01:59:20.845043 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="075528cd-eab2-47bb-8aa0-3d39002fb3d1" containerName="controller-manager" Oct 02 01:59:20 crc kubenswrapper[4885]: I1002 01:59:20.845824 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-f8bcd69bb-lzxjz" Oct 02 01:59:20 crc kubenswrapper[4885]: I1002 01:59:20.848990 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 02 01:59:20 crc kubenswrapper[4885]: I1002 01:59:20.849335 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 02 01:59:20 crc kubenswrapper[4885]: I1002 01:59:20.849003 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 02 01:59:20 crc kubenswrapper[4885]: I1002 01:59:20.849644 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 02 01:59:20 crc kubenswrapper[4885]: I1002 01:59:20.849733 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 02 01:59:20 crc kubenswrapper[4885]: I1002 01:59:20.849895 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 02 01:59:20 crc kubenswrapper[4885]: I1002 01:59:20.850339 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-f8bcd69bb-lzxjz"] Oct 02 01:59:20 crc kubenswrapper[4885]: I1002 01:59:20.858254 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 02 01:59:20 crc kubenswrapper[4885]: I1002 01:59:20.897125 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fd6d9349-9419-433f-9e19-df003ca0912b-client-ca\") pod \"controller-manager-f8bcd69bb-lzxjz\" (UID: \"fd6d9349-9419-433f-9e19-df003ca0912b\") " pod="openshift-controller-manager/controller-manager-f8bcd69bb-lzxjz" Oct 02 01:59:20 crc kubenswrapper[4885]: I1002 01:59:20.897169 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd6d9349-9419-433f-9e19-df003ca0912b-config\") pod \"controller-manager-f8bcd69bb-lzxjz\" (UID: \"fd6d9349-9419-433f-9e19-df003ca0912b\") " pod="openshift-controller-manager/controller-manager-f8bcd69bb-lzxjz" Oct 02 01:59:20 crc kubenswrapper[4885]: I1002 01:59:20.938658 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-6c54f448c5-tknpj" Oct 02 01:59:20 crc kubenswrapper[4885]: I1002 01:59:20.999196 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/fd6d9349-9419-433f-9e19-df003ca0912b-proxy-ca-bundles\") pod \"controller-manager-f8bcd69bb-lzxjz\" (UID: \"fd6d9349-9419-433f-9e19-df003ca0912b\") " pod="openshift-controller-manager/controller-manager-f8bcd69bb-lzxjz" Oct 02 01:59:20 crc kubenswrapper[4885]: I1002 01:59:20.999338 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fd6d9349-9419-433f-9e19-df003ca0912b-client-ca\") pod \"controller-manager-f8bcd69bb-lzxjz\" (UID: \"fd6d9349-9419-433f-9e19-df003ca0912b\") " pod="openshift-controller-manager/controller-manager-f8bcd69bb-lzxjz" Oct 02 01:59:20 crc kubenswrapper[4885]: I1002 01:59:20.999376 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd6d9349-9419-433f-9e19-df003ca0912b-config\") pod \"controller-manager-f8bcd69bb-lzxjz\" (UID: \"fd6d9349-9419-433f-9e19-df003ca0912b\") " pod="openshift-controller-manager/controller-manager-f8bcd69bb-lzxjz" Oct 02 01:59:20 crc kubenswrapper[4885]: I1002 01:59:20.999451 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-25jvn\" (UniqueName: \"kubernetes.io/projected/fd6d9349-9419-433f-9e19-df003ca0912b-kube-api-access-25jvn\") pod \"controller-manager-f8bcd69bb-lzxjz\" (UID: \"fd6d9349-9419-433f-9e19-df003ca0912b\") " pod="openshift-controller-manager/controller-manager-f8bcd69bb-lzxjz" Oct 02 01:59:20 crc kubenswrapper[4885]: I1002 01:59:20.999501 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fd6d9349-9419-433f-9e19-df003ca0912b-serving-cert\") pod \"controller-manager-f8bcd69bb-lzxjz\" (UID: \"fd6d9349-9419-433f-9e19-df003ca0912b\") " pod="openshift-controller-manager/controller-manager-f8bcd69bb-lzxjz" Oct 02 01:59:21 crc kubenswrapper[4885]: I1002 01:59:21.002170 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fd6d9349-9419-433f-9e19-df003ca0912b-client-ca\") pod \"controller-manager-f8bcd69bb-lzxjz\" (UID: \"fd6d9349-9419-433f-9e19-df003ca0912b\") " pod="openshift-controller-manager/controller-manager-f8bcd69bb-lzxjz" Oct 02 01:59:21 crc kubenswrapper[4885]: I1002 01:59:21.003361 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd6d9349-9419-433f-9e19-df003ca0912b-config\") pod \"controller-manager-f8bcd69bb-lzxjz\" (UID: \"fd6d9349-9419-433f-9e19-df003ca0912b\") " pod="openshift-controller-manager/controller-manager-f8bcd69bb-lzxjz" Oct 02 01:59:21 crc kubenswrapper[4885]: I1002 01:59:21.101098 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/fd6d9349-9419-433f-9e19-df003ca0912b-proxy-ca-bundles\") pod \"controller-manager-f8bcd69bb-lzxjz\" (UID: \"fd6d9349-9419-433f-9e19-df003ca0912b\") " pod="openshift-controller-manager/controller-manager-f8bcd69bb-lzxjz" Oct 02 01:59:21 crc kubenswrapper[4885]: I1002 01:59:21.101223 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-25jvn\" (UniqueName: \"kubernetes.io/projected/fd6d9349-9419-433f-9e19-df003ca0912b-kube-api-access-25jvn\") pod \"controller-manager-f8bcd69bb-lzxjz\" (UID: \"fd6d9349-9419-433f-9e19-df003ca0912b\") " pod="openshift-controller-manager/controller-manager-f8bcd69bb-lzxjz" Oct 02 01:59:21 crc kubenswrapper[4885]: I1002 01:59:21.101326 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fd6d9349-9419-433f-9e19-df003ca0912b-serving-cert\") pod \"controller-manager-f8bcd69bb-lzxjz\" (UID: \"fd6d9349-9419-433f-9e19-df003ca0912b\") " pod="openshift-controller-manager/controller-manager-f8bcd69bb-lzxjz" Oct 02 01:59:21 crc kubenswrapper[4885]: I1002 01:59:21.102278 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/fd6d9349-9419-433f-9e19-df003ca0912b-proxy-ca-bundles\") pod \"controller-manager-f8bcd69bb-lzxjz\" (UID: \"fd6d9349-9419-433f-9e19-df003ca0912b\") " pod="openshift-controller-manager/controller-manager-f8bcd69bb-lzxjz" Oct 02 01:59:21 crc kubenswrapper[4885]: I1002 01:59:21.107585 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fd6d9349-9419-433f-9e19-df003ca0912b-serving-cert\") pod \"controller-manager-f8bcd69bb-lzxjz\" (UID: \"fd6d9349-9419-433f-9e19-df003ca0912b\") " pod="openshift-controller-manager/controller-manager-f8bcd69bb-lzxjz" Oct 02 01:59:21 crc kubenswrapper[4885]: I1002 01:59:21.135530 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-25jvn\" (UniqueName: \"kubernetes.io/projected/fd6d9349-9419-433f-9e19-df003ca0912b-kube-api-access-25jvn\") pod \"controller-manager-f8bcd69bb-lzxjz\" (UID: \"fd6d9349-9419-433f-9e19-df003ca0912b\") " pod="openshift-controller-manager/controller-manager-f8bcd69bb-lzxjz" Oct 02 01:59:21 crc kubenswrapper[4885]: I1002 01:59:21.180254 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-f8bcd69bb-lzxjz" Oct 02 01:59:21 crc kubenswrapper[4885]: I1002 01:59:21.667643 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-f8bcd69bb-lzxjz"] Oct 02 01:59:21 crc kubenswrapper[4885]: I1002 01:59:21.799024 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-ggsh4"] Oct 02 01:59:21 crc kubenswrapper[4885]: I1002 01:59:21.801187 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-ggsh4" Oct 02 01:59:21 crc kubenswrapper[4885]: I1002 01:59:21.804054 4885 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-jrkmx" Oct 02 01:59:21 crc kubenswrapper[4885]: I1002 01:59:21.804242 4885 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Oct 02 01:59:21 crc kubenswrapper[4885]: I1002 01:59:21.804439 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Oct 02 01:59:21 crc kubenswrapper[4885]: I1002 01:59:21.809099 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-q7xtq"] Oct 02 01:59:21 crc kubenswrapper[4885]: I1002 01:59:21.809675 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-q7xtq" Oct 02 01:59:21 crc kubenswrapper[4885]: I1002 01:59:21.809963 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/05715b27-682b-46a1-80b5-667667147ce0-frr-conf\") pod \"frr-k8s-ggsh4\" (UID: \"05715b27-682b-46a1-80b5-667667147ce0\") " pod="metallb-system/frr-k8s-ggsh4" Oct 02 01:59:21 crc kubenswrapper[4885]: I1002 01:59:21.810025 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/05715b27-682b-46a1-80b5-667667147ce0-metrics-certs\") pod \"frr-k8s-ggsh4\" (UID: \"05715b27-682b-46a1-80b5-667667147ce0\") " pod="metallb-system/frr-k8s-ggsh4" Oct 02 01:59:21 crc kubenswrapper[4885]: I1002 01:59:21.810059 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2wrjp\" (UniqueName: \"kubernetes.io/projected/05715b27-682b-46a1-80b5-667667147ce0-kube-api-access-2wrjp\") pod \"frr-k8s-ggsh4\" (UID: \"05715b27-682b-46a1-80b5-667667147ce0\") " pod="metallb-system/frr-k8s-ggsh4" Oct 02 01:59:21 crc kubenswrapper[4885]: I1002 01:59:21.810127 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/05715b27-682b-46a1-80b5-667667147ce0-reloader\") pod \"frr-k8s-ggsh4\" (UID: \"05715b27-682b-46a1-80b5-667667147ce0\") " pod="metallb-system/frr-k8s-ggsh4" Oct 02 01:59:21 crc kubenswrapper[4885]: I1002 01:59:21.810326 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/05715b27-682b-46a1-80b5-667667147ce0-frr-startup\") pod \"frr-k8s-ggsh4\" (UID: \"05715b27-682b-46a1-80b5-667667147ce0\") " pod="metallb-system/frr-k8s-ggsh4" Oct 02 01:59:21 crc kubenswrapper[4885]: I1002 01:59:21.810468 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/05715b27-682b-46a1-80b5-667667147ce0-metrics\") pod \"frr-k8s-ggsh4\" (UID: \"05715b27-682b-46a1-80b5-667667147ce0\") " pod="metallb-system/frr-k8s-ggsh4" Oct 02 01:59:21 crc kubenswrapper[4885]: I1002 01:59:21.810506 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/05715b27-682b-46a1-80b5-667667147ce0-frr-sockets\") pod \"frr-k8s-ggsh4\" (UID: \"05715b27-682b-46a1-80b5-667667147ce0\") " pod="metallb-system/frr-k8s-ggsh4" Oct 02 01:59:21 crc kubenswrapper[4885]: I1002 01:59:21.813885 4885 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Oct 02 01:59:21 crc kubenswrapper[4885]: I1002 01:59:21.818246 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-q7xtq"] Oct 02 01:59:21 crc kubenswrapper[4885]: I1002 01:59:21.840323 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-f76746f8f-znjc9"] Oct 02 01:59:21 crc kubenswrapper[4885]: I1002 01:59:21.840919 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-f76746f8f-znjc9" Oct 02 01:59:21 crc kubenswrapper[4885]: I1002 01:59:21.843158 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 02 01:59:21 crc kubenswrapper[4885]: I1002 01:59:21.843864 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 02 01:59:21 crc kubenswrapper[4885]: I1002 01:59:21.844108 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 02 01:59:21 crc kubenswrapper[4885]: I1002 01:59:21.844364 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 02 01:59:21 crc kubenswrapper[4885]: I1002 01:59:21.844624 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 02 01:59:21 crc kubenswrapper[4885]: I1002 01:59:21.845476 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 02 01:59:21 crc kubenswrapper[4885]: I1002 01:59:21.859096 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-f76746f8f-znjc9"] Oct 02 01:59:21 crc kubenswrapper[4885]: I1002 01:59:21.902522 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-5h9mk"] Oct 02 01:59:21 crc kubenswrapper[4885]: I1002 01:59:21.903300 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-5h9mk" Oct 02 01:59:21 crc kubenswrapper[4885]: I1002 01:59:21.906651 4885 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Oct 02 01:59:21 crc kubenswrapper[4885]: I1002 01:59:21.906766 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Oct 02 01:59:21 crc kubenswrapper[4885]: I1002 01:59:21.906770 4885 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Oct 02 01:59:21 crc kubenswrapper[4885]: I1002 01:59:21.907792 4885 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-4gp46" Oct 02 01:59:21 crc kubenswrapper[4885]: I1002 01:59:21.911467 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/05715b27-682b-46a1-80b5-667667147ce0-metrics\") pod \"frr-k8s-ggsh4\" (UID: \"05715b27-682b-46a1-80b5-667667147ce0\") " pod="metallb-system/frr-k8s-ggsh4" Oct 02 01:59:21 crc kubenswrapper[4885]: I1002 01:59:21.911501 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/05715b27-682b-46a1-80b5-667667147ce0-frr-sockets\") pod \"frr-k8s-ggsh4\" (UID: \"05715b27-682b-46a1-80b5-667667147ce0\") " pod="metallb-system/frr-k8s-ggsh4" Oct 02 01:59:21 crc kubenswrapper[4885]: I1002 01:59:21.911539 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/e3d6e1b0-5963-4e8f-bf3d-eb0e88ec0395-memberlist\") pod \"speaker-5h9mk\" (UID: \"e3d6e1b0-5963-4e8f-bf3d-eb0e88ec0395\") " pod="metallb-system/speaker-5h9mk" Oct 02 01:59:21 crc kubenswrapper[4885]: I1002 01:59:21.911557 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/05715b27-682b-46a1-80b5-667667147ce0-frr-conf\") pod \"frr-k8s-ggsh4\" (UID: \"05715b27-682b-46a1-80b5-667667147ce0\") " pod="metallb-system/frr-k8s-ggsh4" Oct 02 01:59:21 crc kubenswrapper[4885]: I1002 01:59:21.911575 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/e3d6e1b0-5963-4e8f-bf3d-eb0e88ec0395-metallb-excludel2\") pod \"speaker-5h9mk\" (UID: \"e3d6e1b0-5963-4e8f-bf3d-eb0e88ec0395\") " pod="metallb-system/speaker-5h9mk" Oct 02 01:59:21 crc kubenswrapper[4885]: I1002 01:59:21.911668 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2678e135-47c6-4727-a31c-cbdfc16448bc-serving-cert\") pod \"route-controller-manager-f76746f8f-znjc9\" (UID: \"2678e135-47c6-4727-a31c-cbdfc16448bc\") " pod="openshift-route-controller-manager/route-controller-manager-f76746f8f-znjc9" Oct 02 01:59:21 crc kubenswrapper[4885]: I1002 01:59:21.911700 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/05715b27-682b-46a1-80b5-667667147ce0-metrics-certs\") pod \"frr-k8s-ggsh4\" (UID: \"05715b27-682b-46a1-80b5-667667147ce0\") " pod="metallb-system/frr-k8s-ggsh4" Oct 02 01:59:21 crc kubenswrapper[4885]: E1002 01:59:21.911790 4885 secret.go:188] Couldn't get secret metallb-system/frr-k8s-certs-secret: secret "frr-k8s-certs-secret" not found Oct 02 01:59:21 crc kubenswrapper[4885]: E1002 01:59:21.911841 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/05715b27-682b-46a1-80b5-667667147ce0-metrics-certs podName:05715b27-682b-46a1-80b5-667667147ce0 nodeName:}" failed. No retries permitted until 2025-10-02 01:59:22.411823614 +0000 UTC m=+751.223571013 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/05715b27-682b-46a1-80b5-667667147ce0-metrics-certs") pod "frr-k8s-ggsh4" (UID: "05715b27-682b-46a1-80b5-667667147ce0") : secret "frr-k8s-certs-secret" not found Oct 02 01:59:21 crc kubenswrapper[4885]: I1002 01:59:21.912038 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/05715b27-682b-46a1-80b5-667667147ce0-frr-conf\") pod \"frr-k8s-ggsh4\" (UID: \"05715b27-682b-46a1-80b5-667667147ce0\") " pod="metallb-system/frr-k8s-ggsh4" Oct 02 01:59:21 crc kubenswrapper[4885]: I1002 01:59:21.912048 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2wrjp\" (UniqueName: \"kubernetes.io/projected/05715b27-682b-46a1-80b5-667667147ce0-kube-api-access-2wrjp\") pod \"frr-k8s-ggsh4\" (UID: \"05715b27-682b-46a1-80b5-667667147ce0\") " pod="metallb-system/frr-k8s-ggsh4" Oct 02 01:59:21 crc kubenswrapper[4885]: I1002 01:59:21.912106 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2678e135-47c6-4727-a31c-cbdfc16448bc-client-ca\") pod \"route-controller-manager-f76746f8f-znjc9\" (UID: \"2678e135-47c6-4727-a31c-cbdfc16448bc\") " pod="openshift-route-controller-manager/route-controller-manager-f76746f8f-znjc9" Oct 02 01:59:21 crc kubenswrapper[4885]: I1002 01:59:21.912135 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e3d6e1b0-5963-4e8f-bf3d-eb0e88ec0395-metrics-certs\") pod \"speaker-5h9mk\" (UID: \"e3d6e1b0-5963-4e8f-bf3d-eb0e88ec0395\") " pod="metallb-system/speaker-5h9mk" Oct 02 01:59:21 crc kubenswrapper[4885]: I1002 01:59:21.912160 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lkfmz\" (UniqueName: \"kubernetes.io/projected/66663f53-9790-48b4-8d16-4675661f340d-kube-api-access-lkfmz\") pod \"frr-k8s-webhook-server-64bf5d555-q7xtq\" (UID: \"66663f53-9790-48b4-8d16-4675661f340d\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-q7xtq" Oct 02 01:59:21 crc kubenswrapper[4885]: I1002 01:59:21.912186 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/66663f53-9790-48b4-8d16-4675661f340d-cert\") pod \"frr-k8s-webhook-server-64bf5d555-q7xtq\" (UID: \"66663f53-9790-48b4-8d16-4675661f340d\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-q7xtq" Oct 02 01:59:21 crc kubenswrapper[4885]: I1002 01:59:21.912213 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r84kv\" (UniqueName: \"kubernetes.io/projected/e3d6e1b0-5963-4e8f-bf3d-eb0e88ec0395-kube-api-access-r84kv\") pod \"speaker-5h9mk\" (UID: \"e3d6e1b0-5963-4e8f-bf3d-eb0e88ec0395\") " pod="metallb-system/speaker-5h9mk" Oct 02 01:59:21 crc kubenswrapper[4885]: I1002 01:59:21.912242 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/05715b27-682b-46a1-80b5-667667147ce0-reloader\") pod \"frr-k8s-ggsh4\" (UID: \"05715b27-682b-46a1-80b5-667667147ce0\") " pod="metallb-system/frr-k8s-ggsh4" Oct 02 01:59:21 crc kubenswrapper[4885]: I1002 01:59:21.912248 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/05715b27-682b-46a1-80b5-667667147ce0-metrics\") pod \"frr-k8s-ggsh4\" (UID: \"05715b27-682b-46a1-80b5-667667147ce0\") " pod="metallb-system/frr-k8s-ggsh4" Oct 02 01:59:21 crc kubenswrapper[4885]: I1002 01:59:21.912328 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/05715b27-682b-46a1-80b5-667667147ce0-frr-startup\") pod \"frr-k8s-ggsh4\" (UID: \"05715b27-682b-46a1-80b5-667667147ce0\") " pod="metallb-system/frr-k8s-ggsh4" Oct 02 01:59:21 crc kubenswrapper[4885]: I1002 01:59:21.912372 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zpplk\" (UniqueName: \"kubernetes.io/projected/2678e135-47c6-4727-a31c-cbdfc16448bc-kube-api-access-zpplk\") pod \"route-controller-manager-f76746f8f-znjc9\" (UID: \"2678e135-47c6-4727-a31c-cbdfc16448bc\") " pod="openshift-route-controller-manager/route-controller-manager-f76746f8f-znjc9" Oct 02 01:59:21 crc kubenswrapper[4885]: I1002 01:59:21.912431 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2678e135-47c6-4727-a31c-cbdfc16448bc-config\") pod \"route-controller-manager-f76746f8f-znjc9\" (UID: \"2678e135-47c6-4727-a31c-cbdfc16448bc\") " pod="openshift-route-controller-manager/route-controller-manager-f76746f8f-znjc9" Oct 02 01:59:21 crc kubenswrapper[4885]: I1002 01:59:21.912434 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/05715b27-682b-46a1-80b5-667667147ce0-frr-sockets\") pod \"frr-k8s-ggsh4\" (UID: \"05715b27-682b-46a1-80b5-667667147ce0\") " pod="metallb-system/frr-k8s-ggsh4" Oct 02 01:59:21 crc kubenswrapper[4885]: I1002 01:59:21.912465 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/05715b27-682b-46a1-80b5-667667147ce0-reloader\") pod \"frr-k8s-ggsh4\" (UID: \"05715b27-682b-46a1-80b5-667667147ce0\") " pod="metallb-system/frr-k8s-ggsh4" Oct 02 01:59:21 crc kubenswrapper[4885]: I1002 01:59:21.913251 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/05715b27-682b-46a1-80b5-667667147ce0-frr-startup\") pod \"frr-k8s-ggsh4\" (UID: \"05715b27-682b-46a1-80b5-667667147ce0\") " pod="metallb-system/frr-k8s-ggsh4" Oct 02 01:59:21 crc kubenswrapper[4885]: I1002 01:59:21.920102 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-68d546b9d8-djznn"] Oct 02 01:59:21 crc kubenswrapper[4885]: I1002 01:59:21.920910 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-djznn" Oct 02 01:59:21 crc kubenswrapper[4885]: I1002 01:59:21.926153 4885 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Oct 02 01:59:21 crc kubenswrapper[4885]: I1002 01:59:21.941760 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-djznn"] Oct 02 01:59:21 crc kubenswrapper[4885]: I1002 01:59:21.955195 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2wrjp\" (UniqueName: \"kubernetes.io/projected/05715b27-682b-46a1-80b5-667667147ce0-kube-api-access-2wrjp\") pod \"frr-k8s-ggsh4\" (UID: \"05715b27-682b-46a1-80b5-667667147ce0\") " pod="metallb-system/frr-k8s-ggsh4" Oct 02 01:59:22 crc kubenswrapper[4885]: I1002 01:59:22.013123 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2678e135-47c6-4727-a31c-cbdfc16448bc-client-ca\") pod \"route-controller-manager-f76746f8f-znjc9\" (UID: \"2678e135-47c6-4727-a31c-cbdfc16448bc\") " pod="openshift-route-controller-manager/route-controller-manager-f76746f8f-znjc9" Oct 02 01:59:22 crc kubenswrapper[4885]: I1002 01:59:22.013161 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e3d6e1b0-5963-4e8f-bf3d-eb0e88ec0395-metrics-certs\") pod \"speaker-5h9mk\" (UID: \"e3d6e1b0-5963-4e8f-bf3d-eb0e88ec0395\") " pod="metallb-system/speaker-5h9mk" Oct 02 01:59:22 crc kubenswrapper[4885]: I1002 01:59:22.013196 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lkfmz\" (UniqueName: \"kubernetes.io/projected/66663f53-9790-48b4-8d16-4675661f340d-kube-api-access-lkfmz\") pod \"frr-k8s-webhook-server-64bf5d555-q7xtq\" (UID: \"66663f53-9790-48b4-8d16-4675661f340d\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-q7xtq" Oct 02 01:59:22 crc kubenswrapper[4885]: I1002 01:59:22.013220 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/66663f53-9790-48b4-8d16-4675661f340d-cert\") pod \"frr-k8s-webhook-server-64bf5d555-q7xtq\" (UID: \"66663f53-9790-48b4-8d16-4675661f340d\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-q7xtq" Oct 02 01:59:22 crc kubenswrapper[4885]: I1002 01:59:22.013235 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r84kv\" (UniqueName: \"kubernetes.io/projected/e3d6e1b0-5963-4e8f-bf3d-eb0e88ec0395-kube-api-access-r84kv\") pod \"speaker-5h9mk\" (UID: \"e3d6e1b0-5963-4e8f-bf3d-eb0e88ec0395\") " pod="metallb-system/speaker-5h9mk" Oct 02 01:59:22 crc kubenswrapper[4885]: I1002 01:59:22.013282 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4nm24\" (UniqueName: \"kubernetes.io/projected/47f27b52-1453-4566-954f-94e6b02b2221-kube-api-access-4nm24\") pod \"controller-68d546b9d8-djznn\" (UID: \"47f27b52-1453-4566-954f-94e6b02b2221\") " pod="metallb-system/controller-68d546b9d8-djznn" Oct 02 01:59:22 crc kubenswrapper[4885]: I1002 01:59:22.013319 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zpplk\" (UniqueName: \"kubernetes.io/projected/2678e135-47c6-4727-a31c-cbdfc16448bc-kube-api-access-zpplk\") pod \"route-controller-manager-f76746f8f-znjc9\" (UID: \"2678e135-47c6-4727-a31c-cbdfc16448bc\") " pod="openshift-route-controller-manager/route-controller-manager-f76746f8f-znjc9" Oct 02 01:59:22 crc kubenswrapper[4885]: I1002 01:59:22.013356 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/47f27b52-1453-4566-954f-94e6b02b2221-metrics-certs\") pod \"controller-68d546b9d8-djznn\" (UID: \"47f27b52-1453-4566-954f-94e6b02b2221\") " pod="metallb-system/controller-68d546b9d8-djznn" Oct 02 01:59:22 crc kubenswrapper[4885]: I1002 01:59:22.013389 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2678e135-47c6-4727-a31c-cbdfc16448bc-config\") pod \"route-controller-manager-f76746f8f-znjc9\" (UID: \"2678e135-47c6-4727-a31c-cbdfc16448bc\") " pod="openshift-route-controller-manager/route-controller-manager-f76746f8f-znjc9" Oct 02 01:59:22 crc kubenswrapper[4885]: E1002 01:59:22.013424 4885 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: secret "frr-k8s-webhook-server-cert" not found Oct 02 01:59:22 crc kubenswrapper[4885]: I1002 01:59:22.013454 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/e3d6e1b0-5963-4e8f-bf3d-eb0e88ec0395-memberlist\") pod \"speaker-5h9mk\" (UID: \"e3d6e1b0-5963-4e8f-bf3d-eb0e88ec0395\") " pod="metallb-system/speaker-5h9mk" Oct 02 01:59:22 crc kubenswrapper[4885]: I1002 01:59:22.013481 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/e3d6e1b0-5963-4e8f-bf3d-eb0e88ec0395-metallb-excludel2\") pod \"speaker-5h9mk\" (UID: \"e3d6e1b0-5963-4e8f-bf3d-eb0e88ec0395\") " pod="metallb-system/speaker-5h9mk" Oct 02 01:59:22 crc kubenswrapper[4885]: E1002 01:59:22.013516 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/66663f53-9790-48b4-8d16-4675661f340d-cert podName:66663f53-9790-48b4-8d16-4675661f340d nodeName:}" failed. No retries permitted until 2025-10-02 01:59:22.513489353 +0000 UTC m=+751.325236752 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/66663f53-9790-48b4-8d16-4675661f340d-cert") pod "frr-k8s-webhook-server-64bf5d555-q7xtq" (UID: "66663f53-9790-48b4-8d16-4675661f340d") : secret "frr-k8s-webhook-server-cert" not found Oct 02 01:59:22 crc kubenswrapper[4885]: I1002 01:59:22.013546 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2678e135-47c6-4727-a31c-cbdfc16448bc-serving-cert\") pod \"route-controller-manager-f76746f8f-znjc9\" (UID: \"2678e135-47c6-4727-a31c-cbdfc16448bc\") " pod="openshift-route-controller-manager/route-controller-manager-f76746f8f-znjc9" Oct 02 01:59:22 crc kubenswrapper[4885]: I1002 01:59:22.013595 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/47f27b52-1453-4566-954f-94e6b02b2221-cert\") pod \"controller-68d546b9d8-djznn\" (UID: \"47f27b52-1453-4566-954f-94e6b02b2221\") " pod="metallb-system/controller-68d546b9d8-djznn" Oct 02 01:59:22 crc kubenswrapper[4885]: E1002 01:59:22.013641 4885 secret.go:188] Couldn't get secret metallb-system/speaker-certs-secret: secret "speaker-certs-secret" not found Oct 02 01:59:22 crc kubenswrapper[4885]: E1002 01:59:22.013694 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e3d6e1b0-5963-4e8f-bf3d-eb0e88ec0395-metrics-certs podName:e3d6e1b0-5963-4e8f-bf3d-eb0e88ec0395 nodeName:}" failed. No retries permitted until 2025-10-02 01:59:22.513676069 +0000 UTC m=+751.325423468 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e3d6e1b0-5963-4e8f-bf3d-eb0e88ec0395-metrics-certs") pod "speaker-5h9mk" (UID: "e3d6e1b0-5963-4e8f-bf3d-eb0e88ec0395") : secret "speaker-certs-secret" not found Oct 02 01:59:22 crc kubenswrapper[4885]: I1002 01:59:22.014025 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2678e135-47c6-4727-a31c-cbdfc16448bc-client-ca\") pod \"route-controller-manager-f76746f8f-znjc9\" (UID: \"2678e135-47c6-4727-a31c-cbdfc16448bc\") " pod="openshift-route-controller-manager/route-controller-manager-f76746f8f-znjc9" Oct 02 01:59:22 crc kubenswrapper[4885]: E1002 01:59:22.014125 4885 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 02 01:59:22 crc kubenswrapper[4885]: E1002 01:59:22.014178 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e3d6e1b0-5963-4e8f-bf3d-eb0e88ec0395-memberlist podName:e3d6e1b0-5963-4e8f-bf3d-eb0e88ec0395 nodeName:}" failed. No retries permitted until 2025-10-02 01:59:22.514162253 +0000 UTC m=+751.325909652 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/e3d6e1b0-5963-4e8f-bf3d-eb0e88ec0395-memberlist") pod "speaker-5h9mk" (UID: "e3d6e1b0-5963-4e8f-bf3d-eb0e88ec0395") : secret "metallb-memberlist" not found Oct 02 01:59:22 crc kubenswrapper[4885]: I1002 01:59:22.014192 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/e3d6e1b0-5963-4e8f-bf3d-eb0e88ec0395-metallb-excludel2\") pod \"speaker-5h9mk\" (UID: \"e3d6e1b0-5963-4e8f-bf3d-eb0e88ec0395\") " pod="metallb-system/speaker-5h9mk" Oct 02 01:59:22 crc kubenswrapper[4885]: I1002 01:59:22.014908 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2678e135-47c6-4727-a31c-cbdfc16448bc-config\") pod \"route-controller-manager-f76746f8f-znjc9\" (UID: \"2678e135-47c6-4727-a31c-cbdfc16448bc\") " pod="openshift-route-controller-manager/route-controller-manager-f76746f8f-znjc9" Oct 02 01:59:22 crc kubenswrapper[4885]: I1002 01:59:22.017294 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2678e135-47c6-4727-a31c-cbdfc16448bc-serving-cert\") pod \"route-controller-manager-f76746f8f-znjc9\" (UID: \"2678e135-47c6-4727-a31c-cbdfc16448bc\") " pod="openshift-route-controller-manager/route-controller-manager-f76746f8f-znjc9" Oct 02 01:59:22 crc kubenswrapper[4885]: I1002 01:59:22.035945 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zpplk\" (UniqueName: \"kubernetes.io/projected/2678e135-47c6-4727-a31c-cbdfc16448bc-kube-api-access-zpplk\") pod \"route-controller-manager-f76746f8f-znjc9\" (UID: \"2678e135-47c6-4727-a31c-cbdfc16448bc\") " pod="openshift-route-controller-manager/route-controller-manager-f76746f8f-znjc9" Oct 02 01:59:22 crc kubenswrapper[4885]: I1002 01:59:22.036045 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lkfmz\" (UniqueName: \"kubernetes.io/projected/66663f53-9790-48b4-8d16-4675661f340d-kube-api-access-lkfmz\") pod \"frr-k8s-webhook-server-64bf5d555-q7xtq\" (UID: \"66663f53-9790-48b4-8d16-4675661f340d\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-q7xtq" Oct 02 01:59:22 crc kubenswrapper[4885]: I1002 01:59:22.036794 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r84kv\" (UniqueName: \"kubernetes.io/projected/e3d6e1b0-5963-4e8f-bf3d-eb0e88ec0395-kube-api-access-r84kv\") pod \"speaker-5h9mk\" (UID: \"e3d6e1b0-5963-4e8f-bf3d-eb0e88ec0395\") " pod="metallb-system/speaker-5h9mk" Oct 02 01:59:22 crc kubenswrapper[4885]: I1002 01:59:22.057180 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="075528cd-eab2-47bb-8aa0-3d39002fb3d1" path="/var/lib/kubelet/pods/075528cd-eab2-47bb-8aa0-3d39002fb3d1/volumes" Oct 02 01:59:22 crc kubenswrapper[4885]: I1002 01:59:22.057820 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="45e90d9b-8520-4b1f-8f16-170787f4c8b6" path="/var/lib/kubelet/pods/45e90d9b-8520-4b1f-8f16-170787f4c8b6/volumes" Oct 02 01:59:22 crc kubenswrapper[4885]: I1002 01:59:22.114606 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4nm24\" (UniqueName: \"kubernetes.io/projected/47f27b52-1453-4566-954f-94e6b02b2221-kube-api-access-4nm24\") pod \"controller-68d546b9d8-djznn\" (UID: \"47f27b52-1453-4566-954f-94e6b02b2221\") " pod="metallb-system/controller-68d546b9d8-djznn" Oct 02 01:59:22 crc kubenswrapper[4885]: I1002 01:59:22.114679 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/47f27b52-1453-4566-954f-94e6b02b2221-metrics-certs\") pod \"controller-68d546b9d8-djznn\" (UID: \"47f27b52-1453-4566-954f-94e6b02b2221\") " pod="metallb-system/controller-68d546b9d8-djznn" Oct 02 01:59:22 crc kubenswrapper[4885]: I1002 01:59:22.114752 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/47f27b52-1453-4566-954f-94e6b02b2221-cert\") pod \"controller-68d546b9d8-djznn\" (UID: \"47f27b52-1453-4566-954f-94e6b02b2221\") " pod="metallb-system/controller-68d546b9d8-djznn" Oct 02 01:59:22 crc kubenswrapper[4885]: E1002 01:59:22.114892 4885 secret.go:188] Couldn't get secret metallb-system/controller-certs-secret: secret "controller-certs-secret" not found Oct 02 01:59:22 crc kubenswrapper[4885]: E1002 01:59:22.114982 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/47f27b52-1453-4566-954f-94e6b02b2221-metrics-certs podName:47f27b52-1453-4566-954f-94e6b02b2221 nodeName:}" failed. No retries permitted until 2025-10-02 01:59:22.614955005 +0000 UTC m=+751.426702404 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/47f27b52-1453-4566-954f-94e6b02b2221-metrics-certs") pod "controller-68d546b9d8-djznn" (UID: "47f27b52-1453-4566-954f-94e6b02b2221") : secret "controller-certs-secret" not found Oct 02 01:59:22 crc kubenswrapper[4885]: I1002 01:59:22.117588 4885 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Oct 02 01:59:22 crc kubenswrapper[4885]: I1002 01:59:22.129666 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/47f27b52-1453-4566-954f-94e6b02b2221-cert\") pod \"controller-68d546b9d8-djznn\" (UID: \"47f27b52-1453-4566-954f-94e6b02b2221\") " pod="metallb-system/controller-68d546b9d8-djznn" Oct 02 01:59:22 crc kubenswrapper[4885]: I1002 01:59:22.136054 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4nm24\" (UniqueName: \"kubernetes.io/projected/47f27b52-1453-4566-954f-94e6b02b2221-kube-api-access-4nm24\") pod \"controller-68d546b9d8-djznn\" (UID: \"47f27b52-1453-4566-954f-94e6b02b2221\") " pod="metallb-system/controller-68d546b9d8-djznn" Oct 02 01:59:22 crc kubenswrapper[4885]: I1002 01:59:22.158972 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-f76746f8f-znjc9" Oct 02 01:59:22 crc kubenswrapper[4885]: I1002 01:59:22.385910 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-f8bcd69bb-lzxjz" event={"ID":"fd6d9349-9419-433f-9e19-df003ca0912b","Type":"ContainerStarted","Data":"5b931bf71b4a90b8ca89dbff03adf417dbfc0230f075cf69ad0f3f26929b4143"} Oct 02 01:59:22 crc kubenswrapper[4885]: I1002 01:59:22.385953 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-f8bcd69bb-lzxjz" event={"ID":"fd6d9349-9419-433f-9e19-df003ca0912b","Type":"ContainerStarted","Data":"02f6bad593b5c16594cba86617ba254d8d87eca7d95bd057adfcd4beedefe750"} Oct 02 01:59:22 crc kubenswrapper[4885]: I1002 01:59:22.386163 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-f8bcd69bb-lzxjz" Oct 02 01:59:22 crc kubenswrapper[4885]: I1002 01:59:22.399353 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-f8bcd69bb-lzxjz" Oct 02 01:59:22 crc kubenswrapper[4885]: I1002 01:59:22.408239 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-f8bcd69bb-lzxjz" podStartSLOduration=3.408224366 podStartE2EDuration="3.408224366s" podCreationTimestamp="2025-10-02 01:59:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:59:22.404189213 +0000 UTC m=+751.215936612" watchObservedRunningTime="2025-10-02 01:59:22.408224366 +0000 UTC m=+751.219971765" Oct 02 01:59:22 crc kubenswrapper[4885]: I1002 01:59:22.417984 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/05715b27-682b-46a1-80b5-667667147ce0-metrics-certs\") pod \"frr-k8s-ggsh4\" (UID: \"05715b27-682b-46a1-80b5-667667147ce0\") " pod="metallb-system/frr-k8s-ggsh4" Oct 02 01:59:22 crc kubenswrapper[4885]: I1002 01:59:22.422794 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/05715b27-682b-46a1-80b5-667667147ce0-metrics-certs\") pod \"frr-k8s-ggsh4\" (UID: \"05715b27-682b-46a1-80b5-667667147ce0\") " pod="metallb-system/frr-k8s-ggsh4" Oct 02 01:59:22 crc kubenswrapper[4885]: I1002 01:59:22.519204 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/e3d6e1b0-5963-4e8f-bf3d-eb0e88ec0395-memberlist\") pod \"speaker-5h9mk\" (UID: \"e3d6e1b0-5963-4e8f-bf3d-eb0e88ec0395\") " pod="metallb-system/speaker-5h9mk" Oct 02 01:59:22 crc kubenswrapper[4885]: I1002 01:59:22.519296 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e3d6e1b0-5963-4e8f-bf3d-eb0e88ec0395-metrics-certs\") pod \"speaker-5h9mk\" (UID: \"e3d6e1b0-5963-4e8f-bf3d-eb0e88ec0395\") " pod="metallb-system/speaker-5h9mk" Oct 02 01:59:22 crc kubenswrapper[4885]: I1002 01:59:22.519318 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/66663f53-9790-48b4-8d16-4675661f340d-cert\") pod \"frr-k8s-webhook-server-64bf5d555-q7xtq\" (UID: \"66663f53-9790-48b4-8d16-4675661f340d\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-q7xtq" Oct 02 01:59:22 crc kubenswrapper[4885]: E1002 01:59:22.520309 4885 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 02 01:59:22 crc kubenswrapper[4885]: E1002 01:59:22.520489 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e3d6e1b0-5963-4e8f-bf3d-eb0e88ec0395-memberlist podName:e3d6e1b0-5963-4e8f-bf3d-eb0e88ec0395 nodeName:}" failed. No retries permitted until 2025-10-02 01:59:23.520457785 +0000 UTC m=+752.332205184 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/e3d6e1b0-5963-4e8f-bf3d-eb0e88ec0395-memberlist") pod "speaker-5h9mk" (UID: "e3d6e1b0-5963-4e8f-bf3d-eb0e88ec0395") : secret "metallb-memberlist" not found Oct 02 01:59:22 crc kubenswrapper[4885]: I1002 01:59:22.522528 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e3d6e1b0-5963-4e8f-bf3d-eb0e88ec0395-metrics-certs\") pod \"speaker-5h9mk\" (UID: \"e3d6e1b0-5963-4e8f-bf3d-eb0e88ec0395\") " pod="metallb-system/speaker-5h9mk" Oct 02 01:59:22 crc kubenswrapper[4885]: I1002 01:59:22.523080 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/66663f53-9790-48b4-8d16-4675661f340d-cert\") pod \"frr-k8s-webhook-server-64bf5d555-q7xtq\" (UID: \"66663f53-9790-48b4-8d16-4675661f340d\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-q7xtq" Oct 02 01:59:22 crc kubenswrapper[4885]: I1002 01:59:22.578131 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-f76746f8f-znjc9"] Oct 02 01:59:22 crc kubenswrapper[4885]: I1002 01:59:22.621174 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/47f27b52-1453-4566-954f-94e6b02b2221-metrics-certs\") pod \"controller-68d546b9d8-djznn\" (UID: \"47f27b52-1453-4566-954f-94e6b02b2221\") " pod="metallb-system/controller-68d546b9d8-djznn" Oct 02 01:59:22 crc kubenswrapper[4885]: I1002 01:59:22.637316 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/47f27b52-1453-4566-954f-94e6b02b2221-metrics-certs\") pod \"controller-68d546b9d8-djznn\" (UID: \"47f27b52-1453-4566-954f-94e6b02b2221\") " pod="metallb-system/controller-68d546b9d8-djznn" Oct 02 01:59:22 crc kubenswrapper[4885]: I1002 01:59:22.717728 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-ggsh4" Oct 02 01:59:22 crc kubenswrapper[4885]: I1002 01:59:22.722701 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-q7xtq" Oct 02 01:59:22 crc kubenswrapper[4885]: I1002 01:59:22.833230 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-djznn" Oct 02 01:59:23 crc kubenswrapper[4885]: I1002 01:59:23.264607 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-q7xtq"] Oct 02 01:59:23 crc kubenswrapper[4885]: W1002 01:59:23.276983 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod66663f53_9790_48b4_8d16_4675661f340d.slice/crio-016633f9962354274843b233cf5246a6b6fd39536d4390e85b232d2abd30b8fa WatchSource:0}: Error finding container 016633f9962354274843b233cf5246a6b6fd39536d4390e85b232d2abd30b8fa: Status 404 returned error can't find the container with id 016633f9962354274843b233cf5246a6b6fd39536d4390e85b232d2abd30b8fa Oct 02 01:59:23 crc kubenswrapper[4885]: I1002 01:59:23.329557 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-djznn"] Oct 02 01:59:23 crc kubenswrapper[4885]: W1002 01:59:23.337442 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod47f27b52_1453_4566_954f_94e6b02b2221.slice/crio-292782734fa2d141e5e39ff015d028ab83019a4f6d3a0334031c774b3447fdaf WatchSource:0}: Error finding container 292782734fa2d141e5e39ff015d028ab83019a4f6d3a0334031c774b3447fdaf: Status 404 returned error can't find the container with id 292782734fa2d141e5e39ff015d028ab83019a4f6d3a0334031c774b3447fdaf Oct 02 01:59:23 crc kubenswrapper[4885]: I1002 01:59:23.393222 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-djznn" event={"ID":"47f27b52-1453-4566-954f-94e6b02b2221","Type":"ContainerStarted","Data":"292782734fa2d141e5e39ff015d028ab83019a4f6d3a0334031c774b3447fdaf"} Oct 02 01:59:23 crc kubenswrapper[4885]: I1002 01:59:23.394228 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-ggsh4" event={"ID":"05715b27-682b-46a1-80b5-667667147ce0","Type":"ContainerStarted","Data":"e02cad2f2fcdace3ac449f5c837f1ce741e7c772bab46dccf85175e8b6d2ca2a"} Oct 02 01:59:23 crc kubenswrapper[4885]: I1002 01:59:23.395393 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-q7xtq" event={"ID":"66663f53-9790-48b4-8d16-4675661f340d","Type":"ContainerStarted","Data":"016633f9962354274843b233cf5246a6b6fd39536d4390e85b232d2abd30b8fa"} Oct 02 01:59:23 crc kubenswrapper[4885]: I1002 01:59:23.397047 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-f76746f8f-znjc9" event={"ID":"2678e135-47c6-4727-a31c-cbdfc16448bc","Type":"ContainerStarted","Data":"2d0f93f8377ded5413115f03f8660dd9142cd92a323db87e21bb3e4f8b6011e4"} Oct 02 01:59:23 crc kubenswrapper[4885]: I1002 01:59:23.397079 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-f76746f8f-znjc9" event={"ID":"2678e135-47c6-4727-a31c-cbdfc16448bc","Type":"ContainerStarted","Data":"de0b0639826811c89a0c59586e2ab53fab5cb73a2fda9332060cfbdbc3e8328c"} Oct 02 01:59:23 crc kubenswrapper[4885]: I1002 01:59:23.413119 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-f76746f8f-znjc9" podStartSLOduration=4.413097545 podStartE2EDuration="4.413097545s" podCreationTimestamp="2025-10-02 01:59:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:59:23.410091473 +0000 UTC m=+752.221838872" watchObservedRunningTime="2025-10-02 01:59:23.413097545 +0000 UTC m=+752.224844944" Oct 02 01:59:23 crc kubenswrapper[4885]: I1002 01:59:23.541544 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/e3d6e1b0-5963-4e8f-bf3d-eb0e88ec0395-memberlist\") pod \"speaker-5h9mk\" (UID: \"e3d6e1b0-5963-4e8f-bf3d-eb0e88ec0395\") " pod="metallb-system/speaker-5h9mk" Oct 02 01:59:23 crc kubenswrapper[4885]: I1002 01:59:23.547959 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/e3d6e1b0-5963-4e8f-bf3d-eb0e88ec0395-memberlist\") pod \"speaker-5h9mk\" (UID: \"e3d6e1b0-5963-4e8f-bf3d-eb0e88ec0395\") " pod="metallb-system/speaker-5h9mk" Oct 02 01:59:23 crc kubenswrapper[4885]: I1002 01:59:23.715104 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-5h9mk" Oct 02 01:59:23 crc kubenswrapper[4885]: W1002 01:59:23.732677 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode3d6e1b0_5963_4e8f_bf3d_eb0e88ec0395.slice/crio-21ff3a3a2817b6fa002d157caaf9cdd4d90d31b4e08d98b7eaae3314d4a1a34c WatchSource:0}: Error finding container 21ff3a3a2817b6fa002d157caaf9cdd4d90d31b4e08d98b7eaae3314d4a1a34c: Status 404 returned error can't find the container with id 21ff3a3a2817b6fa002d157caaf9cdd4d90d31b4e08d98b7eaae3314d4a1a34c Oct 02 01:59:24 crc kubenswrapper[4885]: I1002 01:59:24.403880 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-djznn" event={"ID":"47f27b52-1453-4566-954f-94e6b02b2221","Type":"ContainerStarted","Data":"a4406f66889640698a3c9b1691d8c065326ec326ca5260daa3c95edb32c74531"} Oct 02 01:59:24 crc kubenswrapper[4885]: I1002 01:59:24.404164 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-djznn" event={"ID":"47f27b52-1453-4566-954f-94e6b02b2221","Type":"ContainerStarted","Data":"e8974e59e04bd0cea00b9aaadca299e4755f396f40c94655c1e468afa5f71763"} Oct 02 01:59:24 crc kubenswrapper[4885]: I1002 01:59:24.404182 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-68d546b9d8-djznn" Oct 02 01:59:24 crc kubenswrapper[4885]: I1002 01:59:24.405931 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-5h9mk" event={"ID":"e3d6e1b0-5963-4e8f-bf3d-eb0e88ec0395","Type":"ContainerStarted","Data":"d137a50cb3de445e834d91d614f337e24239ee29a9cc07c4a128a07bc91813dc"} Oct 02 01:59:24 crc kubenswrapper[4885]: I1002 01:59:24.405967 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-5h9mk" event={"ID":"e3d6e1b0-5963-4e8f-bf3d-eb0e88ec0395","Type":"ContainerStarted","Data":"fd788afa6276a44341820da1ee929b075dc69ac11c6168ebbf752c9d8c8b4227"} Oct 02 01:59:24 crc kubenswrapper[4885]: I1002 01:59:24.405976 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-5h9mk" event={"ID":"e3d6e1b0-5963-4e8f-bf3d-eb0e88ec0395","Type":"ContainerStarted","Data":"21ff3a3a2817b6fa002d157caaf9cdd4d90d31b4e08d98b7eaae3314d4a1a34c"} Oct 02 01:59:24 crc kubenswrapper[4885]: I1002 01:59:24.406326 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-f76746f8f-znjc9" Oct 02 01:59:24 crc kubenswrapper[4885]: I1002 01:59:24.411370 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-f76746f8f-znjc9" Oct 02 01:59:24 crc kubenswrapper[4885]: I1002 01:59:24.420229 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-68d546b9d8-djznn" podStartSLOduration=3.420205841 podStartE2EDuration="3.420205841s" podCreationTimestamp="2025-10-02 01:59:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:59:24.418771478 +0000 UTC m=+753.230518877" watchObservedRunningTime="2025-10-02 01:59:24.420205841 +0000 UTC m=+753.231953240" Oct 02 01:59:24 crc kubenswrapper[4885]: I1002 01:59:24.438596 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-5h9mk" podStartSLOduration=3.438576499 podStartE2EDuration="3.438576499s" podCreationTimestamp="2025-10-02 01:59:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:59:24.43694094 +0000 UTC m=+753.248688339" watchObservedRunningTime="2025-10-02 01:59:24.438576499 +0000 UTC m=+753.250323898" Oct 02 01:59:27 crc kubenswrapper[4885]: I1002 01:59:27.138405 4885 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 02 01:59:32 crc kubenswrapper[4885]: I1002 01:59:32.475561 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-q7xtq" event={"ID":"66663f53-9790-48b4-8d16-4675661f340d","Type":"ContainerStarted","Data":"870f24958de39c5cf74c6723cb8b7bc05f92c88aed16f133df18d9dee447cb1e"} Oct 02 01:59:32 crc kubenswrapper[4885]: I1002 01:59:32.475999 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-q7xtq" Oct 02 01:59:32 crc kubenswrapper[4885]: I1002 01:59:32.478309 4885 generic.go:334] "Generic (PLEG): container finished" podID="05715b27-682b-46a1-80b5-667667147ce0" containerID="3596d6266ffb958c3ca1545ee5c8ef6301eff5a4f5649aebab03b42da9566e59" exitCode=0 Oct 02 01:59:32 crc kubenswrapper[4885]: I1002 01:59:32.478405 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-ggsh4" event={"ID":"05715b27-682b-46a1-80b5-667667147ce0","Type":"ContainerDied","Data":"3596d6266ffb958c3ca1545ee5c8ef6301eff5a4f5649aebab03b42da9566e59"} Oct 02 01:59:32 crc kubenswrapper[4885]: I1002 01:59:32.498828 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-q7xtq" podStartSLOduration=3.192925196 podStartE2EDuration="11.498812987s" podCreationTimestamp="2025-10-02 01:59:21 +0000 UTC" firstStartedPulling="2025-10-02 01:59:23.278557637 +0000 UTC m=+752.090305036" lastFinishedPulling="2025-10-02 01:59:31.584445418 +0000 UTC m=+760.396192827" observedRunningTime="2025-10-02 01:59:32.497661873 +0000 UTC m=+761.309409282" watchObservedRunningTime="2025-10-02 01:59:32.498812987 +0000 UTC m=+761.310560396" Oct 02 01:59:33 crc kubenswrapper[4885]: I1002 01:59:33.489118 4885 generic.go:334] "Generic (PLEG): container finished" podID="05715b27-682b-46a1-80b5-667667147ce0" containerID="1c7959544b8c80d82cbd0ce7b8da9ff71c0a3a852dd2f6b7db1fd268519d9768" exitCode=0 Oct 02 01:59:33 crc kubenswrapper[4885]: I1002 01:59:33.489180 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-ggsh4" event={"ID":"05715b27-682b-46a1-80b5-667667147ce0","Type":"ContainerDied","Data":"1c7959544b8c80d82cbd0ce7b8da9ff71c0a3a852dd2f6b7db1fd268519d9768"} Oct 02 01:59:33 crc kubenswrapper[4885]: I1002 01:59:33.715683 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-5h9mk" Oct 02 01:59:34 crc kubenswrapper[4885]: I1002 01:59:34.502970 4885 generic.go:334] "Generic (PLEG): container finished" podID="05715b27-682b-46a1-80b5-667667147ce0" containerID="c81a4f7751c0a9679882484853a763e2582fd4406cf67210149482ab65f86677" exitCode=0 Oct 02 01:59:34 crc kubenswrapper[4885]: I1002 01:59:34.503057 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-ggsh4" event={"ID":"05715b27-682b-46a1-80b5-667667147ce0","Type":"ContainerDied","Data":"c81a4f7751c0a9679882484853a763e2582fd4406cf67210149482ab65f86677"} Oct 02 01:59:35 crc kubenswrapper[4885]: I1002 01:59:35.513570 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-ggsh4" event={"ID":"05715b27-682b-46a1-80b5-667667147ce0","Type":"ContainerStarted","Data":"f1a81fad3ffa1b15c86e27ca3d481232ebcf580370f6adaa833bd75e102e9111"} Oct 02 01:59:35 crc kubenswrapper[4885]: I1002 01:59:35.514564 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-ggsh4" event={"ID":"05715b27-682b-46a1-80b5-667667147ce0","Type":"ContainerStarted","Data":"16a7fc680ff4bbfddd76048fe05a232b5ac4c9163cd2d1ca0cc5c6f41f2446c2"} Oct 02 01:59:35 crc kubenswrapper[4885]: I1002 01:59:35.514831 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-ggsh4" event={"ID":"05715b27-682b-46a1-80b5-667667147ce0","Type":"ContainerStarted","Data":"560b10869b4943a8191b46a44553d7f1bd84e51517acf44b7e45240f39c67bbe"} Oct 02 01:59:35 crc kubenswrapper[4885]: I1002 01:59:35.515231 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-ggsh4" event={"ID":"05715b27-682b-46a1-80b5-667667147ce0","Type":"ContainerStarted","Data":"6ef393c76588515031f864c6570e52439b21d1a1251da8fcf841a3b1482ddf1b"} Oct 02 01:59:36 crc kubenswrapper[4885]: I1002 01:59:36.529380 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-ggsh4" event={"ID":"05715b27-682b-46a1-80b5-667667147ce0","Type":"ContainerStarted","Data":"fa60024dcf52abdd87c01e04918c2ef3774bdaf029b204e278a55c359529a0d7"} Oct 02 01:59:37 crc kubenswrapper[4885]: I1002 01:59:37.058080 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-jfpnq"] Oct 02 01:59:37 crc kubenswrapper[4885]: I1002 01:59:37.059448 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jfpnq" Oct 02 01:59:37 crc kubenswrapper[4885]: I1002 01:59:37.083087 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jfpnq"] Oct 02 01:59:37 crc kubenswrapper[4885]: I1002 01:59:37.159358 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5b96649-fe77-44fd-aa6f-42d3cea5af1f-utilities\") pod \"redhat-marketplace-jfpnq\" (UID: \"e5b96649-fe77-44fd-aa6f-42d3cea5af1f\") " pod="openshift-marketplace/redhat-marketplace-jfpnq" Oct 02 01:59:37 crc kubenswrapper[4885]: I1002 01:59:37.159853 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7gv9f\" (UniqueName: \"kubernetes.io/projected/e5b96649-fe77-44fd-aa6f-42d3cea5af1f-kube-api-access-7gv9f\") pod \"redhat-marketplace-jfpnq\" (UID: \"e5b96649-fe77-44fd-aa6f-42d3cea5af1f\") " pod="openshift-marketplace/redhat-marketplace-jfpnq" Oct 02 01:59:37 crc kubenswrapper[4885]: I1002 01:59:37.160513 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5b96649-fe77-44fd-aa6f-42d3cea5af1f-catalog-content\") pod \"redhat-marketplace-jfpnq\" (UID: \"e5b96649-fe77-44fd-aa6f-42d3cea5af1f\") " pod="openshift-marketplace/redhat-marketplace-jfpnq" Oct 02 01:59:37 crc kubenswrapper[4885]: I1002 01:59:37.261760 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5b96649-fe77-44fd-aa6f-42d3cea5af1f-catalog-content\") pod \"redhat-marketplace-jfpnq\" (UID: \"e5b96649-fe77-44fd-aa6f-42d3cea5af1f\") " pod="openshift-marketplace/redhat-marketplace-jfpnq" Oct 02 01:59:37 crc kubenswrapper[4885]: I1002 01:59:37.261826 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5b96649-fe77-44fd-aa6f-42d3cea5af1f-utilities\") pod \"redhat-marketplace-jfpnq\" (UID: \"e5b96649-fe77-44fd-aa6f-42d3cea5af1f\") " pod="openshift-marketplace/redhat-marketplace-jfpnq" Oct 02 01:59:37 crc kubenswrapper[4885]: I1002 01:59:37.261877 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7gv9f\" (UniqueName: \"kubernetes.io/projected/e5b96649-fe77-44fd-aa6f-42d3cea5af1f-kube-api-access-7gv9f\") pod \"redhat-marketplace-jfpnq\" (UID: \"e5b96649-fe77-44fd-aa6f-42d3cea5af1f\") " pod="openshift-marketplace/redhat-marketplace-jfpnq" Oct 02 01:59:37 crc kubenswrapper[4885]: I1002 01:59:37.262466 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5b96649-fe77-44fd-aa6f-42d3cea5af1f-utilities\") pod \"redhat-marketplace-jfpnq\" (UID: \"e5b96649-fe77-44fd-aa6f-42d3cea5af1f\") " pod="openshift-marketplace/redhat-marketplace-jfpnq" Oct 02 01:59:37 crc kubenswrapper[4885]: I1002 01:59:37.263054 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5b96649-fe77-44fd-aa6f-42d3cea5af1f-catalog-content\") pod \"redhat-marketplace-jfpnq\" (UID: \"e5b96649-fe77-44fd-aa6f-42d3cea5af1f\") " pod="openshift-marketplace/redhat-marketplace-jfpnq" Oct 02 01:59:37 crc kubenswrapper[4885]: I1002 01:59:37.296342 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7gv9f\" (UniqueName: \"kubernetes.io/projected/e5b96649-fe77-44fd-aa6f-42d3cea5af1f-kube-api-access-7gv9f\") pod \"redhat-marketplace-jfpnq\" (UID: \"e5b96649-fe77-44fd-aa6f-42d3cea5af1f\") " pod="openshift-marketplace/redhat-marketplace-jfpnq" Oct 02 01:59:37 crc kubenswrapper[4885]: I1002 01:59:37.391930 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jfpnq" Oct 02 01:59:37 crc kubenswrapper[4885]: I1002 01:59:37.564993 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-ggsh4" event={"ID":"05715b27-682b-46a1-80b5-667667147ce0","Type":"ContainerStarted","Data":"082896d9ef88be47c2a76ecb1ce76a55964c9f761648799abae4c8bc8fe4e6cd"} Oct 02 01:59:37 crc kubenswrapper[4885]: I1002 01:59:37.565554 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-ggsh4" Oct 02 01:59:37 crc kubenswrapper[4885]: I1002 01:59:37.718349 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-ggsh4" Oct 02 01:59:37 crc kubenswrapper[4885]: I1002 01:59:37.767371 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-ggsh4" Oct 02 01:59:37 crc kubenswrapper[4885]: I1002 01:59:37.797070 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-ggsh4" podStartSLOduration=8.100993989 podStartE2EDuration="16.797052174s" podCreationTimestamp="2025-10-02 01:59:21 +0000 UTC" firstStartedPulling="2025-10-02 01:59:22.927157911 +0000 UTC m=+751.738905310" lastFinishedPulling="2025-10-02 01:59:31.623216096 +0000 UTC m=+760.434963495" observedRunningTime="2025-10-02 01:59:37.587829137 +0000 UTC m=+766.399576556" watchObservedRunningTime="2025-10-02 01:59:37.797052174 +0000 UTC m=+766.608799573" Oct 02 01:59:37 crc kubenswrapper[4885]: I1002 01:59:37.871283 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jfpnq"] Oct 02 01:59:38 crc kubenswrapper[4885]: I1002 01:59:38.575189 4885 generic.go:334] "Generic (PLEG): container finished" podID="e5b96649-fe77-44fd-aa6f-42d3cea5af1f" containerID="10f0ab335b054a366152391c4e4610134c705a453de2a73ddf77e3105d8e3b01" exitCode=0 Oct 02 01:59:38 crc kubenswrapper[4885]: I1002 01:59:38.575313 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jfpnq" event={"ID":"e5b96649-fe77-44fd-aa6f-42d3cea5af1f","Type":"ContainerDied","Data":"10f0ab335b054a366152391c4e4610134c705a453de2a73ddf77e3105d8e3b01"} Oct 02 01:59:38 crc kubenswrapper[4885]: I1002 01:59:38.575376 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jfpnq" event={"ID":"e5b96649-fe77-44fd-aa6f-42d3cea5af1f","Type":"ContainerStarted","Data":"96d0326c5fb7935e7ed0901e638c0a3b7b7ca6d629d643694b934ee9ee45b673"} Oct 02 01:59:39 crc kubenswrapper[4885]: I1002 01:59:39.593458 4885 generic.go:334] "Generic (PLEG): container finished" podID="e5b96649-fe77-44fd-aa6f-42d3cea5af1f" containerID="3bb0f0f7d548c8cfab60f6f626c5cc2103310e876a421c17ebf9062ddd08953c" exitCode=0 Oct 02 01:59:39 crc kubenswrapper[4885]: I1002 01:59:39.594230 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jfpnq" event={"ID":"e5b96649-fe77-44fd-aa6f-42d3cea5af1f","Type":"ContainerDied","Data":"3bb0f0f7d548c8cfab60f6f626c5cc2103310e876a421c17ebf9062ddd08953c"} Oct 02 01:59:40 crc kubenswrapper[4885]: I1002 01:59:40.605091 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jfpnq" event={"ID":"e5b96649-fe77-44fd-aa6f-42d3cea5af1f","Type":"ContainerStarted","Data":"204df8de01ee415f4f877e501451357423388f066b0e9618c234d0d409364e44"} Oct 02 01:59:40 crc kubenswrapper[4885]: I1002 01:59:40.637067 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-jfpnq" podStartSLOduration=2.208119363 podStartE2EDuration="3.637042305s" podCreationTimestamp="2025-10-02 01:59:37 +0000 UTC" firstStartedPulling="2025-10-02 01:59:38.578130773 +0000 UTC m=+767.389878192" lastFinishedPulling="2025-10-02 01:59:40.007053695 +0000 UTC m=+768.818801134" observedRunningTime="2025-10-02 01:59:40.633524828 +0000 UTC m=+769.445272257" watchObservedRunningTime="2025-10-02 01:59:40.637042305 +0000 UTC m=+769.448789744" Oct 02 01:59:42 crc kubenswrapper[4885]: I1002 01:59:42.733922 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-q7xtq" Oct 02 01:59:42 crc kubenswrapper[4885]: I1002 01:59:42.841947 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-68d546b9d8-djznn" Oct 02 01:59:43 crc kubenswrapper[4885]: I1002 01:59:43.265734 4885 patch_prober.go:28] interesting pod/machine-config-daemon-rttx8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 01:59:43 crc kubenswrapper[4885]: I1002 01:59:43.265850 4885 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 01:59:43 crc kubenswrapper[4885]: I1002 01:59:43.723531 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-5h9mk" Oct 02 01:59:46 crc kubenswrapper[4885]: I1002 01:59:46.624459 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-grlfd"] Oct 02 01:59:46 crc kubenswrapper[4885]: I1002 01:59:46.625974 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-grlfd" Oct 02 01:59:46 crc kubenswrapper[4885]: I1002 01:59:46.630353 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Oct 02 01:59:46 crc kubenswrapper[4885]: I1002 01:59:46.633989 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Oct 02 01:59:46 crc kubenswrapper[4885]: I1002 01:59:46.637084 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-grlfd"] Oct 02 01:59:46 crc kubenswrapper[4885]: I1002 01:59:46.798134 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rhwsh\" (UniqueName: \"kubernetes.io/projected/6446b34a-3b5d-4a49-a81b-6ef0e5613bea-kube-api-access-rhwsh\") pod \"openstack-operator-index-grlfd\" (UID: \"6446b34a-3b5d-4a49-a81b-6ef0e5613bea\") " pod="openstack-operators/openstack-operator-index-grlfd" Oct 02 01:59:46 crc kubenswrapper[4885]: I1002 01:59:46.899577 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rhwsh\" (UniqueName: \"kubernetes.io/projected/6446b34a-3b5d-4a49-a81b-6ef0e5613bea-kube-api-access-rhwsh\") pod \"openstack-operator-index-grlfd\" (UID: \"6446b34a-3b5d-4a49-a81b-6ef0e5613bea\") " pod="openstack-operators/openstack-operator-index-grlfd" Oct 02 01:59:46 crc kubenswrapper[4885]: I1002 01:59:46.931797 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rhwsh\" (UniqueName: \"kubernetes.io/projected/6446b34a-3b5d-4a49-a81b-6ef0e5613bea-kube-api-access-rhwsh\") pod \"openstack-operator-index-grlfd\" (UID: \"6446b34a-3b5d-4a49-a81b-6ef0e5613bea\") " pod="openstack-operators/openstack-operator-index-grlfd" Oct 02 01:59:46 crc kubenswrapper[4885]: I1002 01:59:46.947057 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-grlfd" Oct 02 01:59:47 crc kubenswrapper[4885]: I1002 01:59:47.392861 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-jfpnq" Oct 02 01:59:47 crc kubenswrapper[4885]: I1002 01:59:47.393671 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-jfpnq" Oct 02 01:59:47 crc kubenswrapper[4885]: I1002 01:59:47.421906 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-grlfd"] Oct 02 01:59:47 crc kubenswrapper[4885]: W1002 01:59:47.431010 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6446b34a_3b5d_4a49_a81b_6ef0e5613bea.slice/crio-0e4d1f350a39ef3d4608181d507082b19ac1410b5d33605c9b186377bcdf6c5d WatchSource:0}: Error finding container 0e4d1f350a39ef3d4608181d507082b19ac1410b5d33605c9b186377bcdf6c5d: Status 404 returned error can't find the container with id 0e4d1f350a39ef3d4608181d507082b19ac1410b5d33605c9b186377bcdf6c5d Oct 02 01:59:47 crc kubenswrapper[4885]: I1002 01:59:47.459786 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-jfpnq" Oct 02 01:59:47 crc kubenswrapper[4885]: I1002 01:59:47.665351 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-grlfd" event={"ID":"6446b34a-3b5d-4a49-a81b-6ef0e5613bea","Type":"ContainerStarted","Data":"0e4d1f350a39ef3d4608181d507082b19ac1410b5d33605c9b186377bcdf6c5d"} Oct 02 01:59:47 crc kubenswrapper[4885]: I1002 01:59:47.723013 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-jfpnq" Oct 02 01:59:50 crc kubenswrapper[4885]: I1002 01:59:50.795044 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-grlfd"] Oct 02 01:59:51 crc kubenswrapper[4885]: I1002 01:59:51.599706 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-zgthg"] Oct 02 01:59:51 crc kubenswrapper[4885]: I1002 01:59:51.600523 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-zgthg" Oct 02 01:59:51 crc kubenswrapper[4885]: I1002 01:59:51.604558 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-q44bl" Oct 02 01:59:51 crc kubenswrapper[4885]: I1002 01:59:51.618369 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-zgthg"] Oct 02 01:59:51 crc kubenswrapper[4885]: I1002 01:59:51.698998 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8sdmm\" (UniqueName: \"kubernetes.io/projected/12809968-8a60-4ffa-9506-0c89ce063382-kube-api-access-8sdmm\") pod \"openstack-operator-index-zgthg\" (UID: \"12809968-8a60-4ffa-9506-0c89ce063382\") " pod="openstack-operators/openstack-operator-index-zgthg" Oct 02 01:59:51 crc kubenswrapper[4885]: I1002 01:59:51.800423 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8sdmm\" (UniqueName: \"kubernetes.io/projected/12809968-8a60-4ffa-9506-0c89ce063382-kube-api-access-8sdmm\") pod \"openstack-operator-index-zgthg\" (UID: \"12809968-8a60-4ffa-9506-0c89ce063382\") " pod="openstack-operators/openstack-operator-index-zgthg" Oct 02 01:59:51 crc kubenswrapper[4885]: I1002 01:59:51.840648 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8sdmm\" (UniqueName: \"kubernetes.io/projected/12809968-8a60-4ffa-9506-0c89ce063382-kube-api-access-8sdmm\") pod \"openstack-operator-index-zgthg\" (UID: \"12809968-8a60-4ffa-9506-0c89ce063382\") " pod="openstack-operators/openstack-operator-index-zgthg" Oct 02 01:59:51 crc kubenswrapper[4885]: I1002 01:59:51.937707 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-zgthg" Oct 02 01:59:52 crc kubenswrapper[4885]: I1002 01:59:52.195302 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jfpnq"] Oct 02 01:59:52 crc kubenswrapper[4885]: I1002 01:59:52.195666 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-jfpnq" podUID="e5b96649-fe77-44fd-aa6f-42d3cea5af1f" containerName="registry-server" containerID="cri-o://204df8de01ee415f4f877e501451357423388f066b0e9618c234d0d409364e44" gracePeriod=2 Oct 02 01:59:52 crc kubenswrapper[4885]: I1002 01:59:52.708805 4885 generic.go:334] "Generic (PLEG): container finished" podID="e5b96649-fe77-44fd-aa6f-42d3cea5af1f" containerID="204df8de01ee415f4f877e501451357423388f066b0e9618c234d0d409364e44" exitCode=0 Oct 02 01:59:52 crc kubenswrapper[4885]: I1002 01:59:52.708878 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jfpnq" event={"ID":"e5b96649-fe77-44fd-aa6f-42d3cea5af1f","Type":"ContainerDied","Data":"204df8de01ee415f4f877e501451357423388f066b0e9618c234d0d409364e44"} Oct 02 01:59:52 crc kubenswrapper[4885]: I1002 01:59:52.721089 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-ggsh4" Oct 02 01:59:53 crc kubenswrapper[4885]: I1002 01:59:53.078186 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jfpnq" Oct 02 01:59:53 crc kubenswrapper[4885]: I1002 01:59:53.118430 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5b96649-fe77-44fd-aa6f-42d3cea5af1f-catalog-content\") pod \"e5b96649-fe77-44fd-aa6f-42d3cea5af1f\" (UID: \"e5b96649-fe77-44fd-aa6f-42d3cea5af1f\") " Oct 02 01:59:53 crc kubenswrapper[4885]: I1002 01:59:53.118491 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7gv9f\" (UniqueName: \"kubernetes.io/projected/e5b96649-fe77-44fd-aa6f-42d3cea5af1f-kube-api-access-7gv9f\") pod \"e5b96649-fe77-44fd-aa6f-42d3cea5af1f\" (UID: \"e5b96649-fe77-44fd-aa6f-42d3cea5af1f\") " Oct 02 01:59:53 crc kubenswrapper[4885]: I1002 01:59:53.118550 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5b96649-fe77-44fd-aa6f-42d3cea5af1f-utilities\") pod \"e5b96649-fe77-44fd-aa6f-42d3cea5af1f\" (UID: \"e5b96649-fe77-44fd-aa6f-42d3cea5af1f\") " Oct 02 01:59:53 crc kubenswrapper[4885]: I1002 01:59:53.119802 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e5b96649-fe77-44fd-aa6f-42d3cea5af1f-utilities" (OuterVolumeSpecName: "utilities") pod "e5b96649-fe77-44fd-aa6f-42d3cea5af1f" (UID: "e5b96649-fe77-44fd-aa6f-42d3cea5af1f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 01:59:53 crc kubenswrapper[4885]: I1002 01:59:53.134738 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e5b96649-fe77-44fd-aa6f-42d3cea5af1f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e5b96649-fe77-44fd-aa6f-42d3cea5af1f" (UID: "e5b96649-fe77-44fd-aa6f-42d3cea5af1f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 01:59:53 crc kubenswrapper[4885]: I1002 01:59:53.160002 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5b96649-fe77-44fd-aa6f-42d3cea5af1f-kube-api-access-7gv9f" (OuterVolumeSpecName: "kube-api-access-7gv9f") pod "e5b96649-fe77-44fd-aa6f-42d3cea5af1f" (UID: "e5b96649-fe77-44fd-aa6f-42d3cea5af1f"). InnerVolumeSpecName "kube-api-access-7gv9f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:59:53 crc kubenswrapper[4885]: I1002 01:59:53.219651 4885 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5b96649-fe77-44fd-aa6f-42d3cea5af1f-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 01:59:53 crc kubenswrapper[4885]: I1002 01:59:53.219876 4885 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5b96649-fe77-44fd-aa6f-42d3cea5af1f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 01:59:53 crc kubenswrapper[4885]: I1002 01:59:53.219885 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7gv9f\" (UniqueName: \"kubernetes.io/projected/e5b96649-fe77-44fd-aa6f-42d3cea5af1f-kube-api-access-7gv9f\") on node \"crc\" DevicePath \"\"" Oct 02 01:59:53 crc kubenswrapper[4885]: I1002 01:59:53.726324 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jfpnq" event={"ID":"e5b96649-fe77-44fd-aa6f-42d3cea5af1f","Type":"ContainerDied","Data":"96d0326c5fb7935e7ed0901e638c0a3b7b7ca6d629d643694b934ee9ee45b673"} Oct 02 01:59:53 crc kubenswrapper[4885]: I1002 01:59:53.726418 4885 scope.go:117] "RemoveContainer" containerID="204df8de01ee415f4f877e501451357423388f066b0e9618c234d0d409364e44" Oct 02 01:59:53 crc kubenswrapper[4885]: I1002 01:59:53.726449 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jfpnq" Oct 02 01:59:53 crc kubenswrapper[4885]: I1002 01:59:53.814308 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jfpnq"] Oct 02 01:59:53 crc kubenswrapper[4885]: I1002 01:59:53.818101 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-jfpnq"] Oct 02 01:59:53 crc kubenswrapper[4885]: I1002 01:59:53.934602 4885 scope.go:117] "RemoveContainer" containerID="3bb0f0f7d548c8cfab60f6f626c5cc2103310e876a421c17ebf9062ddd08953c" Oct 02 01:59:53 crc kubenswrapper[4885]: I1002 01:59:53.955998 4885 scope.go:117] "RemoveContainer" containerID="10f0ab335b054a366152391c4e4610134c705a453de2a73ddf77e3105d8e3b01" Oct 02 01:59:54 crc kubenswrapper[4885]: I1002 01:59:54.059403 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e5b96649-fe77-44fd-aa6f-42d3cea5af1f" path="/var/lib/kubelet/pods/e5b96649-fe77-44fd-aa6f-42d3cea5af1f/volumes" Oct 02 01:59:54 crc kubenswrapper[4885]: I1002 01:59:54.203886 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-zgthg"] Oct 02 01:59:54 crc kubenswrapper[4885]: I1002 01:59:54.738283 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-zgthg" event={"ID":"12809968-8a60-4ffa-9506-0c89ce063382","Type":"ContainerStarted","Data":"0fbdd74f16d658708a62c29beb2fe6c7a8f792d2262a76b396906088f2798ee4"} Oct 02 01:59:54 crc kubenswrapper[4885]: I1002 01:59:54.740570 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-grlfd" event={"ID":"6446b34a-3b5d-4a49-a81b-6ef0e5613bea","Type":"ContainerStarted","Data":"a04ff8c3588e8c2fdd178148ec914abc771c2bb413600e624709f6849aeb939c"} Oct 02 01:59:54 crc kubenswrapper[4885]: I1002 01:59:54.741173 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-grlfd" podUID="6446b34a-3b5d-4a49-a81b-6ef0e5613bea" containerName="registry-server" containerID="cri-o://a04ff8c3588e8c2fdd178148ec914abc771c2bb413600e624709f6849aeb939c" gracePeriod=2 Oct 02 01:59:54 crc kubenswrapper[4885]: I1002 01:59:54.768383 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-grlfd" podStartSLOduration=2.02825901 podStartE2EDuration="8.76835266s" podCreationTimestamp="2025-10-02 01:59:46 +0000 UTC" firstStartedPulling="2025-10-02 01:59:47.433385376 +0000 UTC m=+776.245132825" lastFinishedPulling="2025-10-02 01:59:54.173479046 +0000 UTC m=+782.985226475" observedRunningTime="2025-10-02 01:59:54.766123932 +0000 UTC m=+783.577871361" watchObservedRunningTime="2025-10-02 01:59:54.76835266 +0000 UTC m=+783.580100099" Oct 02 01:59:55 crc kubenswrapper[4885]: I1002 01:59:55.336168 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-grlfd" Oct 02 01:59:55 crc kubenswrapper[4885]: I1002 01:59:55.457583 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rhwsh\" (UniqueName: \"kubernetes.io/projected/6446b34a-3b5d-4a49-a81b-6ef0e5613bea-kube-api-access-rhwsh\") pod \"6446b34a-3b5d-4a49-a81b-6ef0e5613bea\" (UID: \"6446b34a-3b5d-4a49-a81b-6ef0e5613bea\") " Oct 02 01:59:55 crc kubenswrapper[4885]: I1002 01:59:55.465781 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6446b34a-3b5d-4a49-a81b-6ef0e5613bea-kube-api-access-rhwsh" (OuterVolumeSpecName: "kube-api-access-rhwsh") pod "6446b34a-3b5d-4a49-a81b-6ef0e5613bea" (UID: "6446b34a-3b5d-4a49-a81b-6ef0e5613bea"). InnerVolumeSpecName "kube-api-access-rhwsh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:59:55 crc kubenswrapper[4885]: I1002 01:59:55.559816 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rhwsh\" (UniqueName: \"kubernetes.io/projected/6446b34a-3b5d-4a49-a81b-6ef0e5613bea-kube-api-access-rhwsh\") on node \"crc\" DevicePath \"\"" Oct 02 01:59:55 crc kubenswrapper[4885]: I1002 01:59:55.750813 4885 generic.go:334] "Generic (PLEG): container finished" podID="6446b34a-3b5d-4a49-a81b-6ef0e5613bea" containerID="a04ff8c3588e8c2fdd178148ec914abc771c2bb413600e624709f6849aeb939c" exitCode=0 Oct 02 01:59:55 crc kubenswrapper[4885]: I1002 01:59:55.751067 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-grlfd" Oct 02 01:59:55 crc kubenswrapper[4885]: I1002 01:59:55.751305 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-grlfd" event={"ID":"6446b34a-3b5d-4a49-a81b-6ef0e5613bea","Type":"ContainerDied","Data":"a04ff8c3588e8c2fdd178148ec914abc771c2bb413600e624709f6849aeb939c"} Oct 02 01:59:55 crc kubenswrapper[4885]: I1002 01:59:55.751526 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-grlfd" event={"ID":"6446b34a-3b5d-4a49-a81b-6ef0e5613bea","Type":"ContainerDied","Data":"0e4d1f350a39ef3d4608181d507082b19ac1410b5d33605c9b186377bcdf6c5d"} Oct 02 01:59:55 crc kubenswrapper[4885]: I1002 01:59:55.751578 4885 scope.go:117] "RemoveContainer" containerID="a04ff8c3588e8c2fdd178148ec914abc771c2bb413600e624709f6849aeb939c" Oct 02 01:59:55 crc kubenswrapper[4885]: I1002 01:59:55.753492 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-zgthg" event={"ID":"12809968-8a60-4ffa-9506-0c89ce063382","Type":"ContainerStarted","Data":"70cab18faf89e3e1891102deea7ed5fad18dc21975ed79209316e5ca738c92de"} Oct 02 01:59:55 crc kubenswrapper[4885]: I1002 01:59:55.777945 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-zgthg" podStartSLOduration=4.460655492 podStartE2EDuration="4.777919751s" podCreationTimestamp="2025-10-02 01:59:51 +0000 UTC" firstStartedPulling="2025-10-02 01:59:54.231966334 +0000 UTC m=+783.043713743" lastFinishedPulling="2025-10-02 01:59:54.549230563 +0000 UTC m=+783.360978002" observedRunningTime="2025-10-02 01:59:55.775742395 +0000 UTC m=+784.587489824" watchObservedRunningTime="2025-10-02 01:59:55.777919751 +0000 UTC m=+784.589667190" Oct 02 01:59:55 crc kubenswrapper[4885]: I1002 01:59:55.779589 4885 scope.go:117] "RemoveContainer" containerID="a04ff8c3588e8c2fdd178148ec914abc771c2bb413600e624709f6849aeb939c" Oct 02 01:59:55 crc kubenswrapper[4885]: E1002 01:59:55.780422 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a04ff8c3588e8c2fdd178148ec914abc771c2bb413600e624709f6849aeb939c\": container with ID starting with a04ff8c3588e8c2fdd178148ec914abc771c2bb413600e624709f6849aeb939c not found: ID does not exist" containerID="a04ff8c3588e8c2fdd178148ec914abc771c2bb413600e624709f6849aeb939c" Oct 02 01:59:55 crc kubenswrapper[4885]: I1002 01:59:55.780656 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a04ff8c3588e8c2fdd178148ec914abc771c2bb413600e624709f6849aeb939c"} err="failed to get container status \"a04ff8c3588e8c2fdd178148ec914abc771c2bb413600e624709f6849aeb939c\": rpc error: code = NotFound desc = could not find container \"a04ff8c3588e8c2fdd178148ec914abc771c2bb413600e624709f6849aeb939c\": container with ID starting with a04ff8c3588e8c2fdd178148ec914abc771c2bb413600e624709f6849aeb939c not found: ID does not exist" Oct 02 01:59:55 crc kubenswrapper[4885]: I1002 01:59:55.800794 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-grlfd"] Oct 02 01:59:55 crc kubenswrapper[4885]: I1002 01:59:55.803121 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-grlfd"] Oct 02 01:59:56 crc kubenswrapper[4885]: I1002 01:59:56.056812 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6446b34a-3b5d-4a49-a81b-6ef0e5613bea" path="/var/lib/kubelet/pods/6446b34a-3b5d-4a49-a81b-6ef0e5613bea/volumes" Oct 02 02:00:00 crc kubenswrapper[4885]: I1002 02:00:00.140290 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322840-56lsm"] Oct 02 02:00:00 crc kubenswrapper[4885]: E1002 02:00:00.140927 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6446b34a-3b5d-4a49-a81b-6ef0e5613bea" containerName="registry-server" Oct 02 02:00:00 crc kubenswrapper[4885]: I1002 02:00:00.140949 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="6446b34a-3b5d-4a49-a81b-6ef0e5613bea" containerName="registry-server" Oct 02 02:00:00 crc kubenswrapper[4885]: E1002 02:00:00.140974 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5b96649-fe77-44fd-aa6f-42d3cea5af1f" containerName="extract-content" Oct 02 02:00:00 crc kubenswrapper[4885]: I1002 02:00:00.140986 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5b96649-fe77-44fd-aa6f-42d3cea5af1f" containerName="extract-content" Oct 02 02:00:00 crc kubenswrapper[4885]: E1002 02:00:00.141005 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5b96649-fe77-44fd-aa6f-42d3cea5af1f" containerName="extract-utilities" Oct 02 02:00:00 crc kubenswrapper[4885]: I1002 02:00:00.141020 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5b96649-fe77-44fd-aa6f-42d3cea5af1f" containerName="extract-utilities" Oct 02 02:00:00 crc kubenswrapper[4885]: E1002 02:00:00.141035 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5b96649-fe77-44fd-aa6f-42d3cea5af1f" containerName="registry-server" Oct 02 02:00:00 crc kubenswrapper[4885]: I1002 02:00:00.141047 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5b96649-fe77-44fd-aa6f-42d3cea5af1f" containerName="registry-server" Oct 02 02:00:00 crc kubenswrapper[4885]: I1002 02:00:00.141242 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="6446b34a-3b5d-4a49-a81b-6ef0e5613bea" containerName="registry-server" Oct 02 02:00:00 crc kubenswrapper[4885]: I1002 02:00:00.141293 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5b96649-fe77-44fd-aa6f-42d3cea5af1f" containerName="registry-server" Oct 02 02:00:00 crc kubenswrapper[4885]: I1002 02:00:00.141900 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322840-56lsm" Oct 02 02:00:00 crc kubenswrapper[4885]: I1002 02:00:00.144645 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 02 02:00:00 crc kubenswrapper[4885]: I1002 02:00:00.146539 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 02 02:00:00 crc kubenswrapper[4885]: I1002 02:00:00.158461 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322840-56lsm"] Oct 02 02:00:00 crc kubenswrapper[4885]: I1002 02:00:00.236228 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-srzf5\" (UniqueName: \"kubernetes.io/projected/e32595a5-56a6-4fb8-8f7d-41363c7f57d2-kube-api-access-srzf5\") pod \"collect-profiles-29322840-56lsm\" (UID: \"e32595a5-56a6-4fb8-8f7d-41363c7f57d2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322840-56lsm" Oct 02 02:00:00 crc kubenswrapper[4885]: I1002 02:00:00.236331 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e32595a5-56a6-4fb8-8f7d-41363c7f57d2-secret-volume\") pod \"collect-profiles-29322840-56lsm\" (UID: \"e32595a5-56a6-4fb8-8f7d-41363c7f57d2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322840-56lsm" Oct 02 02:00:00 crc kubenswrapper[4885]: I1002 02:00:00.236369 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e32595a5-56a6-4fb8-8f7d-41363c7f57d2-config-volume\") pod \"collect-profiles-29322840-56lsm\" (UID: \"e32595a5-56a6-4fb8-8f7d-41363c7f57d2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322840-56lsm" Oct 02 02:00:00 crc kubenswrapper[4885]: I1002 02:00:00.338839 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e32595a5-56a6-4fb8-8f7d-41363c7f57d2-secret-volume\") pod \"collect-profiles-29322840-56lsm\" (UID: \"e32595a5-56a6-4fb8-8f7d-41363c7f57d2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322840-56lsm" Oct 02 02:00:00 crc kubenswrapper[4885]: I1002 02:00:00.338942 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e32595a5-56a6-4fb8-8f7d-41363c7f57d2-config-volume\") pod \"collect-profiles-29322840-56lsm\" (UID: \"e32595a5-56a6-4fb8-8f7d-41363c7f57d2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322840-56lsm" Oct 02 02:00:00 crc kubenswrapper[4885]: I1002 02:00:00.339086 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-srzf5\" (UniqueName: \"kubernetes.io/projected/e32595a5-56a6-4fb8-8f7d-41363c7f57d2-kube-api-access-srzf5\") pod \"collect-profiles-29322840-56lsm\" (UID: \"e32595a5-56a6-4fb8-8f7d-41363c7f57d2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322840-56lsm" Oct 02 02:00:00 crc kubenswrapper[4885]: I1002 02:00:00.340901 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e32595a5-56a6-4fb8-8f7d-41363c7f57d2-config-volume\") pod \"collect-profiles-29322840-56lsm\" (UID: \"e32595a5-56a6-4fb8-8f7d-41363c7f57d2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322840-56lsm" Oct 02 02:00:00 crc kubenswrapper[4885]: I1002 02:00:00.348715 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e32595a5-56a6-4fb8-8f7d-41363c7f57d2-secret-volume\") pod \"collect-profiles-29322840-56lsm\" (UID: \"e32595a5-56a6-4fb8-8f7d-41363c7f57d2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322840-56lsm" Oct 02 02:00:00 crc kubenswrapper[4885]: I1002 02:00:00.368236 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-srzf5\" (UniqueName: \"kubernetes.io/projected/e32595a5-56a6-4fb8-8f7d-41363c7f57d2-kube-api-access-srzf5\") pod \"collect-profiles-29322840-56lsm\" (UID: \"e32595a5-56a6-4fb8-8f7d-41363c7f57d2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322840-56lsm" Oct 02 02:00:00 crc kubenswrapper[4885]: I1002 02:00:00.471362 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322840-56lsm" Oct 02 02:00:00 crc kubenswrapper[4885]: I1002 02:00:00.967211 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322840-56lsm"] Oct 02 02:00:00 crc kubenswrapper[4885]: W1002 02:00:00.985096 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode32595a5_56a6_4fb8_8f7d_41363c7f57d2.slice/crio-80d792037d06c627a6ef445405b7df309c4f662493faa1e09272a6222330e57f WatchSource:0}: Error finding container 80d792037d06c627a6ef445405b7df309c4f662493faa1e09272a6222330e57f: Status 404 returned error can't find the container with id 80d792037d06c627a6ef445405b7df309c4f662493faa1e09272a6222330e57f Oct 02 02:00:01 crc kubenswrapper[4885]: I1002 02:00:01.807035 4885 generic.go:334] "Generic (PLEG): container finished" podID="e32595a5-56a6-4fb8-8f7d-41363c7f57d2" containerID="ecd566acdf250db83dd67b6c57afa0780263d338c8eee6e4a4cd7c37be95e031" exitCode=0 Oct 02 02:00:01 crc kubenswrapper[4885]: I1002 02:00:01.807110 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29322840-56lsm" event={"ID":"e32595a5-56a6-4fb8-8f7d-41363c7f57d2","Type":"ContainerDied","Data":"ecd566acdf250db83dd67b6c57afa0780263d338c8eee6e4a4cd7c37be95e031"} Oct 02 02:00:01 crc kubenswrapper[4885]: I1002 02:00:01.807562 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29322840-56lsm" event={"ID":"e32595a5-56a6-4fb8-8f7d-41363c7f57d2","Type":"ContainerStarted","Data":"80d792037d06c627a6ef445405b7df309c4f662493faa1e09272a6222330e57f"} Oct 02 02:00:01 crc kubenswrapper[4885]: I1002 02:00:01.937886 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-zgthg" Oct 02 02:00:01 crc kubenswrapper[4885]: I1002 02:00:01.937957 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-zgthg" Oct 02 02:00:01 crc kubenswrapper[4885]: I1002 02:00:01.979633 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-zgthg" Oct 02 02:00:02 crc kubenswrapper[4885]: I1002 02:00:02.859299 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-zgthg" Oct 02 02:00:03 crc kubenswrapper[4885]: I1002 02:00:03.172316 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322840-56lsm" Oct 02 02:00:03 crc kubenswrapper[4885]: I1002 02:00:03.287997 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e32595a5-56a6-4fb8-8f7d-41363c7f57d2-config-volume\") pod \"e32595a5-56a6-4fb8-8f7d-41363c7f57d2\" (UID: \"e32595a5-56a6-4fb8-8f7d-41363c7f57d2\") " Oct 02 02:00:03 crc kubenswrapper[4885]: I1002 02:00:03.288158 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-srzf5\" (UniqueName: \"kubernetes.io/projected/e32595a5-56a6-4fb8-8f7d-41363c7f57d2-kube-api-access-srzf5\") pod \"e32595a5-56a6-4fb8-8f7d-41363c7f57d2\" (UID: \"e32595a5-56a6-4fb8-8f7d-41363c7f57d2\") " Oct 02 02:00:03 crc kubenswrapper[4885]: I1002 02:00:03.288291 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e32595a5-56a6-4fb8-8f7d-41363c7f57d2-secret-volume\") pod \"e32595a5-56a6-4fb8-8f7d-41363c7f57d2\" (UID: \"e32595a5-56a6-4fb8-8f7d-41363c7f57d2\") " Oct 02 02:00:03 crc kubenswrapper[4885]: I1002 02:00:03.289170 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e32595a5-56a6-4fb8-8f7d-41363c7f57d2-config-volume" (OuterVolumeSpecName: "config-volume") pod "e32595a5-56a6-4fb8-8f7d-41363c7f57d2" (UID: "e32595a5-56a6-4fb8-8f7d-41363c7f57d2"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:00:03 crc kubenswrapper[4885]: I1002 02:00:03.293615 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e32595a5-56a6-4fb8-8f7d-41363c7f57d2-kube-api-access-srzf5" (OuterVolumeSpecName: "kube-api-access-srzf5") pod "e32595a5-56a6-4fb8-8f7d-41363c7f57d2" (UID: "e32595a5-56a6-4fb8-8f7d-41363c7f57d2"). InnerVolumeSpecName "kube-api-access-srzf5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:00:03 crc kubenswrapper[4885]: I1002 02:00:03.293942 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e32595a5-56a6-4fb8-8f7d-41363c7f57d2-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "e32595a5-56a6-4fb8-8f7d-41363c7f57d2" (UID: "e32595a5-56a6-4fb8-8f7d-41363c7f57d2"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:00:03 crc kubenswrapper[4885]: I1002 02:00:03.389566 4885 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e32595a5-56a6-4fb8-8f7d-41363c7f57d2-config-volume\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:03 crc kubenswrapper[4885]: I1002 02:00:03.389606 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-srzf5\" (UniqueName: \"kubernetes.io/projected/e32595a5-56a6-4fb8-8f7d-41363c7f57d2-kube-api-access-srzf5\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:03 crc kubenswrapper[4885]: I1002 02:00:03.389621 4885 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e32595a5-56a6-4fb8-8f7d-41363c7f57d2-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:03 crc kubenswrapper[4885]: I1002 02:00:03.838164 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322840-56lsm" Oct 02 02:00:03 crc kubenswrapper[4885]: I1002 02:00:03.838434 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29322840-56lsm" event={"ID":"e32595a5-56a6-4fb8-8f7d-41363c7f57d2","Type":"ContainerDied","Data":"80d792037d06c627a6ef445405b7df309c4f662493faa1e09272a6222330e57f"} Oct 02 02:00:03 crc kubenswrapper[4885]: I1002 02:00:03.838777 4885 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="80d792037d06c627a6ef445405b7df309c4f662493faa1e09272a6222330e57f" Oct 02 02:00:06 crc kubenswrapper[4885]: I1002 02:00:06.007241 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-l4ktn"] Oct 02 02:00:06 crc kubenswrapper[4885]: E1002 02:00:06.007644 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e32595a5-56a6-4fb8-8f7d-41363c7f57d2" containerName="collect-profiles" Oct 02 02:00:06 crc kubenswrapper[4885]: I1002 02:00:06.007664 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="e32595a5-56a6-4fb8-8f7d-41363c7f57d2" containerName="collect-profiles" Oct 02 02:00:06 crc kubenswrapper[4885]: I1002 02:00:06.007902 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="e32595a5-56a6-4fb8-8f7d-41363c7f57d2" containerName="collect-profiles" Oct 02 02:00:06 crc kubenswrapper[4885]: I1002 02:00:06.009237 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l4ktn" Oct 02 02:00:06 crc kubenswrapper[4885]: I1002 02:00:06.020464 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-l4ktn"] Oct 02 02:00:06 crc kubenswrapper[4885]: I1002 02:00:06.168192 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pmc4n\" (UniqueName: \"kubernetes.io/projected/c5b45f09-e4c2-465b-9ba1-6d9375e5b246-kube-api-access-pmc4n\") pod \"redhat-operators-l4ktn\" (UID: \"c5b45f09-e4c2-465b-9ba1-6d9375e5b246\") " pod="openshift-marketplace/redhat-operators-l4ktn" Oct 02 02:00:06 crc kubenswrapper[4885]: I1002 02:00:06.168303 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c5b45f09-e4c2-465b-9ba1-6d9375e5b246-utilities\") pod \"redhat-operators-l4ktn\" (UID: \"c5b45f09-e4c2-465b-9ba1-6d9375e5b246\") " pod="openshift-marketplace/redhat-operators-l4ktn" Oct 02 02:00:06 crc kubenswrapper[4885]: I1002 02:00:06.168344 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c5b45f09-e4c2-465b-9ba1-6d9375e5b246-catalog-content\") pod \"redhat-operators-l4ktn\" (UID: \"c5b45f09-e4c2-465b-9ba1-6d9375e5b246\") " pod="openshift-marketplace/redhat-operators-l4ktn" Oct 02 02:00:06 crc kubenswrapper[4885]: I1002 02:00:06.269894 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c5b45f09-e4c2-465b-9ba1-6d9375e5b246-catalog-content\") pod \"redhat-operators-l4ktn\" (UID: \"c5b45f09-e4c2-465b-9ba1-6d9375e5b246\") " pod="openshift-marketplace/redhat-operators-l4ktn" Oct 02 02:00:06 crc kubenswrapper[4885]: I1002 02:00:06.270108 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pmc4n\" (UniqueName: \"kubernetes.io/projected/c5b45f09-e4c2-465b-9ba1-6d9375e5b246-kube-api-access-pmc4n\") pod \"redhat-operators-l4ktn\" (UID: \"c5b45f09-e4c2-465b-9ba1-6d9375e5b246\") " pod="openshift-marketplace/redhat-operators-l4ktn" Oct 02 02:00:06 crc kubenswrapper[4885]: I1002 02:00:06.270202 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c5b45f09-e4c2-465b-9ba1-6d9375e5b246-utilities\") pod \"redhat-operators-l4ktn\" (UID: \"c5b45f09-e4c2-465b-9ba1-6d9375e5b246\") " pod="openshift-marketplace/redhat-operators-l4ktn" Oct 02 02:00:06 crc kubenswrapper[4885]: I1002 02:00:06.271114 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c5b45f09-e4c2-465b-9ba1-6d9375e5b246-utilities\") pod \"redhat-operators-l4ktn\" (UID: \"c5b45f09-e4c2-465b-9ba1-6d9375e5b246\") " pod="openshift-marketplace/redhat-operators-l4ktn" Oct 02 02:00:06 crc kubenswrapper[4885]: I1002 02:00:06.271481 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c5b45f09-e4c2-465b-9ba1-6d9375e5b246-catalog-content\") pod \"redhat-operators-l4ktn\" (UID: \"c5b45f09-e4c2-465b-9ba1-6d9375e5b246\") " pod="openshift-marketplace/redhat-operators-l4ktn" Oct 02 02:00:06 crc kubenswrapper[4885]: I1002 02:00:06.307151 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pmc4n\" (UniqueName: \"kubernetes.io/projected/c5b45f09-e4c2-465b-9ba1-6d9375e5b246-kube-api-access-pmc4n\") pod \"redhat-operators-l4ktn\" (UID: \"c5b45f09-e4c2-465b-9ba1-6d9375e5b246\") " pod="openshift-marketplace/redhat-operators-l4ktn" Oct 02 02:00:06 crc kubenswrapper[4885]: I1002 02:00:06.339192 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l4ktn" Oct 02 02:00:06 crc kubenswrapper[4885]: I1002 02:00:06.885686 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-l4ktn"] Oct 02 02:00:07 crc kubenswrapper[4885]: I1002 02:00:07.871313 4885 generic.go:334] "Generic (PLEG): container finished" podID="c5b45f09-e4c2-465b-9ba1-6d9375e5b246" containerID="9b1bbf0001f658947690f50f678f07febae92b609dfa7e44fe9fc13fa6d0a87d" exitCode=0 Oct 02 02:00:07 crc kubenswrapper[4885]: I1002 02:00:07.871375 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l4ktn" event={"ID":"c5b45f09-e4c2-465b-9ba1-6d9375e5b246","Type":"ContainerDied","Data":"9b1bbf0001f658947690f50f678f07febae92b609dfa7e44fe9fc13fa6d0a87d"} Oct 02 02:00:07 crc kubenswrapper[4885]: I1002 02:00:07.871414 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l4ktn" event={"ID":"c5b45f09-e4c2-465b-9ba1-6d9375e5b246","Type":"ContainerStarted","Data":"c496fc7fbf7beaebf66b1a9be9db6719544dc23bb84233b475f16f21131fdd00"} Oct 02 02:00:08 crc kubenswrapper[4885]: I1002 02:00:08.651651 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cbbec88b66c61e3d10a0bbc868e336160e79c262c68f9747aec27f06598t8b5"] Oct 02 02:00:08 crc kubenswrapper[4885]: I1002 02:00:08.654352 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cbbec88b66c61e3d10a0bbc868e336160e79c262c68f9747aec27f06598t8b5" Oct 02 02:00:08 crc kubenswrapper[4885]: I1002 02:00:08.659930 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-d82vk" Oct 02 02:00:08 crc kubenswrapper[4885]: I1002 02:00:08.663346 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cbbec88b66c61e3d10a0bbc868e336160e79c262c68f9747aec27f06598t8b5"] Oct 02 02:00:08 crc kubenswrapper[4885]: I1002 02:00:08.810643 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5b1a29e8-0ed9-4c6a-bcc5-65849560bd7a-util\") pod \"cbbec88b66c61e3d10a0bbc868e336160e79c262c68f9747aec27f06598t8b5\" (UID: \"5b1a29e8-0ed9-4c6a-bcc5-65849560bd7a\") " pod="openstack-operators/cbbec88b66c61e3d10a0bbc868e336160e79c262c68f9747aec27f06598t8b5" Oct 02 02:00:08 crc kubenswrapper[4885]: I1002 02:00:08.810725 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5b1a29e8-0ed9-4c6a-bcc5-65849560bd7a-bundle\") pod \"cbbec88b66c61e3d10a0bbc868e336160e79c262c68f9747aec27f06598t8b5\" (UID: \"5b1a29e8-0ed9-4c6a-bcc5-65849560bd7a\") " pod="openstack-operators/cbbec88b66c61e3d10a0bbc868e336160e79c262c68f9747aec27f06598t8b5" Oct 02 02:00:08 crc kubenswrapper[4885]: I1002 02:00:08.810959 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mtkpj\" (UniqueName: \"kubernetes.io/projected/5b1a29e8-0ed9-4c6a-bcc5-65849560bd7a-kube-api-access-mtkpj\") pod \"cbbec88b66c61e3d10a0bbc868e336160e79c262c68f9747aec27f06598t8b5\" (UID: \"5b1a29e8-0ed9-4c6a-bcc5-65849560bd7a\") " pod="openstack-operators/cbbec88b66c61e3d10a0bbc868e336160e79c262c68f9747aec27f06598t8b5" Oct 02 02:00:08 crc kubenswrapper[4885]: I1002 02:00:08.912585 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mtkpj\" (UniqueName: \"kubernetes.io/projected/5b1a29e8-0ed9-4c6a-bcc5-65849560bd7a-kube-api-access-mtkpj\") pod \"cbbec88b66c61e3d10a0bbc868e336160e79c262c68f9747aec27f06598t8b5\" (UID: \"5b1a29e8-0ed9-4c6a-bcc5-65849560bd7a\") " pod="openstack-operators/cbbec88b66c61e3d10a0bbc868e336160e79c262c68f9747aec27f06598t8b5" Oct 02 02:00:08 crc kubenswrapper[4885]: I1002 02:00:08.912825 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5b1a29e8-0ed9-4c6a-bcc5-65849560bd7a-util\") pod \"cbbec88b66c61e3d10a0bbc868e336160e79c262c68f9747aec27f06598t8b5\" (UID: \"5b1a29e8-0ed9-4c6a-bcc5-65849560bd7a\") " pod="openstack-operators/cbbec88b66c61e3d10a0bbc868e336160e79c262c68f9747aec27f06598t8b5" Oct 02 02:00:08 crc kubenswrapper[4885]: I1002 02:00:08.912895 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5b1a29e8-0ed9-4c6a-bcc5-65849560bd7a-bundle\") pod \"cbbec88b66c61e3d10a0bbc868e336160e79c262c68f9747aec27f06598t8b5\" (UID: \"5b1a29e8-0ed9-4c6a-bcc5-65849560bd7a\") " pod="openstack-operators/cbbec88b66c61e3d10a0bbc868e336160e79c262c68f9747aec27f06598t8b5" Oct 02 02:00:08 crc kubenswrapper[4885]: I1002 02:00:08.913721 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5b1a29e8-0ed9-4c6a-bcc5-65849560bd7a-util\") pod \"cbbec88b66c61e3d10a0bbc868e336160e79c262c68f9747aec27f06598t8b5\" (UID: \"5b1a29e8-0ed9-4c6a-bcc5-65849560bd7a\") " pod="openstack-operators/cbbec88b66c61e3d10a0bbc868e336160e79c262c68f9747aec27f06598t8b5" Oct 02 02:00:08 crc kubenswrapper[4885]: I1002 02:00:08.913828 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5b1a29e8-0ed9-4c6a-bcc5-65849560bd7a-bundle\") pod \"cbbec88b66c61e3d10a0bbc868e336160e79c262c68f9747aec27f06598t8b5\" (UID: \"5b1a29e8-0ed9-4c6a-bcc5-65849560bd7a\") " pod="openstack-operators/cbbec88b66c61e3d10a0bbc868e336160e79c262c68f9747aec27f06598t8b5" Oct 02 02:00:08 crc kubenswrapper[4885]: I1002 02:00:08.945952 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mtkpj\" (UniqueName: \"kubernetes.io/projected/5b1a29e8-0ed9-4c6a-bcc5-65849560bd7a-kube-api-access-mtkpj\") pod \"cbbec88b66c61e3d10a0bbc868e336160e79c262c68f9747aec27f06598t8b5\" (UID: \"5b1a29e8-0ed9-4c6a-bcc5-65849560bd7a\") " pod="openstack-operators/cbbec88b66c61e3d10a0bbc868e336160e79c262c68f9747aec27f06598t8b5" Oct 02 02:00:08 crc kubenswrapper[4885]: I1002 02:00:08.985903 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cbbec88b66c61e3d10a0bbc868e336160e79c262c68f9747aec27f06598t8b5" Oct 02 02:00:09 crc kubenswrapper[4885]: I1002 02:00:09.278628 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cbbec88b66c61e3d10a0bbc868e336160e79c262c68f9747aec27f06598t8b5"] Oct 02 02:00:09 crc kubenswrapper[4885]: W1002 02:00:09.290057 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5b1a29e8_0ed9_4c6a_bcc5_65849560bd7a.slice/crio-70da3440b0fcea4e24e37276afddb595109809a81fdad0310cb2ad9d22572498 WatchSource:0}: Error finding container 70da3440b0fcea4e24e37276afddb595109809a81fdad0310cb2ad9d22572498: Status 404 returned error can't find the container with id 70da3440b0fcea4e24e37276afddb595109809a81fdad0310cb2ad9d22572498 Oct 02 02:00:09 crc kubenswrapper[4885]: I1002 02:00:09.892737 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l4ktn" event={"ID":"c5b45f09-e4c2-465b-9ba1-6d9375e5b246","Type":"ContainerStarted","Data":"ee20aaaecc06fe69d98a88b051af7d44717f934f7cb4a9627f51add2b914d933"} Oct 02 02:00:09 crc kubenswrapper[4885]: I1002 02:00:09.895037 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cbbec88b66c61e3d10a0bbc868e336160e79c262c68f9747aec27f06598t8b5" event={"ID":"5b1a29e8-0ed9-4c6a-bcc5-65849560bd7a","Type":"ContainerDied","Data":"663d031ce517b385dcab5098d2f7c70f4a130ad4b6ac61165d6fa141de04c11e"} Oct 02 02:00:09 crc kubenswrapper[4885]: I1002 02:00:09.896444 4885 generic.go:334] "Generic (PLEG): container finished" podID="5b1a29e8-0ed9-4c6a-bcc5-65849560bd7a" containerID="663d031ce517b385dcab5098d2f7c70f4a130ad4b6ac61165d6fa141de04c11e" exitCode=0 Oct 02 02:00:09 crc kubenswrapper[4885]: I1002 02:00:09.896527 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cbbec88b66c61e3d10a0bbc868e336160e79c262c68f9747aec27f06598t8b5" event={"ID":"5b1a29e8-0ed9-4c6a-bcc5-65849560bd7a","Type":"ContainerStarted","Data":"70da3440b0fcea4e24e37276afddb595109809a81fdad0310cb2ad9d22572498"} Oct 02 02:00:10 crc kubenswrapper[4885]: I1002 02:00:10.907026 4885 generic.go:334] "Generic (PLEG): container finished" podID="c5b45f09-e4c2-465b-9ba1-6d9375e5b246" containerID="ee20aaaecc06fe69d98a88b051af7d44717f934f7cb4a9627f51add2b914d933" exitCode=0 Oct 02 02:00:10 crc kubenswrapper[4885]: I1002 02:00:10.907096 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l4ktn" event={"ID":"c5b45f09-e4c2-465b-9ba1-6d9375e5b246","Type":"ContainerDied","Data":"ee20aaaecc06fe69d98a88b051af7d44717f934f7cb4a9627f51add2b914d933"} Oct 02 02:00:10 crc kubenswrapper[4885]: I1002 02:00:10.910587 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cbbec88b66c61e3d10a0bbc868e336160e79c262c68f9747aec27f06598t8b5" event={"ID":"5b1a29e8-0ed9-4c6a-bcc5-65849560bd7a","Type":"ContainerStarted","Data":"dc085993ba8567bf5df51917b9448400137c6c573ab890b6a1039dcba537f763"} Oct 02 02:00:11 crc kubenswrapper[4885]: I1002 02:00:11.919703 4885 generic.go:334] "Generic (PLEG): container finished" podID="5b1a29e8-0ed9-4c6a-bcc5-65849560bd7a" containerID="dc085993ba8567bf5df51917b9448400137c6c573ab890b6a1039dcba537f763" exitCode=0 Oct 02 02:00:11 crc kubenswrapper[4885]: I1002 02:00:11.919903 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cbbec88b66c61e3d10a0bbc868e336160e79c262c68f9747aec27f06598t8b5" event={"ID":"5b1a29e8-0ed9-4c6a-bcc5-65849560bd7a","Type":"ContainerDied","Data":"dc085993ba8567bf5df51917b9448400137c6c573ab890b6a1039dcba537f763"} Oct 02 02:00:12 crc kubenswrapper[4885]: I1002 02:00:12.930213 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l4ktn" event={"ID":"c5b45f09-e4c2-465b-9ba1-6d9375e5b246","Type":"ContainerStarted","Data":"f62c769ad94bc9604c8d96c4278eb0a1b6bbd82ff458de630362b9b7fa9a57cb"} Oct 02 02:00:12 crc kubenswrapper[4885]: I1002 02:00:12.934305 4885 generic.go:334] "Generic (PLEG): container finished" podID="5b1a29e8-0ed9-4c6a-bcc5-65849560bd7a" containerID="2d2326aba25609724e4891511a70f844ebe2b3de1146f769730af7489d17a5fd" exitCode=0 Oct 02 02:00:12 crc kubenswrapper[4885]: I1002 02:00:12.934366 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cbbec88b66c61e3d10a0bbc868e336160e79c262c68f9747aec27f06598t8b5" event={"ID":"5b1a29e8-0ed9-4c6a-bcc5-65849560bd7a","Type":"ContainerDied","Data":"2d2326aba25609724e4891511a70f844ebe2b3de1146f769730af7489d17a5fd"} Oct 02 02:00:12 crc kubenswrapper[4885]: I1002 02:00:12.959109 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-l4ktn" podStartSLOduration=3.976265853 podStartE2EDuration="7.959085686s" podCreationTimestamp="2025-10-02 02:00:05 +0000 UTC" firstStartedPulling="2025-10-02 02:00:07.874206941 +0000 UTC m=+796.685954380" lastFinishedPulling="2025-10-02 02:00:11.857026804 +0000 UTC m=+800.668774213" observedRunningTime="2025-10-02 02:00:12.953458545 +0000 UTC m=+801.765206034" watchObservedRunningTime="2025-10-02 02:00:12.959085686 +0000 UTC m=+801.770833125" Oct 02 02:00:13 crc kubenswrapper[4885]: I1002 02:00:13.265555 4885 patch_prober.go:28] interesting pod/machine-config-daemon-rttx8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 02:00:13 crc kubenswrapper[4885]: I1002 02:00:13.265642 4885 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 02:00:13 crc kubenswrapper[4885]: I1002 02:00:13.265707 4885 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" Oct 02 02:00:13 crc kubenswrapper[4885]: I1002 02:00:13.266469 4885 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"83cae60dd6556fc08f2eddf3fd26fd0abb56b3d0a4d62534298ea5f48fc39e18"} pod="openshift-machine-config-operator/machine-config-daemon-rttx8" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 02:00:13 crc kubenswrapper[4885]: I1002 02:00:13.266562 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" containerName="machine-config-daemon" containerID="cri-o://83cae60dd6556fc08f2eddf3fd26fd0abb56b3d0a4d62534298ea5f48fc39e18" gracePeriod=600 Oct 02 02:00:13 crc kubenswrapper[4885]: I1002 02:00:13.943828 4885 generic.go:334] "Generic (PLEG): container finished" podID="d169c8ba-000a-4218-8e3b-4ae53035b110" containerID="83cae60dd6556fc08f2eddf3fd26fd0abb56b3d0a4d62534298ea5f48fc39e18" exitCode=0 Oct 02 02:00:13 crc kubenswrapper[4885]: I1002 02:00:13.943888 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" event={"ID":"d169c8ba-000a-4218-8e3b-4ae53035b110","Type":"ContainerDied","Data":"83cae60dd6556fc08f2eddf3fd26fd0abb56b3d0a4d62534298ea5f48fc39e18"} Oct 02 02:00:13 crc kubenswrapper[4885]: I1002 02:00:13.944228 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" event={"ID":"d169c8ba-000a-4218-8e3b-4ae53035b110","Type":"ContainerStarted","Data":"4708a92cba43cfdb8c289fa5ff76e06b7b83a40fb102347f523fcefc325254bd"} Oct 02 02:00:13 crc kubenswrapper[4885]: I1002 02:00:13.944251 4885 scope.go:117] "RemoveContainer" containerID="a1c63629d5a2ca3023ebf7324c43c17375f971212dce224761d6f69e9b4ea0ce" Oct 02 02:00:14 crc kubenswrapper[4885]: I1002 02:00:14.284251 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cbbec88b66c61e3d10a0bbc868e336160e79c262c68f9747aec27f06598t8b5" Oct 02 02:00:14 crc kubenswrapper[4885]: I1002 02:00:14.400298 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mtkpj\" (UniqueName: \"kubernetes.io/projected/5b1a29e8-0ed9-4c6a-bcc5-65849560bd7a-kube-api-access-mtkpj\") pod \"5b1a29e8-0ed9-4c6a-bcc5-65849560bd7a\" (UID: \"5b1a29e8-0ed9-4c6a-bcc5-65849560bd7a\") " Oct 02 02:00:14 crc kubenswrapper[4885]: I1002 02:00:14.400433 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5b1a29e8-0ed9-4c6a-bcc5-65849560bd7a-bundle\") pod \"5b1a29e8-0ed9-4c6a-bcc5-65849560bd7a\" (UID: \"5b1a29e8-0ed9-4c6a-bcc5-65849560bd7a\") " Oct 02 02:00:14 crc kubenswrapper[4885]: I1002 02:00:14.400584 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5b1a29e8-0ed9-4c6a-bcc5-65849560bd7a-util\") pod \"5b1a29e8-0ed9-4c6a-bcc5-65849560bd7a\" (UID: \"5b1a29e8-0ed9-4c6a-bcc5-65849560bd7a\") " Oct 02 02:00:14 crc kubenswrapper[4885]: I1002 02:00:14.402335 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5b1a29e8-0ed9-4c6a-bcc5-65849560bd7a-bundle" (OuterVolumeSpecName: "bundle") pod "5b1a29e8-0ed9-4c6a-bcc5-65849560bd7a" (UID: "5b1a29e8-0ed9-4c6a-bcc5-65849560bd7a"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:00:14 crc kubenswrapper[4885]: I1002 02:00:14.410653 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b1a29e8-0ed9-4c6a-bcc5-65849560bd7a-kube-api-access-mtkpj" (OuterVolumeSpecName: "kube-api-access-mtkpj") pod "5b1a29e8-0ed9-4c6a-bcc5-65849560bd7a" (UID: "5b1a29e8-0ed9-4c6a-bcc5-65849560bd7a"). InnerVolumeSpecName "kube-api-access-mtkpj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:00:14 crc kubenswrapper[4885]: I1002 02:00:14.426111 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5b1a29e8-0ed9-4c6a-bcc5-65849560bd7a-util" (OuterVolumeSpecName: "util") pod "5b1a29e8-0ed9-4c6a-bcc5-65849560bd7a" (UID: "5b1a29e8-0ed9-4c6a-bcc5-65849560bd7a"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:00:14 crc kubenswrapper[4885]: I1002 02:00:14.501905 4885 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5b1a29e8-0ed9-4c6a-bcc5-65849560bd7a-util\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:14 crc kubenswrapper[4885]: I1002 02:00:14.501954 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mtkpj\" (UniqueName: \"kubernetes.io/projected/5b1a29e8-0ed9-4c6a-bcc5-65849560bd7a-kube-api-access-mtkpj\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:14 crc kubenswrapper[4885]: I1002 02:00:14.501978 4885 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5b1a29e8-0ed9-4c6a-bcc5-65849560bd7a-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:14 crc kubenswrapper[4885]: I1002 02:00:14.957921 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cbbec88b66c61e3d10a0bbc868e336160e79c262c68f9747aec27f06598t8b5" event={"ID":"5b1a29e8-0ed9-4c6a-bcc5-65849560bd7a","Type":"ContainerDied","Data":"70da3440b0fcea4e24e37276afddb595109809a81fdad0310cb2ad9d22572498"} Oct 02 02:00:14 crc kubenswrapper[4885]: I1002 02:00:14.957977 4885 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="70da3440b0fcea4e24e37276afddb595109809a81fdad0310cb2ad9d22572498" Oct 02 02:00:14 crc kubenswrapper[4885]: I1002 02:00:14.958013 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cbbec88b66c61e3d10a0bbc868e336160e79c262c68f9747aec27f06598t8b5" Oct 02 02:00:16 crc kubenswrapper[4885]: I1002 02:00:16.340515 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-l4ktn" Oct 02 02:00:16 crc kubenswrapper[4885]: I1002 02:00:16.341341 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-l4ktn" Oct 02 02:00:17 crc kubenswrapper[4885]: I1002 02:00:17.398150 4885 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-l4ktn" podUID="c5b45f09-e4c2-465b-9ba1-6d9375e5b246" containerName="registry-server" probeResult="failure" output=< Oct 02 02:00:17 crc kubenswrapper[4885]: timeout: failed to connect service ":50051" within 1s Oct 02 02:00:17 crc kubenswrapper[4885]: > Oct 02 02:00:17 crc kubenswrapper[4885]: I1002 02:00:17.950830 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-6bcc7488c-bc4gx"] Oct 02 02:00:17 crc kubenswrapper[4885]: E1002 02:00:17.951386 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b1a29e8-0ed9-4c6a-bcc5-65849560bd7a" containerName="extract" Oct 02 02:00:17 crc kubenswrapper[4885]: I1002 02:00:17.951499 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b1a29e8-0ed9-4c6a-bcc5-65849560bd7a" containerName="extract" Oct 02 02:00:17 crc kubenswrapper[4885]: E1002 02:00:17.951601 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b1a29e8-0ed9-4c6a-bcc5-65849560bd7a" containerName="pull" Oct 02 02:00:17 crc kubenswrapper[4885]: I1002 02:00:17.951677 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b1a29e8-0ed9-4c6a-bcc5-65849560bd7a" containerName="pull" Oct 02 02:00:17 crc kubenswrapper[4885]: E1002 02:00:17.951756 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b1a29e8-0ed9-4c6a-bcc5-65849560bd7a" containerName="util" Oct 02 02:00:17 crc kubenswrapper[4885]: I1002 02:00:17.951837 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b1a29e8-0ed9-4c6a-bcc5-65849560bd7a" containerName="util" Oct 02 02:00:17 crc kubenswrapper[4885]: I1002 02:00:17.952030 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b1a29e8-0ed9-4c6a-bcc5-65849560bd7a" containerName="extract" Oct 02 02:00:17 crc kubenswrapper[4885]: I1002 02:00:17.952820 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-6bcc7488c-bc4gx" Oct 02 02:00:17 crc kubenswrapper[4885]: I1002 02:00:17.954982 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-94w4d" Oct 02 02:00:17 crc kubenswrapper[4885]: I1002 02:00:17.970165 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-6bcc7488c-bc4gx"] Oct 02 02:00:18 crc kubenswrapper[4885]: I1002 02:00:18.050841 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wct2f\" (UniqueName: \"kubernetes.io/projected/02e27ea3-228f-4f78-9135-deaff888afd2-kube-api-access-wct2f\") pod \"openstack-operator-controller-operator-6bcc7488c-bc4gx\" (UID: \"02e27ea3-228f-4f78-9135-deaff888afd2\") " pod="openstack-operators/openstack-operator-controller-operator-6bcc7488c-bc4gx" Oct 02 02:00:18 crc kubenswrapper[4885]: I1002 02:00:18.152180 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wct2f\" (UniqueName: \"kubernetes.io/projected/02e27ea3-228f-4f78-9135-deaff888afd2-kube-api-access-wct2f\") pod \"openstack-operator-controller-operator-6bcc7488c-bc4gx\" (UID: \"02e27ea3-228f-4f78-9135-deaff888afd2\") " pod="openstack-operators/openstack-operator-controller-operator-6bcc7488c-bc4gx" Oct 02 02:00:18 crc kubenswrapper[4885]: I1002 02:00:18.178124 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wct2f\" (UniqueName: \"kubernetes.io/projected/02e27ea3-228f-4f78-9135-deaff888afd2-kube-api-access-wct2f\") pod \"openstack-operator-controller-operator-6bcc7488c-bc4gx\" (UID: \"02e27ea3-228f-4f78-9135-deaff888afd2\") " pod="openstack-operators/openstack-operator-controller-operator-6bcc7488c-bc4gx" Oct 02 02:00:18 crc kubenswrapper[4885]: I1002 02:00:18.313526 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-6bcc7488c-bc4gx" Oct 02 02:00:18 crc kubenswrapper[4885]: I1002 02:00:18.565797 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-6bcc7488c-bc4gx"] Oct 02 02:00:18 crc kubenswrapper[4885]: W1002 02:00:18.581405 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod02e27ea3_228f_4f78_9135_deaff888afd2.slice/crio-3d6b99da8e9a41a22ecccb029baa4df8616557c713ad6518398cb1c4bfd01b8a WatchSource:0}: Error finding container 3d6b99da8e9a41a22ecccb029baa4df8616557c713ad6518398cb1c4bfd01b8a: Status 404 returned error can't find the container with id 3d6b99da8e9a41a22ecccb029baa4df8616557c713ad6518398cb1c4bfd01b8a Oct 02 02:00:19 crc kubenswrapper[4885]: I1002 02:00:19.031854 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-6bcc7488c-bc4gx" event={"ID":"02e27ea3-228f-4f78-9135-deaff888afd2","Type":"ContainerStarted","Data":"3d6b99da8e9a41a22ecccb029baa4df8616557c713ad6518398cb1c4bfd01b8a"} Oct 02 02:00:20 crc kubenswrapper[4885]: I1002 02:00:20.401700 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-5sx8b"] Oct 02 02:00:20 crc kubenswrapper[4885]: I1002 02:00:20.403493 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5sx8b" Oct 02 02:00:20 crc kubenswrapper[4885]: I1002 02:00:20.409064 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5sx8b"] Oct 02 02:00:20 crc kubenswrapper[4885]: I1002 02:00:20.494456 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4488b5d0-76fc-453f-911e-ac68246b7c11-utilities\") pod \"certified-operators-5sx8b\" (UID: \"4488b5d0-76fc-453f-911e-ac68246b7c11\") " pod="openshift-marketplace/certified-operators-5sx8b" Oct 02 02:00:20 crc kubenswrapper[4885]: I1002 02:00:20.494540 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4488b5d0-76fc-453f-911e-ac68246b7c11-catalog-content\") pod \"certified-operators-5sx8b\" (UID: \"4488b5d0-76fc-453f-911e-ac68246b7c11\") " pod="openshift-marketplace/certified-operators-5sx8b" Oct 02 02:00:20 crc kubenswrapper[4885]: I1002 02:00:20.494600 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p5xd4\" (UniqueName: \"kubernetes.io/projected/4488b5d0-76fc-453f-911e-ac68246b7c11-kube-api-access-p5xd4\") pod \"certified-operators-5sx8b\" (UID: \"4488b5d0-76fc-453f-911e-ac68246b7c11\") " pod="openshift-marketplace/certified-operators-5sx8b" Oct 02 02:00:20 crc kubenswrapper[4885]: I1002 02:00:20.596301 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p5xd4\" (UniqueName: \"kubernetes.io/projected/4488b5d0-76fc-453f-911e-ac68246b7c11-kube-api-access-p5xd4\") pod \"certified-operators-5sx8b\" (UID: \"4488b5d0-76fc-453f-911e-ac68246b7c11\") " pod="openshift-marketplace/certified-operators-5sx8b" Oct 02 02:00:20 crc kubenswrapper[4885]: I1002 02:00:20.596370 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4488b5d0-76fc-453f-911e-ac68246b7c11-utilities\") pod \"certified-operators-5sx8b\" (UID: \"4488b5d0-76fc-453f-911e-ac68246b7c11\") " pod="openshift-marketplace/certified-operators-5sx8b" Oct 02 02:00:20 crc kubenswrapper[4885]: I1002 02:00:20.596412 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4488b5d0-76fc-453f-911e-ac68246b7c11-catalog-content\") pod \"certified-operators-5sx8b\" (UID: \"4488b5d0-76fc-453f-911e-ac68246b7c11\") " pod="openshift-marketplace/certified-operators-5sx8b" Oct 02 02:00:20 crc kubenswrapper[4885]: I1002 02:00:20.596876 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4488b5d0-76fc-453f-911e-ac68246b7c11-utilities\") pod \"certified-operators-5sx8b\" (UID: \"4488b5d0-76fc-453f-911e-ac68246b7c11\") " pod="openshift-marketplace/certified-operators-5sx8b" Oct 02 02:00:20 crc kubenswrapper[4885]: I1002 02:00:20.598299 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4488b5d0-76fc-453f-911e-ac68246b7c11-catalog-content\") pod \"certified-operators-5sx8b\" (UID: \"4488b5d0-76fc-453f-911e-ac68246b7c11\") " pod="openshift-marketplace/certified-operators-5sx8b" Oct 02 02:00:20 crc kubenswrapper[4885]: I1002 02:00:20.613857 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p5xd4\" (UniqueName: \"kubernetes.io/projected/4488b5d0-76fc-453f-911e-ac68246b7c11-kube-api-access-p5xd4\") pod \"certified-operators-5sx8b\" (UID: \"4488b5d0-76fc-453f-911e-ac68246b7c11\") " pod="openshift-marketplace/certified-operators-5sx8b" Oct 02 02:00:20 crc kubenswrapper[4885]: I1002 02:00:20.719836 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5sx8b" Oct 02 02:00:24 crc kubenswrapper[4885]: I1002 02:00:24.316915 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5sx8b"] Oct 02 02:00:25 crc kubenswrapper[4885]: I1002 02:00:25.079529 4885 generic.go:334] "Generic (PLEG): container finished" podID="4488b5d0-76fc-453f-911e-ac68246b7c11" containerID="ebd2c1bfeaca97477ee5221df53322c296cfa7c2d63023869c12d9a49669c1b1" exitCode=0 Oct 02 02:00:25 crc kubenswrapper[4885]: I1002 02:00:25.079633 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5sx8b" event={"ID":"4488b5d0-76fc-453f-911e-ac68246b7c11","Type":"ContainerDied","Data":"ebd2c1bfeaca97477ee5221df53322c296cfa7c2d63023869c12d9a49669c1b1"} Oct 02 02:00:25 crc kubenswrapper[4885]: I1002 02:00:25.080115 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5sx8b" event={"ID":"4488b5d0-76fc-453f-911e-ac68246b7c11","Type":"ContainerStarted","Data":"82a9c3434dcdde1a7ab3855bbd739926655d7c7b095838314bbc5fd0ec677ef3"} Oct 02 02:00:25 crc kubenswrapper[4885]: I1002 02:00:25.082101 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-6bcc7488c-bc4gx" event={"ID":"02e27ea3-228f-4f78-9135-deaff888afd2","Type":"ContainerStarted","Data":"b8cd8f8a5fb5d5022fce4e9e2303fd224891992bb78eecedb6f08cbd3978bdb5"} Oct 02 02:00:25 crc kubenswrapper[4885]: I1002 02:00:25.798753 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-qtvx4"] Oct 02 02:00:25 crc kubenswrapper[4885]: I1002 02:00:25.799810 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qtvx4" Oct 02 02:00:25 crc kubenswrapper[4885]: I1002 02:00:25.819846 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-qtvx4"] Oct 02 02:00:25 crc kubenswrapper[4885]: I1002 02:00:25.878901 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-phv7p\" (UniqueName: \"kubernetes.io/projected/608b9eff-222e-4bbf-a865-5f38a2f2804c-kube-api-access-phv7p\") pod \"community-operators-qtvx4\" (UID: \"608b9eff-222e-4bbf-a865-5f38a2f2804c\") " pod="openshift-marketplace/community-operators-qtvx4" Oct 02 02:00:25 crc kubenswrapper[4885]: I1002 02:00:25.878952 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/608b9eff-222e-4bbf-a865-5f38a2f2804c-utilities\") pod \"community-operators-qtvx4\" (UID: \"608b9eff-222e-4bbf-a865-5f38a2f2804c\") " pod="openshift-marketplace/community-operators-qtvx4" Oct 02 02:00:25 crc kubenswrapper[4885]: I1002 02:00:25.879011 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/608b9eff-222e-4bbf-a865-5f38a2f2804c-catalog-content\") pod \"community-operators-qtvx4\" (UID: \"608b9eff-222e-4bbf-a865-5f38a2f2804c\") " pod="openshift-marketplace/community-operators-qtvx4" Oct 02 02:00:25 crc kubenswrapper[4885]: I1002 02:00:25.980820 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/608b9eff-222e-4bbf-a865-5f38a2f2804c-catalog-content\") pod \"community-operators-qtvx4\" (UID: \"608b9eff-222e-4bbf-a865-5f38a2f2804c\") " pod="openshift-marketplace/community-operators-qtvx4" Oct 02 02:00:25 crc kubenswrapper[4885]: I1002 02:00:25.980925 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-phv7p\" (UniqueName: \"kubernetes.io/projected/608b9eff-222e-4bbf-a865-5f38a2f2804c-kube-api-access-phv7p\") pod \"community-operators-qtvx4\" (UID: \"608b9eff-222e-4bbf-a865-5f38a2f2804c\") " pod="openshift-marketplace/community-operators-qtvx4" Oct 02 02:00:25 crc kubenswrapper[4885]: I1002 02:00:25.980956 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/608b9eff-222e-4bbf-a865-5f38a2f2804c-utilities\") pod \"community-operators-qtvx4\" (UID: \"608b9eff-222e-4bbf-a865-5f38a2f2804c\") " pod="openshift-marketplace/community-operators-qtvx4" Oct 02 02:00:25 crc kubenswrapper[4885]: I1002 02:00:25.981483 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/608b9eff-222e-4bbf-a865-5f38a2f2804c-utilities\") pod \"community-operators-qtvx4\" (UID: \"608b9eff-222e-4bbf-a865-5f38a2f2804c\") " pod="openshift-marketplace/community-operators-qtvx4" Oct 02 02:00:25 crc kubenswrapper[4885]: I1002 02:00:25.981483 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/608b9eff-222e-4bbf-a865-5f38a2f2804c-catalog-content\") pod \"community-operators-qtvx4\" (UID: \"608b9eff-222e-4bbf-a865-5f38a2f2804c\") " pod="openshift-marketplace/community-operators-qtvx4" Oct 02 02:00:26 crc kubenswrapper[4885]: I1002 02:00:26.008239 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-phv7p\" (UniqueName: \"kubernetes.io/projected/608b9eff-222e-4bbf-a865-5f38a2f2804c-kube-api-access-phv7p\") pod \"community-operators-qtvx4\" (UID: \"608b9eff-222e-4bbf-a865-5f38a2f2804c\") " pod="openshift-marketplace/community-operators-qtvx4" Oct 02 02:00:26 crc kubenswrapper[4885]: I1002 02:00:26.159966 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qtvx4" Oct 02 02:00:26 crc kubenswrapper[4885]: I1002 02:00:26.378388 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-l4ktn" Oct 02 02:00:26 crc kubenswrapper[4885]: I1002 02:00:26.415343 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-l4ktn" Oct 02 02:00:27 crc kubenswrapper[4885]: I1002 02:00:27.097291 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-6bcc7488c-bc4gx" event={"ID":"02e27ea3-228f-4f78-9135-deaff888afd2","Type":"ContainerStarted","Data":"1f1a5281609e17053e75edf8ed5a20c44ea8b42e2db111dfbce2e0d46b9139f4"} Oct 02 02:00:27 crc kubenswrapper[4885]: I1002 02:00:27.097901 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-6bcc7488c-bc4gx" Oct 02 02:00:27 crc kubenswrapper[4885]: I1002 02:00:27.098875 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5sx8b" event={"ID":"4488b5d0-76fc-453f-911e-ac68246b7c11","Type":"ContainerStarted","Data":"9d2cec7e22005559f0203f5aa0cb20614e243a84dc5ab19b6bc04239b5751fc4"} Oct 02 02:00:27 crc kubenswrapper[4885]: I1002 02:00:27.146563 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-6bcc7488c-bc4gx" podStartSLOduration=1.9734545890000001 podStartE2EDuration="10.146546719s" podCreationTimestamp="2025-10-02 02:00:17 +0000 UTC" firstStartedPulling="2025-10-02 02:00:18.583474181 +0000 UTC m=+807.395221590" lastFinishedPulling="2025-10-02 02:00:26.756566281 +0000 UTC m=+815.568313720" observedRunningTime="2025-10-02 02:00:27.143350552 +0000 UTC m=+815.955097961" watchObservedRunningTime="2025-10-02 02:00:27.146546719 +0000 UTC m=+815.958294118" Oct 02 02:00:27 crc kubenswrapper[4885]: I1002 02:00:27.234449 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-qtvx4"] Oct 02 02:00:27 crc kubenswrapper[4885]: W1002 02:00:27.259526 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod608b9eff_222e_4bbf_a865_5f38a2f2804c.slice/crio-4b49595145ab12d8ee2469102d607699bc350a3dee92f5c435a2776f04297684 WatchSource:0}: Error finding container 4b49595145ab12d8ee2469102d607699bc350a3dee92f5c435a2776f04297684: Status 404 returned error can't find the container with id 4b49595145ab12d8ee2469102d607699bc350a3dee92f5c435a2776f04297684 Oct 02 02:00:28 crc kubenswrapper[4885]: I1002 02:00:28.109218 4885 generic.go:334] "Generic (PLEG): container finished" podID="608b9eff-222e-4bbf-a865-5f38a2f2804c" containerID="c87fe689549c8e215196c5e90b63df242884efbd0a1bdde37b353cfcd428b4a7" exitCode=0 Oct 02 02:00:28 crc kubenswrapper[4885]: I1002 02:00:28.109309 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qtvx4" event={"ID":"608b9eff-222e-4bbf-a865-5f38a2f2804c","Type":"ContainerDied","Data":"c87fe689549c8e215196c5e90b63df242884efbd0a1bdde37b353cfcd428b4a7"} Oct 02 02:00:28 crc kubenswrapper[4885]: I1002 02:00:28.109816 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qtvx4" event={"ID":"608b9eff-222e-4bbf-a865-5f38a2f2804c","Type":"ContainerStarted","Data":"4b49595145ab12d8ee2469102d607699bc350a3dee92f5c435a2776f04297684"} Oct 02 02:00:28 crc kubenswrapper[4885]: I1002 02:00:28.122496 4885 generic.go:334] "Generic (PLEG): container finished" podID="4488b5d0-76fc-453f-911e-ac68246b7c11" containerID="9d2cec7e22005559f0203f5aa0cb20614e243a84dc5ab19b6bc04239b5751fc4" exitCode=0 Oct 02 02:00:28 crc kubenswrapper[4885]: I1002 02:00:28.122598 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5sx8b" event={"ID":"4488b5d0-76fc-453f-911e-ac68246b7c11","Type":"ContainerDied","Data":"9d2cec7e22005559f0203f5aa0cb20614e243a84dc5ab19b6bc04239b5751fc4"} Oct 02 02:00:29 crc kubenswrapper[4885]: I1002 02:00:29.132659 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5sx8b" event={"ID":"4488b5d0-76fc-453f-911e-ac68246b7c11","Type":"ContainerStarted","Data":"4501ce9da15b842883419622bff4b06d9cfaad9dfe8921ab35d2ae472aae0ad4"} Oct 02 02:00:29 crc kubenswrapper[4885]: I1002 02:00:29.168681 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-5sx8b" podStartSLOduration=5.583651057 podStartE2EDuration="9.168628222s" podCreationTimestamp="2025-10-02 02:00:20 +0000 UTC" firstStartedPulling="2025-10-02 02:00:25.082230503 +0000 UTC m=+813.893977932" lastFinishedPulling="2025-10-02 02:00:28.667207668 +0000 UTC m=+817.478955097" observedRunningTime="2025-10-02 02:00:29.160696091 +0000 UTC m=+817.972443530" watchObservedRunningTime="2025-10-02 02:00:29.168628222 +0000 UTC m=+817.980375691" Oct 02 02:00:30 crc kubenswrapper[4885]: I1002 02:00:30.143292 4885 generic.go:334] "Generic (PLEG): container finished" podID="608b9eff-222e-4bbf-a865-5f38a2f2804c" containerID="9f576da10f62db0a271609a872a249694acd748e059eaeb5d0ade3b66d5a2581" exitCode=0 Oct 02 02:00:30 crc kubenswrapper[4885]: I1002 02:00:30.143369 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qtvx4" event={"ID":"608b9eff-222e-4bbf-a865-5f38a2f2804c","Type":"ContainerDied","Data":"9f576da10f62db0a271609a872a249694acd748e059eaeb5d0ade3b66d5a2581"} Oct 02 02:00:30 crc kubenswrapper[4885]: I1002 02:00:30.720581 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-5sx8b" Oct 02 02:00:30 crc kubenswrapper[4885]: I1002 02:00:30.725957 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-5sx8b" Oct 02 02:00:31 crc kubenswrapper[4885]: I1002 02:00:31.158170 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qtvx4" event={"ID":"608b9eff-222e-4bbf-a865-5f38a2f2804c","Type":"ContainerStarted","Data":"68732860d2e2372b56ccfa48eca93b3aeb272f61d1ee58953443ac0ef92267cd"} Oct 02 02:00:31 crc kubenswrapper[4885]: I1002 02:00:31.204855 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-qtvx4" podStartSLOduration=3.665544357 podStartE2EDuration="6.204827363s" podCreationTimestamp="2025-10-02 02:00:25 +0000 UTC" firstStartedPulling="2025-10-02 02:00:28.112944279 +0000 UTC m=+816.924691708" lastFinishedPulling="2025-10-02 02:00:30.652227275 +0000 UTC m=+819.463974714" observedRunningTime="2025-10-02 02:00:31.19944893 +0000 UTC m=+820.011196369" watchObservedRunningTime="2025-10-02 02:00:31.204827363 +0000 UTC m=+820.016574792" Oct 02 02:00:31 crc kubenswrapper[4885]: I1002 02:00:31.790351 4885 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-5sx8b" podUID="4488b5d0-76fc-453f-911e-ac68246b7c11" containerName="registry-server" probeResult="failure" output=< Oct 02 02:00:31 crc kubenswrapper[4885]: timeout: failed to connect service ":50051" within 1s Oct 02 02:00:31 crc kubenswrapper[4885]: > Oct 02 02:00:32 crc kubenswrapper[4885]: I1002 02:00:32.194084 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-l4ktn"] Oct 02 02:00:32 crc kubenswrapper[4885]: I1002 02:00:32.194892 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-l4ktn" podUID="c5b45f09-e4c2-465b-9ba1-6d9375e5b246" containerName="registry-server" containerID="cri-o://f62c769ad94bc9604c8d96c4278eb0a1b6bbd82ff458de630362b9b7fa9a57cb" gracePeriod=2 Oct 02 02:00:33 crc kubenswrapper[4885]: I1002 02:00:33.141070 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l4ktn" Oct 02 02:00:33 crc kubenswrapper[4885]: I1002 02:00:33.176595 4885 generic.go:334] "Generic (PLEG): container finished" podID="c5b45f09-e4c2-465b-9ba1-6d9375e5b246" containerID="f62c769ad94bc9604c8d96c4278eb0a1b6bbd82ff458de630362b9b7fa9a57cb" exitCode=0 Oct 02 02:00:33 crc kubenswrapper[4885]: I1002 02:00:33.176651 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l4ktn" event={"ID":"c5b45f09-e4c2-465b-9ba1-6d9375e5b246","Type":"ContainerDied","Data":"f62c769ad94bc9604c8d96c4278eb0a1b6bbd82ff458de630362b9b7fa9a57cb"} Oct 02 02:00:33 crc kubenswrapper[4885]: I1002 02:00:33.176679 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l4ktn" event={"ID":"c5b45f09-e4c2-465b-9ba1-6d9375e5b246","Type":"ContainerDied","Data":"c496fc7fbf7beaebf66b1a9be9db6719544dc23bb84233b475f16f21131fdd00"} Oct 02 02:00:33 crc kubenswrapper[4885]: I1002 02:00:33.176695 4885 scope.go:117] "RemoveContainer" containerID="f62c769ad94bc9604c8d96c4278eb0a1b6bbd82ff458de630362b9b7fa9a57cb" Oct 02 02:00:33 crc kubenswrapper[4885]: I1002 02:00:33.176801 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l4ktn" Oct 02 02:00:33 crc kubenswrapper[4885]: I1002 02:00:33.207042 4885 scope.go:117] "RemoveContainer" containerID="ee20aaaecc06fe69d98a88b051af7d44717f934f7cb4a9627f51add2b914d933" Oct 02 02:00:33 crc kubenswrapper[4885]: I1002 02:00:33.231882 4885 scope.go:117] "RemoveContainer" containerID="9b1bbf0001f658947690f50f678f07febae92b609dfa7e44fe9fc13fa6d0a87d" Oct 02 02:00:33 crc kubenswrapper[4885]: I1002 02:00:33.250405 4885 scope.go:117] "RemoveContainer" containerID="f62c769ad94bc9604c8d96c4278eb0a1b6bbd82ff458de630362b9b7fa9a57cb" Oct 02 02:00:33 crc kubenswrapper[4885]: E1002 02:00:33.251022 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f62c769ad94bc9604c8d96c4278eb0a1b6bbd82ff458de630362b9b7fa9a57cb\": container with ID starting with f62c769ad94bc9604c8d96c4278eb0a1b6bbd82ff458de630362b9b7fa9a57cb not found: ID does not exist" containerID="f62c769ad94bc9604c8d96c4278eb0a1b6bbd82ff458de630362b9b7fa9a57cb" Oct 02 02:00:33 crc kubenswrapper[4885]: I1002 02:00:33.251077 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f62c769ad94bc9604c8d96c4278eb0a1b6bbd82ff458de630362b9b7fa9a57cb"} err="failed to get container status \"f62c769ad94bc9604c8d96c4278eb0a1b6bbd82ff458de630362b9b7fa9a57cb\": rpc error: code = NotFound desc = could not find container \"f62c769ad94bc9604c8d96c4278eb0a1b6bbd82ff458de630362b9b7fa9a57cb\": container with ID starting with f62c769ad94bc9604c8d96c4278eb0a1b6bbd82ff458de630362b9b7fa9a57cb not found: ID does not exist" Oct 02 02:00:33 crc kubenswrapper[4885]: I1002 02:00:33.251111 4885 scope.go:117] "RemoveContainer" containerID="ee20aaaecc06fe69d98a88b051af7d44717f934f7cb4a9627f51add2b914d933" Oct 02 02:00:33 crc kubenswrapper[4885]: E1002 02:00:33.251472 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ee20aaaecc06fe69d98a88b051af7d44717f934f7cb4a9627f51add2b914d933\": container with ID starting with ee20aaaecc06fe69d98a88b051af7d44717f934f7cb4a9627f51add2b914d933 not found: ID does not exist" containerID="ee20aaaecc06fe69d98a88b051af7d44717f934f7cb4a9627f51add2b914d933" Oct 02 02:00:33 crc kubenswrapper[4885]: I1002 02:00:33.251509 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee20aaaecc06fe69d98a88b051af7d44717f934f7cb4a9627f51add2b914d933"} err="failed to get container status \"ee20aaaecc06fe69d98a88b051af7d44717f934f7cb4a9627f51add2b914d933\": rpc error: code = NotFound desc = could not find container \"ee20aaaecc06fe69d98a88b051af7d44717f934f7cb4a9627f51add2b914d933\": container with ID starting with ee20aaaecc06fe69d98a88b051af7d44717f934f7cb4a9627f51add2b914d933 not found: ID does not exist" Oct 02 02:00:33 crc kubenswrapper[4885]: I1002 02:00:33.251534 4885 scope.go:117] "RemoveContainer" containerID="9b1bbf0001f658947690f50f678f07febae92b609dfa7e44fe9fc13fa6d0a87d" Oct 02 02:00:33 crc kubenswrapper[4885]: E1002 02:00:33.251835 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9b1bbf0001f658947690f50f678f07febae92b609dfa7e44fe9fc13fa6d0a87d\": container with ID starting with 9b1bbf0001f658947690f50f678f07febae92b609dfa7e44fe9fc13fa6d0a87d not found: ID does not exist" containerID="9b1bbf0001f658947690f50f678f07febae92b609dfa7e44fe9fc13fa6d0a87d" Oct 02 02:00:33 crc kubenswrapper[4885]: I1002 02:00:33.251950 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b1bbf0001f658947690f50f678f07febae92b609dfa7e44fe9fc13fa6d0a87d"} err="failed to get container status \"9b1bbf0001f658947690f50f678f07febae92b609dfa7e44fe9fc13fa6d0a87d\": rpc error: code = NotFound desc = could not find container \"9b1bbf0001f658947690f50f678f07febae92b609dfa7e44fe9fc13fa6d0a87d\": container with ID starting with 9b1bbf0001f658947690f50f678f07febae92b609dfa7e44fe9fc13fa6d0a87d not found: ID does not exist" Oct 02 02:00:33 crc kubenswrapper[4885]: I1002 02:00:33.272976 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c5b45f09-e4c2-465b-9ba1-6d9375e5b246-utilities\") pod \"c5b45f09-e4c2-465b-9ba1-6d9375e5b246\" (UID: \"c5b45f09-e4c2-465b-9ba1-6d9375e5b246\") " Oct 02 02:00:33 crc kubenswrapper[4885]: I1002 02:00:33.273043 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c5b45f09-e4c2-465b-9ba1-6d9375e5b246-catalog-content\") pod \"c5b45f09-e4c2-465b-9ba1-6d9375e5b246\" (UID: \"c5b45f09-e4c2-465b-9ba1-6d9375e5b246\") " Oct 02 02:00:33 crc kubenswrapper[4885]: I1002 02:00:33.273094 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pmc4n\" (UniqueName: \"kubernetes.io/projected/c5b45f09-e4c2-465b-9ba1-6d9375e5b246-kube-api-access-pmc4n\") pod \"c5b45f09-e4c2-465b-9ba1-6d9375e5b246\" (UID: \"c5b45f09-e4c2-465b-9ba1-6d9375e5b246\") " Oct 02 02:00:33 crc kubenswrapper[4885]: I1002 02:00:33.274373 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c5b45f09-e4c2-465b-9ba1-6d9375e5b246-utilities" (OuterVolumeSpecName: "utilities") pod "c5b45f09-e4c2-465b-9ba1-6d9375e5b246" (UID: "c5b45f09-e4c2-465b-9ba1-6d9375e5b246"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:00:33 crc kubenswrapper[4885]: I1002 02:00:33.282489 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c5b45f09-e4c2-465b-9ba1-6d9375e5b246-kube-api-access-pmc4n" (OuterVolumeSpecName: "kube-api-access-pmc4n") pod "c5b45f09-e4c2-465b-9ba1-6d9375e5b246" (UID: "c5b45f09-e4c2-465b-9ba1-6d9375e5b246"). InnerVolumeSpecName "kube-api-access-pmc4n". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:00:33 crc kubenswrapper[4885]: I1002 02:00:33.351853 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c5b45f09-e4c2-465b-9ba1-6d9375e5b246-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c5b45f09-e4c2-465b-9ba1-6d9375e5b246" (UID: "c5b45f09-e4c2-465b-9ba1-6d9375e5b246"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:00:33 crc kubenswrapper[4885]: I1002 02:00:33.374745 4885 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c5b45f09-e4c2-465b-9ba1-6d9375e5b246-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:33 crc kubenswrapper[4885]: I1002 02:00:33.374771 4885 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c5b45f09-e4c2-465b-9ba1-6d9375e5b246-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:33 crc kubenswrapper[4885]: I1002 02:00:33.374782 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pmc4n\" (UniqueName: \"kubernetes.io/projected/c5b45f09-e4c2-465b-9ba1-6d9375e5b246-kube-api-access-pmc4n\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:33 crc kubenswrapper[4885]: I1002 02:00:33.520866 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-l4ktn"] Oct 02 02:00:33 crc kubenswrapper[4885]: I1002 02:00:33.528908 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-l4ktn"] Oct 02 02:00:34 crc kubenswrapper[4885]: I1002 02:00:34.061160 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c5b45f09-e4c2-465b-9ba1-6d9375e5b246" path="/var/lib/kubelet/pods/c5b45f09-e4c2-465b-9ba1-6d9375e5b246/volumes" Oct 02 02:00:36 crc kubenswrapper[4885]: I1002 02:00:36.160616 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-qtvx4" Oct 02 02:00:36 crc kubenswrapper[4885]: I1002 02:00:36.161555 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-qtvx4" Oct 02 02:00:36 crc kubenswrapper[4885]: I1002 02:00:36.231531 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-qtvx4" Oct 02 02:00:36 crc kubenswrapper[4885]: I1002 02:00:36.301973 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-qtvx4" Oct 02 02:00:37 crc kubenswrapper[4885]: I1002 02:00:37.390538 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-qtvx4"] Oct 02 02:00:38 crc kubenswrapper[4885]: I1002 02:00:38.221596 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-qtvx4" podUID="608b9eff-222e-4bbf-a865-5f38a2f2804c" containerName="registry-server" containerID="cri-o://68732860d2e2372b56ccfa48eca93b3aeb272f61d1ee58953443ac0ef92267cd" gracePeriod=2 Oct 02 02:00:38 crc kubenswrapper[4885]: I1002 02:00:38.318142 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-6bcc7488c-bc4gx" Oct 02 02:00:39 crc kubenswrapper[4885]: I1002 02:00:39.232060 4885 generic.go:334] "Generic (PLEG): container finished" podID="608b9eff-222e-4bbf-a865-5f38a2f2804c" containerID="68732860d2e2372b56ccfa48eca93b3aeb272f61d1ee58953443ac0ef92267cd" exitCode=0 Oct 02 02:00:39 crc kubenswrapper[4885]: I1002 02:00:39.232155 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qtvx4" event={"ID":"608b9eff-222e-4bbf-a865-5f38a2f2804c","Type":"ContainerDied","Data":"68732860d2e2372b56ccfa48eca93b3aeb272f61d1ee58953443ac0ef92267cd"} Oct 02 02:00:39 crc kubenswrapper[4885]: I1002 02:00:39.749231 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qtvx4" Oct 02 02:00:39 crc kubenswrapper[4885]: I1002 02:00:39.868409 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/608b9eff-222e-4bbf-a865-5f38a2f2804c-catalog-content\") pod \"608b9eff-222e-4bbf-a865-5f38a2f2804c\" (UID: \"608b9eff-222e-4bbf-a865-5f38a2f2804c\") " Oct 02 02:00:39 crc kubenswrapper[4885]: I1002 02:00:39.868452 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/608b9eff-222e-4bbf-a865-5f38a2f2804c-utilities\") pod \"608b9eff-222e-4bbf-a865-5f38a2f2804c\" (UID: \"608b9eff-222e-4bbf-a865-5f38a2f2804c\") " Oct 02 02:00:39 crc kubenswrapper[4885]: I1002 02:00:39.868548 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-phv7p\" (UniqueName: \"kubernetes.io/projected/608b9eff-222e-4bbf-a865-5f38a2f2804c-kube-api-access-phv7p\") pod \"608b9eff-222e-4bbf-a865-5f38a2f2804c\" (UID: \"608b9eff-222e-4bbf-a865-5f38a2f2804c\") " Oct 02 02:00:39 crc kubenswrapper[4885]: I1002 02:00:39.869254 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/608b9eff-222e-4bbf-a865-5f38a2f2804c-utilities" (OuterVolumeSpecName: "utilities") pod "608b9eff-222e-4bbf-a865-5f38a2f2804c" (UID: "608b9eff-222e-4bbf-a865-5f38a2f2804c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:00:39 crc kubenswrapper[4885]: I1002 02:00:39.878522 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/608b9eff-222e-4bbf-a865-5f38a2f2804c-kube-api-access-phv7p" (OuterVolumeSpecName: "kube-api-access-phv7p") pod "608b9eff-222e-4bbf-a865-5f38a2f2804c" (UID: "608b9eff-222e-4bbf-a865-5f38a2f2804c"). InnerVolumeSpecName "kube-api-access-phv7p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:00:39 crc kubenswrapper[4885]: I1002 02:00:39.970172 4885 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/608b9eff-222e-4bbf-a865-5f38a2f2804c-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:39 crc kubenswrapper[4885]: I1002 02:00:39.970211 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-phv7p\" (UniqueName: \"kubernetes.io/projected/608b9eff-222e-4bbf-a865-5f38a2f2804c-kube-api-access-phv7p\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:40 crc kubenswrapper[4885]: I1002 02:00:40.242924 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qtvx4" event={"ID":"608b9eff-222e-4bbf-a865-5f38a2f2804c","Type":"ContainerDied","Data":"4b49595145ab12d8ee2469102d607699bc350a3dee92f5c435a2776f04297684"} Oct 02 02:00:40 crc kubenswrapper[4885]: I1002 02:00:40.243032 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qtvx4" Oct 02 02:00:40 crc kubenswrapper[4885]: I1002 02:00:40.243044 4885 scope.go:117] "RemoveContainer" containerID="68732860d2e2372b56ccfa48eca93b3aeb272f61d1ee58953443ac0ef92267cd" Oct 02 02:00:40 crc kubenswrapper[4885]: I1002 02:00:40.273011 4885 scope.go:117] "RemoveContainer" containerID="9f576da10f62db0a271609a872a249694acd748e059eaeb5d0ade3b66d5a2581" Oct 02 02:00:40 crc kubenswrapper[4885]: I1002 02:00:40.308572 4885 scope.go:117] "RemoveContainer" containerID="c87fe689549c8e215196c5e90b63df242884efbd0a1bdde37b353cfcd428b4a7" Oct 02 02:00:40 crc kubenswrapper[4885]: I1002 02:00:40.747324 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/608b9eff-222e-4bbf-a865-5f38a2f2804c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "608b9eff-222e-4bbf-a865-5f38a2f2804c" (UID: "608b9eff-222e-4bbf-a865-5f38a2f2804c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:00:40 crc kubenswrapper[4885]: I1002 02:00:40.781789 4885 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/608b9eff-222e-4bbf-a865-5f38a2f2804c-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:40 crc kubenswrapper[4885]: I1002 02:00:40.809034 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-5sx8b" Oct 02 02:00:40 crc kubenswrapper[4885]: I1002 02:00:40.844361 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-5sx8b" Oct 02 02:00:40 crc kubenswrapper[4885]: I1002 02:00:40.873676 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-qtvx4"] Oct 02 02:00:40 crc kubenswrapper[4885]: I1002 02:00:40.876981 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-qtvx4"] Oct 02 02:00:42 crc kubenswrapper[4885]: I1002 02:00:42.054404 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="608b9eff-222e-4bbf-a865-5f38a2f2804c" path="/var/lib/kubelet/pods/608b9eff-222e-4bbf-a865-5f38a2f2804c/volumes" Oct 02 02:00:42 crc kubenswrapper[4885]: I1002 02:00:42.587878 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5sx8b"] Oct 02 02:00:42 crc kubenswrapper[4885]: I1002 02:00:42.588077 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-5sx8b" podUID="4488b5d0-76fc-453f-911e-ac68246b7c11" containerName="registry-server" containerID="cri-o://4501ce9da15b842883419622bff4b06d9cfaad9dfe8921ab35d2ae472aae0ad4" gracePeriod=2 Oct 02 02:00:43 crc kubenswrapper[4885]: I1002 02:00:43.154063 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5sx8b" Oct 02 02:00:43 crc kubenswrapper[4885]: I1002 02:00:43.273605 4885 generic.go:334] "Generic (PLEG): container finished" podID="4488b5d0-76fc-453f-911e-ac68246b7c11" containerID="4501ce9da15b842883419622bff4b06d9cfaad9dfe8921ab35d2ae472aae0ad4" exitCode=0 Oct 02 02:00:43 crc kubenswrapper[4885]: I1002 02:00:43.273641 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5sx8b" event={"ID":"4488b5d0-76fc-453f-911e-ac68246b7c11","Type":"ContainerDied","Data":"4501ce9da15b842883419622bff4b06d9cfaad9dfe8921ab35d2ae472aae0ad4"} Oct 02 02:00:43 crc kubenswrapper[4885]: I1002 02:00:43.273663 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5sx8b" event={"ID":"4488b5d0-76fc-453f-911e-ac68246b7c11","Type":"ContainerDied","Data":"82a9c3434dcdde1a7ab3855bbd739926655d7c7b095838314bbc5fd0ec677ef3"} Oct 02 02:00:43 crc kubenswrapper[4885]: I1002 02:00:43.273679 4885 scope.go:117] "RemoveContainer" containerID="4501ce9da15b842883419622bff4b06d9cfaad9dfe8921ab35d2ae472aae0ad4" Oct 02 02:00:43 crc kubenswrapper[4885]: I1002 02:00:43.273677 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5sx8b" Oct 02 02:00:43 crc kubenswrapper[4885]: I1002 02:00:43.291398 4885 scope.go:117] "RemoveContainer" containerID="9d2cec7e22005559f0203f5aa0cb20614e243a84dc5ab19b6bc04239b5751fc4" Oct 02 02:00:43 crc kubenswrapper[4885]: I1002 02:00:43.311931 4885 scope.go:117] "RemoveContainer" containerID="ebd2c1bfeaca97477ee5221df53322c296cfa7c2d63023869c12d9a49669c1b1" Oct 02 02:00:43 crc kubenswrapper[4885]: I1002 02:00:43.315441 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4488b5d0-76fc-453f-911e-ac68246b7c11-utilities\") pod \"4488b5d0-76fc-453f-911e-ac68246b7c11\" (UID: \"4488b5d0-76fc-453f-911e-ac68246b7c11\") " Oct 02 02:00:43 crc kubenswrapper[4885]: I1002 02:00:43.315487 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p5xd4\" (UniqueName: \"kubernetes.io/projected/4488b5d0-76fc-453f-911e-ac68246b7c11-kube-api-access-p5xd4\") pod \"4488b5d0-76fc-453f-911e-ac68246b7c11\" (UID: \"4488b5d0-76fc-453f-911e-ac68246b7c11\") " Oct 02 02:00:43 crc kubenswrapper[4885]: I1002 02:00:43.315609 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4488b5d0-76fc-453f-911e-ac68246b7c11-catalog-content\") pod \"4488b5d0-76fc-453f-911e-ac68246b7c11\" (UID: \"4488b5d0-76fc-453f-911e-ac68246b7c11\") " Oct 02 02:00:43 crc kubenswrapper[4885]: I1002 02:00:43.316677 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4488b5d0-76fc-453f-911e-ac68246b7c11-utilities" (OuterVolumeSpecName: "utilities") pod "4488b5d0-76fc-453f-911e-ac68246b7c11" (UID: "4488b5d0-76fc-453f-911e-ac68246b7c11"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:00:43 crc kubenswrapper[4885]: I1002 02:00:43.324480 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4488b5d0-76fc-453f-911e-ac68246b7c11-kube-api-access-p5xd4" (OuterVolumeSpecName: "kube-api-access-p5xd4") pod "4488b5d0-76fc-453f-911e-ac68246b7c11" (UID: "4488b5d0-76fc-453f-911e-ac68246b7c11"). InnerVolumeSpecName "kube-api-access-p5xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:00:43 crc kubenswrapper[4885]: I1002 02:00:43.340296 4885 scope.go:117] "RemoveContainer" containerID="4501ce9da15b842883419622bff4b06d9cfaad9dfe8921ab35d2ae472aae0ad4" Oct 02 02:00:43 crc kubenswrapper[4885]: E1002 02:00:43.340719 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4501ce9da15b842883419622bff4b06d9cfaad9dfe8921ab35d2ae472aae0ad4\": container with ID starting with 4501ce9da15b842883419622bff4b06d9cfaad9dfe8921ab35d2ae472aae0ad4 not found: ID does not exist" containerID="4501ce9da15b842883419622bff4b06d9cfaad9dfe8921ab35d2ae472aae0ad4" Oct 02 02:00:43 crc kubenswrapper[4885]: I1002 02:00:43.340757 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4501ce9da15b842883419622bff4b06d9cfaad9dfe8921ab35d2ae472aae0ad4"} err="failed to get container status \"4501ce9da15b842883419622bff4b06d9cfaad9dfe8921ab35d2ae472aae0ad4\": rpc error: code = NotFound desc = could not find container \"4501ce9da15b842883419622bff4b06d9cfaad9dfe8921ab35d2ae472aae0ad4\": container with ID starting with 4501ce9da15b842883419622bff4b06d9cfaad9dfe8921ab35d2ae472aae0ad4 not found: ID does not exist" Oct 02 02:00:43 crc kubenswrapper[4885]: I1002 02:00:43.340783 4885 scope.go:117] "RemoveContainer" containerID="9d2cec7e22005559f0203f5aa0cb20614e243a84dc5ab19b6bc04239b5751fc4" Oct 02 02:00:43 crc kubenswrapper[4885]: E1002 02:00:43.341087 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9d2cec7e22005559f0203f5aa0cb20614e243a84dc5ab19b6bc04239b5751fc4\": container with ID starting with 9d2cec7e22005559f0203f5aa0cb20614e243a84dc5ab19b6bc04239b5751fc4 not found: ID does not exist" containerID="9d2cec7e22005559f0203f5aa0cb20614e243a84dc5ab19b6bc04239b5751fc4" Oct 02 02:00:43 crc kubenswrapper[4885]: I1002 02:00:43.341119 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d2cec7e22005559f0203f5aa0cb20614e243a84dc5ab19b6bc04239b5751fc4"} err="failed to get container status \"9d2cec7e22005559f0203f5aa0cb20614e243a84dc5ab19b6bc04239b5751fc4\": rpc error: code = NotFound desc = could not find container \"9d2cec7e22005559f0203f5aa0cb20614e243a84dc5ab19b6bc04239b5751fc4\": container with ID starting with 9d2cec7e22005559f0203f5aa0cb20614e243a84dc5ab19b6bc04239b5751fc4 not found: ID does not exist" Oct 02 02:00:43 crc kubenswrapper[4885]: I1002 02:00:43.341140 4885 scope.go:117] "RemoveContainer" containerID="ebd2c1bfeaca97477ee5221df53322c296cfa7c2d63023869c12d9a49669c1b1" Oct 02 02:00:43 crc kubenswrapper[4885]: E1002 02:00:43.341452 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ebd2c1bfeaca97477ee5221df53322c296cfa7c2d63023869c12d9a49669c1b1\": container with ID starting with ebd2c1bfeaca97477ee5221df53322c296cfa7c2d63023869c12d9a49669c1b1 not found: ID does not exist" containerID="ebd2c1bfeaca97477ee5221df53322c296cfa7c2d63023869c12d9a49669c1b1" Oct 02 02:00:43 crc kubenswrapper[4885]: I1002 02:00:43.341499 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ebd2c1bfeaca97477ee5221df53322c296cfa7c2d63023869c12d9a49669c1b1"} err="failed to get container status \"ebd2c1bfeaca97477ee5221df53322c296cfa7c2d63023869c12d9a49669c1b1\": rpc error: code = NotFound desc = could not find container \"ebd2c1bfeaca97477ee5221df53322c296cfa7c2d63023869c12d9a49669c1b1\": container with ID starting with ebd2c1bfeaca97477ee5221df53322c296cfa7c2d63023869c12d9a49669c1b1 not found: ID does not exist" Oct 02 02:00:43 crc kubenswrapper[4885]: I1002 02:00:43.364804 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4488b5d0-76fc-453f-911e-ac68246b7c11-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4488b5d0-76fc-453f-911e-ac68246b7c11" (UID: "4488b5d0-76fc-453f-911e-ac68246b7c11"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:00:43 crc kubenswrapper[4885]: I1002 02:00:43.416811 4885 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4488b5d0-76fc-453f-911e-ac68246b7c11-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:43 crc kubenswrapper[4885]: I1002 02:00:43.417009 4885 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4488b5d0-76fc-453f-911e-ac68246b7c11-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:43 crc kubenswrapper[4885]: I1002 02:00:43.417065 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p5xd4\" (UniqueName: \"kubernetes.io/projected/4488b5d0-76fc-453f-911e-ac68246b7c11-kube-api-access-p5xd4\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:43 crc kubenswrapper[4885]: I1002 02:00:43.599702 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5sx8b"] Oct 02 02:00:43 crc kubenswrapper[4885]: I1002 02:00:43.606493 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-5sx8b"] Oct 02 02:00:44 crc kubenswrapper[4885]: I1002 02:00:44.053750 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4488b5d0-76fc-453f-911e-ac68246b7c11" path="/var/lib/kubelet/pods/4488b5d0-76fc-453f-911e-ac68246b7c11/volumes" Oct 02 02:00:52 crc kubenswrapper[4885]: I1002 02:00:52.655360 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-644bddb6d8-kw92z"] Oct 02 02:00:52 crc kubenswrapper[4885]: E1002 02:00:52.656657 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4488b5d0-76fc-453f-911e-ac68246b7c11" containerName="extract-content" Oct 02 02:00:52 crc kubenswrapper[4885]: I1002 02:00:52.656683 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="4488b5d0-76fc-453f-911e-ac68246b7c11" containerName="extract-content" Oct 02 02:00:52 crc kubenswrapper[4885]: E1002 02:00:52.656704 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5b45f09-e4c2-465b-9ba1-6d9375e5b246" containerName="extract-content" Oct 02 02:00:52 crc kubenswrapper[4885]: I1002 02:00:52.656718 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5b45f09-e4c2-465b-9ba1-6d9375e5b246" containerName="extract-content" Oct 02 02:00:52 crc kubenswrapper[4885]: E1002 02:00:52.656745 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="608b9eff-222e-4bbf-a865-5f38a2f2804c" containerName="registry-server" Oct 02 02:00:52 crc kubenswrapper[4885]: I1002 02:00:52.656759 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="608b9eff-222e-4bbf-a865-5f38a2f2804c" containerName="registry-server" Oct 02 02:00:52 crc kubenswrapper[4885]: E1002 02:00:52.656775 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4488b5d0-76fc-453f-911e-ac68246b7c11" containerName="extract-utilities" Oct 02 02:00:52 crc kubenswrapper[4885]: I1002 02:00:52.656789 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="4488b5d0-76fc-453f-911e-ac68246b7c11" containerName="extract-utilities" Oct 02 02:00:52 crc kubenswrapper[4885]: E1002 02:00:52.656820 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5b45f09-e4c2-465b-9ba1-6d9375e5b246" containerName="extract-utilities" Oct 02 02:00:52 crc kubenswrapper[4885]: I1002 02:00:52.656834 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5b45f09-e4c2-465b-9ba1-6d9375e5b246" containerName="extract-utilities" Oct 02 02:00:52 crc kubenswrapper[4885]: E1002 02:00:52.656855 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5b45f09-e4c2-465b-9ba1-6d9375e5b246" containerName="registry-server" Oct 02 02:00:52 crc kubenswrapper[4885]: I1002 02:00:52.656871 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5b45f09-e4c2-465b-9ba1-6d9375e5b246" containerName="registry-server" Oct 02 02:00:52 crc kubenswrapper[4885]: E1002 02:00:52.656897 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="608b9eff-222e-4bbf-a865-5f38a2f2804c" containerName="extract-content" Oct 02 02:00:52 crc kubenswrapper[4885]: I1002 02:00:52.656910 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="608b9eff-222e-4bbf-a865-5f38a2f2804c" containerName="extract-content" Oct 02 02:00:52 crc kubenswrapper[4885]: E1002 02:00:52.656931 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="608b9eff-222e-4bbf-a865-5f38a2f2804c" containerName="extract-utilities" Oct 02 02:00:52 crc kubenswrapper[4885]: I1002 02:00:52.656945 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="608b9eff-222e-4bbf-a865-5f38a2f2804c" containerName="extract-utilities" Oct 02 02:00:52 crc kubenswrapper[4885]: E1002 02:00:52.656960 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4488b5d0-76fc-453f-911e-ac68246b7c11" containerName="registry-server" Oct 02 02:00:52 crc kubenswrapper[4885]: I1002 02:00:52.656974 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="4488b5d0-76fc-453f-911e-ac68246b7c11" containerName="registry-server" Oct 02 02:00:52 crc kubenswrapper[4885]: I1002 02:00:52.657250 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="608b9eff-222e-4bbf-a865-5f38a2f2804c" containerName="registry-server" Oct 02 02:00:52 crc kubenswrapper[4885]: I1002 02:00:52.657326 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5b45f09-e4c2-465b-9ba1-6d9375e5b246" containerName="registry-server" Oct 02 02:00:52 crc kubenswrapper[4885]: I1002 02:00:52.657357 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="4488b5d0-76fc-453f-911e-ac68246b7c11" containerName="registry-server" Oct 02 02:00:52 crc kubenswrapper[4885]: I1002 02:00:52.658696 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-kw92z" Oct 02 02:00:52 crc kubenswrapper[4885]: I1002 02:00:52.659291 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-6ff8b75857-wnxfk"] Oct 02 02:00:52 crc kubenswrapper[4885]: I1002 02:00:52.660181 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-wnxfk" Oct 02 02:00:52 crc kubenswrapper[4885]: I1002 02:00:52.662046 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-jgb2t" Oct 02 02:00:52 crc kubenswrapper[4885]: I1002 02:00:52.662097 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-tw6t6" Oct 02 02:00:52 crc kubenswrapper[4885]: I1002 02:00:52.670875 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-84f4f7b77b-mw6h7"] Oct 02 02:00:52 crc kubenswrapper[4885]: I1002 02:00:52.671955 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-mw6h7" Oct 02 02:00:52 crc kubenswrapper[4885]: I1002 02:00:52.679114 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-644bddb6d8-kw92z"] Oct 02 02:00:52 crc kubenswrapper[4885]: I1002 02:00:52.679447 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-rf2wq" Oct 02 02:00:52 crc kubenswrapper[4885]: I1002 02:00:52.739513 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-84f4f7b77b-mw6h7"] Oct 02 02:00:52 crc kubenswrapper[4885]: I1002 02:00:52.743876 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-6ff8b75857-wnxfk"] Oct 02 02:00:52 crc kubenswrapper[4885]: I1002 02:00:52.758335 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-84958c4d49-2dsv9"] Oct 02 02:00:52 crc kubenswrapper[4885]: I1002 02:00:52.759449 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-2dsv9" Oct 02 02:00:52 crc kubenswrapper[4885]: I1002 02:00:52.762783 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-t7ppr" Oct 02 02:00:52 crc kubenswrapper[4885]: I1002 02:00:52.765094 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-5d889d78cf-862d6"] Oct 02 02:00:52 crc kubenswrapper[4885]: I1002 02:00:52.766065 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-862d6" Oct 02 02:00:52 crc kubenswrapper[4885]: I1002 02:00:52.767610 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-z9c8l" Oct 02 02:00:52 crc kubenswrapper[4885]: I1002 02:00:52.771525 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-25r4q\" (UniqueName: \"kubernetes.io/projected/e5874982-d78d-4b7c-bf0a-19a7572dddad-kube-api-access-25r4q\") pod \"cinder-operator-controller-manager-644bddb6d8-kw92z\" (UID: \"e5874982-d78d-4b7c-bf0a-19a7572dddad\") " pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-kw92z" Oct 02 02:00:52 crc kubenswrapper[4885]: I1002 02:00:52.771575 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2lb7w\" (UniqueName: \"kubernetes.io/projected/cc08d74a-adca-4ae7-98c6-16c7037bc0ca-kube-api-access-2lb7w\") pod \"designate-operator-controller-manager-84f4f7b77b-mw6h7\" (UID: \"cc08d74a-adca-4ae7-98c6-16c7037bc0ca\") " pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-mw6h7" Oct 02 02:00:52 crc kubenswrapper[4885]: I1002 02:00:52.771598 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9rv2n\" (UniqueName: \"kubernetes.io/projected/8113ab96-dde5-4f9f-b922-687e7a17afa7-kube-api-access-9rv2n\") pod \"barbican-operator-controller-manager-6ff8b75857-wnxfk\" (UID: \"8113ab96-dde5-4f9f-b922-687e7a17afa7\") " pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-wnxfk" Oct 02 02:00:52 crc kubenswrapper[4885]: I1002 02:00:52.779353 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-84958c4d49-2dsv9"] Oct 02 02:00:52 crc kubenswrapper[4885]: I1002 02:00:52.790878 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-9f4696d94-pv94t"] Oct 02 02:00:52 crc kubenswrapper[4885]: I1002 02:00:52.791994 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-pv94t" Oct 02 02:00:52 crc kubenswrapper[4885]: I1002 02:00:52.796168 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-xjnvf" Oct 02 02:00:52 crc kubenswrapper[4885]: I1002 02:00:52.806321 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5d889d78cf-862d6"] Oct 02 02:00:52 crc kubenswrapper[4885]: I1002 02:00:52.819335 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-9d6c5db85-9fjl5"] Oct 02 02:00:52 crc kubenswrapper[4885]: I1002 02:00:52.820178 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-9fjl5" Oct 02 02:00:52 crc kubenswrapper[4885]: I1002 02:00:52.827503 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5cd4858477-mjtzp"] Oct 02 02:00:52 crc kubenswrapper[4885]: I1002 02:00:52.827617 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Oct 02 02:00:52 crc kubenswrapper[4885]: I1002 02:00:52.827732 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-s9ddd" Oct 02 02:00:52 crc kubenswrapper[4885]: I1002 02:00:52.828438 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-5cd4858477-mjtzp" Oct 02 02:00:52 crc kubenswrapper[4885]: I1002 02:00:52.844017 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-bwjgd" Oct 02 02:00:52 crc kubenswrapper[4885]: I1002 02:00:52.873238 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-5bd55b4bff-pww46"] Oct 02 02:00:52 crc kubenswrapper[4885]: I1002 02:00:52.874474 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-pww46" Oct 02 02:00:52 crc kubenswrapper[4885]: I1002 02:00:52.877582 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-b24np" Oct 02 02:00:52 crc kubenswrapper[4885]: I1002 02:00:52.878142 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gm5kf\" (UniqueName: \"kubernetes.io/projected/1e64a13b-65af-4460-80a8-4a4659edd8aa-kube-api-access-gm5kf\") pod \"horizon-operator-controller-manager-9f4696d94-pv94t\" (UID: \"1e64a13b-65af-4460-80a8-4a4659edd8aa\") " pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-pv94t" Oct 02 02:00:52 crc kubenswrapper[4885]: I1002 02:00:52.878188 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7fqlj\" (UniqueName: \"kubernetes.io/projected/45e282ed-8195-40dc-86cf-1177f3ebeb65-kube-api-access-7fqlj\") pod \"infra-operator-controller-manager-9d6c5db85-9fjl5\" (UID: \"45e282ed-8195-40dc-86cf-1177f3ebeb65\") " pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-9fjl5" Oct 02 02:00:52 crc kubenswrapper[4885]: I1002 02:00:52.878211 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nxcjp\" (UniqueName: \"kubernetes.io/projected/78117b64-3b3a-4892-90de-0dea12fa2602-kube-api-access-nxcjp\") pod \"heat-operator-controller-manager-5d889d78cf-862d6\" (UID: \"78117b64-3b3a-4892-90de-0dea12fa2602\") " pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-862d6" Oct 02 02:00:52 crc kubenswrapper[4885]: I1002 02:00:52.878242 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-25r4q\" (UniqueName: \"kubernetes.io/projected/e5874982-d78d-4b7c-bf0a-19a7572dddad-kube-api-access-25r4q\") pod \"cinder-operator-controller-manager-644bddb6d8-kw92z\" (UID: \"e5874982-d78d-4b7c-bf0a-19a7572dddad\") " pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-kw92z" Oct 02 02:00:52 crc kubenswrapper[4885]: I1002 02:00:52.878298 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5zkvr\" (UniqueName: \"kubernetes.io/projected/bbd60151-4081-46a8-85c0-0783604392a1-kube-api-access-5zkvr\") pod \"ironic-operator-controller-manager-5cd4858477-mjtzp\" (UID: \"bbd60151-4081-46a8-85c0-0783604392a1\") " pod="openstack-operators/ironic-operator-controller-manager-5cd4858477-mjtzp" Oct 02 02:00:52 crc kubenswrapper[4885]: I1002 02:00:52.878322 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2lb7w\" (UniqueName: \"kubernetes.io/projected/cc08d74a-adca-4ae7-98c6-16c7037bc0ca-kube-api-access-2lb7w\") pod \"designate-operator-controller-manager-84f4f7b77b-mw6h7\" (UID: \"cc08d74a-adca-4ae7-98c6-16c7037bc0ca\") " pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-mw6h7" Oct 02 02:00:52 crc kubenswrapper[4885]: I1002 02:00:52.878343 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9rv2n\" (UniqueName: \"kubernetes.io/projected/8113ab96-dde5-4f9f-b922-687e7a17afa7-kube-api-access-9rv2n\") pod \"barbican-operator-controller-manager-6ff8b75857-wnxfk\" (UID: \"8113ab96-dde5-4f9f-b922-687e7a17afa7\") " pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-wnxfk" Oct 02 02:00:52 crc kubenswrapper[4885]: I1002 02:00:52.878362 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rrm9q\" (UniqueName: \"kubernetes.io/projected/ff10db28-28ca-4f65-8c24-2ed813d7cc5c-kube-api-access-rrm9q\") pod \"glance-operator-controller-manager-84958c4d49-2dsv9\" (UID: \"ff10db28-28ca-4f65-8c24-2ed813d7cc5c\") " pod="openstack-operators/glance-operator-controller-manager-84958c4d49-2dsv9" Oct 02 02:00:52 crc kubenswrapper[4885]: I1002 02:00:52.878404 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/45e282ed-8195-40dc-86cf-1177f3ebeb65-cert\") pod \"infra-operator-controller-manager-9d6c5db85-9fjl5\" (UID: \"45e282ed-8195-40dc-86cf-1177f3ebeb65\") " pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-9fjl5" Oct 02 02:00:52 crc kubenswrapper[4885]: I1002 02:00:52.908191 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-6d68dbc695-9zf8p"] Oct 02 02:00:52 crc kubenswrapper[4885]: I1002 02:00:52.918676 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-25r4q\" (UniqueName: \"kubernetes.io/projected/e5874982-d78d-4b7c-bf0a-19a7572dddad-kube-api-access-25r4q\") pod \"cinder-operator-controller-manager-644bddb6d8-kw92z\" (UID: \"e5874982-d78d-4b7c-bf0a-19a7572dddad\") " pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-kw92z" Oct 02 02:00:52 crc kubenswrapper[4885]: I1002 02:00:52.923485 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9rv2n\" (UniqueName: \"kubernetes.io/projected/8113ab96-dde5-4f9f-b922-687e7a17afa7-kube-api-access-9rv2n\") pod \"barbican-operator-controller-manager-6ff8b75857-wnxfk\" (UID: \"8113ab96-dde5-4f9f-b922-687e7a17afa7\") " pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-wnxfk" Oct 02 02:00:52 crc kubenswrapper[4885]: I1002 02:00:52.924977 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2lb7w\" (UniqueName: \"kubernetes.io/projected/cc08d74a-adca-4ae7-98c6-16c7037bc0ca-kube-api-access-2lb7w\") pod \"designate-operator-controller-manager-84f4f7b77b-mw6h7\" (UID: \"cc08d74a-adca-4ae7-98c6-16c7037bc0ca\") " pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-mw6h7" Oct 02 02:00:52 crc kubenswrapper[4885]: I1002 02:00:52.945023 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-9zf8p" Oct 02 02:00:52 crc kubenswrapper[4885]: I1002 02:00:52.950388 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5cd4858477-mjtzp"] Oct 02 02:00:52 crc kubenswrapper[4885]: I1002 02:00:52.956901 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-nvz8x" Oct 02 02:00:52 crc kubenswrapper[4885]: I1002 02:00:52.957055 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-88c7-nz599"] Oct 02 02:00:52 crc kubenswrapper[4885]: I1002 02:00:52.958119 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-88c7-nz599" Oct 02 02:00:52 crc kubenswrapper[4885]: I1002 02:00:52.962908 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-5bd55b4bff-pww46"] Oct 02 02:00:52 crc kubenswrapper[4885]: I1002 02:00:52.969214 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-jjm7n" Oct 02 02:00:52 crc kubenswrapper[4885]: I1002 02:00:52.975360 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-9d6c5db85-9fjl5"] Oct 02 02:00:52 crc kubenswrapper[4885]: I1002 02:00:52.980562 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rrm9q\" (UniqueName: \"kubernetes.io/projected/ff10db28-28ca-4f65-8c24-2ed813d7cc5c-kube-api-access-rrm9q\") pod \"glance-operator-controller-manager-84958c4d49-2dsv9\" (UID: \"ff10db28-28ca-4f65-8c24-2ed813d7cc5c\") " pod="openstack-operators/glance-operator-controller-manager-84958c4d49-2dsv9" Oct 02 02:00:52 crc kubenswrapper[4885]: I1002 02:00:52.980684 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dcmdc\" (UniqueName: \"kubernetes.io/projected/bcb6c44c-c620-465c-8db3-6d770deda522-kube-api-access-dcmdc\") pod \"manila-operator-controller-manager-6d68dbc695-9zf8p\" (UID: \"bcb6c44c-c620-465c-8db3-6d770deda522\") " pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-9zf8p" Oct 02 02:00:52 crc kubenswrapper[4885]: I1002 02:00:52.980715 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lmgrw\" (UniqueName: \"kubernetes.io/projected/d2263844-1435-4352-b773-04e8d10f35a0-kube-api-access-lmgrw\") pod \"keystone-operator-controller-manager-5bd55b4bff-pww46\" (UID: \"d2263844-1435-4352-b773-04e8d10f35a0\") " pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-pww46" Oct 02 02:00:52 crc kubenswrapper[4885]: I1002 02:00:52.980739 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/45e282ed-8195-40dc-86cf-1177f3ebeb65-cert\") pod \"infra-operator-controller-manager-9d6c5db85-9fjl5\" (UID: \"45e282ed-8195-40dc-86cf-1177f3ebeb65\") " pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-9fjl5" Oct 02 02:00:52 crc kubenswrapper[4885]: I1002 02:00:52.980786 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gm5kf\" (UniqueName: \"kubernetes.io/projected/1e64a13b-65af-4460-80a8-4a4659edd8aa-kube-api-access-gm5kf\") pod \"horizon-operator-controller-manager-9f4696d94-pv94t\" (UID: \"1e64a13b-65af-4460-80a8-4a4659edd8aa\") " pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-pv94t" Oct 02 02:00:52 crc kubenswrapper[4885]: I1002 02:00:52.980809 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7fqlj\" (UniqueName: \"kubernetes.io/projected/45e282ed-8195-40dc-86cf-1177f3ebeb65-kube-api-access-7fqlj\") pod \"infra-operator-controller-manager-9d6c5db85-9fjl5\" (UID: \"45e282ed-8195-40dc-86cf-1177f3ebeb65\") " pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-9fjl5" Oct 02 02:00:52 crc kubenswrapper[4885]: I1002 02:00:52.980830 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nxcjp\" (UniqueName: \"kubernetes.io/projected/78117b64-3b3a-4892-90de-0dea12fa2602-kube-api-access-nxcjp\") pod \"heat-operator-controller-manager-5d889d78cf-862d6\" (UID: \"78117b64-3b3a-4892-90de-0dea12fa2602\") " pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-862d6" Oct 02 02:00:52 crc kubenswrapper[4885]: E1002 02:00:52.980836 4885 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Oct 02 02:00:52 crc kubenswrapper[4885]: I1002 02:00:52.980856 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5zkvr\" (UniqueName: \"kubernetes.io/projected/bbd60151-4081-46a8-85c0-0783604392a1-kube-api-access-5zkvr\") pod \"ironic-operator-controller-manager-5cd4858477-mjtzp\" (UID: \"bbd60151-4081-46a8-85c0-0783604392a1\") " pod="openstack-operators/ironic-operator-controller-manager-5cd4858477-mjtzp" Oct 02 02:00:52 crc kubenswrapper[4885]: E1002 02:00:52.980892 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/45e282ed-8195-40dc-86cf-1177f3ebeb65-cert podName:45e282ed-8195-40dc-86cf-1177f3ebeb65 nodeName:}" failed. No retries permitted until 2025-10-02 02:00:53.480873572 +0000 UTC m=+842.292620961 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/45e282ed-8195-40dc-86cf-1177f3ebeb65-cert") pod "infra-operator-controller-manager-9d6c5db85-9fjl5" (UID: "45e282ed-8195-40dc-86cf-1177f3ebeb65") : secret "infra-operator-webhook-server-cert" not found Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.000001 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-849d5b9b84-clsdk"] Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.001084 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-clsdk" Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.006595 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nxcjp\" (UniqueName: \"kubernetes.io/projected/78117b64-3b3a-4892-90de-0dea12fa2602-kube-api-access-nxcjp\") pod \"heat-operator-controller-manager-5d889d78cf-862d6\" (UID: \"78117b64-3b3a-4892-90de-0dea12fa2602\") " pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-862d6" Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.007042 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rrm9q\" (UniqueName: \"kubernetes.io/projected/ff10db28-28ca-4f65-8c24-2ed813d7cc5c-kube-api-access-rrm9q\") pod \"glance-operator-controller-manager-84958c4d49-2dsv9\" (UID: \"ff10db28-28ca-4f65-8c24-2ed813d7cc5c\") " pod="openstack-operators/glance-operator-controller-manager-84958c4d49-2dsv9" Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.008578 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-mnpk6" Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.012150 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gm5kf\" (UniqueName: \"kubernetes.io/projected/1e64a13b-65af-4460-80a8-4a4659edd8aa-kube-api-access-gm5kf\") pod \"horizon-operator-controller-manager-9f4696d94-pv94t\" (UID: \"1e64a13b-65af-4460-80a8-4a4659edd8aa\") " pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-pv94t" Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.014443 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-9f4696d94-pv94t"] Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.014943 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7fqlj\" (UniqueName: \"kubernetes.io/projected/45e282ed-8195-40dc-86cf-1177f3ebeb65-kube-api-access-7fqlj\") pod \"infra-operator-controller-manager-9d6c5db85-9fjl5\" (UID: \"45e282ed-8195-40dc-86cf-1177f3ebeb65\") " pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-9fjl5" Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.029544 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5zkvr\" (UniqueName: \"kubernetes.io/projected/bbd60151-4081-46a8-85c0-0783604392a1-kube-api-access-5zkvr\") pod \"ironic-operator-controller-manager-5cd4858477-mjtzp\" (UID: \"bbd60151-4081-46a8-85c0-0783604392a1\") " pod="openstack-operators/ironic-operator-controller-manager-5cd4858477-mjtzp" Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.037180 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-64cd67b5cb-hwrj4"] Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.038105 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-hwrj4" Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.038250 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-kw92z" Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.039429 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-qmfqg" Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.040721 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-7b787867f4-kcpkv"] Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.041611 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-kcpkv" Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.043085 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-qnwsb" Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.045111 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-849d5b9b84-clsdk"] Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.048775 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-6d68dbc695-9zf8p"] Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.049055 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-wnxfk" Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.057503 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-mw6h7" Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.063944 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-88c7-nz599"] Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.073693 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-2dsv9" Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.079937 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-7b787867f4-kcpkv"] Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.081759 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9vxfb\" (UniqueName: \"kubernetes.io/projected/e6673002-4d33-4c32-8eb4-f1727339b76d-kube-api-access-9vxfb\") pod \"nova-operator-controller-manager-64cd67b5cb-hwrj4\" (UID: \"e6673002-4d33-4c32-8eb4-f1727339b76d\") " pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-hwrj4" Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.081803 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-642bx\" (UniqueName: \"kubernetes.io/projected/ae73c7b6-6b7e-4ca9-8a4e-0694ab34d9a9-kube-api-access-642bx\") pod \"neutron-operator-controller-manager-849d5b9b84-clsdk\" (UID: \"ae73c7b6-6b7e-4ca9-8a4e-0694ab34d9a9\") " pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-clsdk" Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.081945 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cx6kn\" (UniqueName: \"kubernetes.io/projected/9d724768-b5a0-4a55-951a-d350127061cb-kube-api-access-cx6kn\") pod \"octavia-operator-controller-manager-7b787867f4-kcpkv\" (UID: \"9d724768-b5a0-4a55-951a-d350127061cb\") " pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-kcpkv" Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.083514 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dcmdc\" (UniqueName: \"kubernetes.io/projected/bcb6c44c-c620-465c-8db3-6d770deda522-kube-api-access-dcmdc\") pod \"manila-operator-controller-manager-6d68dbc695-9zf8p\" (UID: \"bcb6c44c-c620-465c-8db3-6d770deda522\") " pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-9zf8p" Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.083561 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lmgrw\" (UniqueName: \"kubernetes.io/projected/d2263844-1435-4352-b773-04e8d10f35a0-kube-api-access-lmgrw\") pod \"keystone-operator-controller-manager-5bd55b4bff-pww46\" (UID: \"d2263844-1435-4352-b773-04e8d10f35a0\") " pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-pww46" Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.084689 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w627l\" (UniqueName: \"kubernetes.io/projected/0f1b84a4-59d6-45fe-9c2f-0bdff82fde6e-kube-api-access-w627l\") pod \"mariadb-operator-controller-manager-88c7-nz599\" (UID: \"0f1b84a4-59d6-45fe-9c2f-0bdff82fde6e\") " pod="openstack-operators/mariadb-operator-controller-manager-88c7-nz599" Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.087676 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-64cd67b5cb-hwrj4"] Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.087895 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-862d6" Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.096340 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-9976ff44c-gl72q"] Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.097658 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-gl72q" Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.099057 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-phjhs" Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.100026 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-wgfjr"] Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.100985 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-wgfjr" Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.102469 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.103166 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dcmdc\" (UniqueName: \"kubernetes.io/projected/bcb6c44c-c620-465c-8db3-6d770deda522-kube-api-access-dcmdc\") pod \"manila-operator-controller-manager-6d68dbc695-9zf8p\" (UID: \"bcb6c44c-c620-465c-8db3-6d770deda522\") " pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-9zf8p" Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.104554 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-9gj95" Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.105216 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lmgrw\" (UniqueName: \"kubernetes.io/projected/d2263844-1435-4352-b773-04e8d10f35a0-kube-api-access-lmgrw\") pod \"keystone-operator-controller-manager-5bd55b4bff-pww46\" (UID: \"d2263844-1435-4352-b773-04e8d10f35a0\") " pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-pww46" Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.105387 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-589c58c6c-r65hz"] Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.107092 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-r65hz" Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.111317 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-fh5xv" Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.120836 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-pv94t" Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.140818 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-9976ff44c-gl72q"] Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.156553 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-wgfjr"] Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.172153 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-84d6b4b759-6f74g"] Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.173215 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-6f74g" Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.175285 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-kpzsm" Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.177485 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-5cd4858477-mjtzp" Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.185653 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cx6kn\" (UniqueName: \"kubernetes.io/projected/9d724768-b5a0-4a55-951a-d350127061cb-kube-api-access-cx6kn\") pod \"octavia-operator-controller-manager-7b787867f4-kcpkv\" (UID: \"9d724768-b5a0-4a55-951a-d350127061cb\") " pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-kcpkv" Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.185700 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-44vfh\" (UniqueName: \"kubernetes.io/projected/6e740063-fb49-4847-ba6d-77b6c089ba50-kube-api-access-44vfh\") pod \"placement-operator-controller-manager-589c58c6c-r65hz\" (UID: \"6e740063-fb49-4847-ba6d-77b6c089ba50\") " pod="openstack-operators/placement-operator-controller-manager-589c58c6c-r65hz" Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.185736 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w627l\" (UniqueName: \"kubernetes.io/projected/0f1b84a4-59d6-45fe-9c2f-0bdff82fde6e-kube-api-access-w627l\") pod \"mariadb-operator-controller-manager-88c7-nz599\" (UID: \"0f1b84a4-59d6-45fe-9c2f-0bdff82fde6e\") " pod="openstack-operators/mariadb-operator-controller-manager-88c7-nz599" Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.187679 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-589c58c6c-r65hz"] Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.187911 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9vxfb\" (UniqueName: \"kubernetes.io/projected/e6673002-4d33-4c32-8eb4-f1727339b76d-kube-api-access-9vxfb\") pod \"nova-operator-controller-manager-64cd67b5cb-hwrj4\" (UID: \"e6673002-4d33-4c32-8eb4-f1727339b76d\") " pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-hwrj4" Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.187960 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-642bx\" (UniqueName: \"kubernetes.io/projected/ae73c7b6-6b7e-4ca9-8a4e-0694ab34d9a9-kube-api-access-642bx\") pod \"neutron-operator-controller-manager-849d5b9b84-clsdk\" (UID: \"ae73c7b6-6b7e-4ca9-8a4e-0694ab34d9a9\") " pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-clsdk" Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.187980 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/dc61bc9b-923b-41cb-b724-62c72caf0586-cert\") pod \"openstack-baremetal-operator-controller-manager-5869cb545-wgfjr\" (UID: \"dc61bc9b-923b-41cb-b724-62c72caf0586\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-wgfjr" Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.188028 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dwm7p\" (UniqueName: \"kubernetes.io/projected/c540bfce-7653-4976-8d6e-36b479c3c711-kube-api-access-dwm7p\") pod \"ovn-operator-controller-manager-9976ff44c-gl72q\" (UID: \"c540bfce-7653-4976-8d6e-36b479c3c711\") " pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-gl72q" Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.188060 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jq7ws\" (UniqueName: \"kubernetes.io/projected/dc61bc9b-923b-41cb-b724-62c72caf0586-kube-api-access-jq7ws\") pod \"openstack-baremetal-operator-controller-manager-5869cb545-wgfjr\" (UID: \"dc61bc9b-923b-41cb-b724-62c72caf0586\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-wgfjr" Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.196838 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-84d6b4b759-6f74g"] Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.208875 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-26g2l"] Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.210952 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-642bx\" (UniqueName: \"kubernetes.io/projected/ae73c7b6-6b7e-4ca9-8a4e-0694ab34d9a9-kube-api-access-642bx\") pod \"neutron-operator-controller-manager-849d5b9b84-clsdk\" (UID: \"ae73c7b6-6b7e-4ca9-8a4e-0694ab34d9a9\") " pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-clsdk" Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.211570 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-26g2l" Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.214116 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-pww46" Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.216663 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-dm969" Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.216813 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w627l\" (UniqueName: \"kubernetes.io/projected/0f1b84a4-59d6-45fe-9c2f-0bdff82fde6e-kube-api-access-w627l\") pod \"mariadb-operator-controller-manager-88c7-nz599\" (UID: \"0f1b84a4-59d6-45fe-9c2f-0bdff82fde6e\") " pod="openstack-operators/mariadb-operator-controller-manager-88c7-nz599" Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.216881 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-26g2l"] Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.217640 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cx6kn\" (UniqueName: \"kubernetes.io/projected/9d724768-b5a0-4a55-951a-d350127061cb-kube-api-access-cx6kn\") pod \"octavia-operator-controller-manager-7b787867f4-kcpkv\" (UID: \"9d724768-b5a0-4a55-951a-d350127061cb\") " pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-kcpkv" Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.219849 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9vxfb\" (UniqueName: \"kubernetes.io/projected/e6673002-4d33-4c32-8eb4-f1727339b76d-kube-api-access-9vxfb\") pod \"nova-operator-controller-manager-64cd67b5cb-hwrj4\" (UID: \"e6673002-4d33-4c32-8eb4-f1727339b76d\") " pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-hwrj4" Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.263070 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-85777745bb-94284"] Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.263941 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-85777745bb-94284" Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.269279 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-r42h6" Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.280409 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-85777745bb-94284"] Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.293052 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-9zf8p" Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.293759 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v7rnh\" (UniqueName: \"kubernetes.io/projected/dd09a3bb-1217-4a3a-822b-2b2f04171271-kube-api-access-v7rnh\") pod \"telemetry-operator-controller-manager-b8d54b5d7-26g2l\" (UID: \"dd09a3bb-1217-4a3a-822b-2b2f04171271\") " pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-26g2l" Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.293815 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-44vfh\" (UniqueName: \"kubernetes.io/projected/6e740063-fb49-4847-ba6d-77b6c089ba50-kube-api-access-44vfh\") pod \"placement-operator-controller-manager-589c58c6c-r65hz\" (UID: \"6e740063-fb49-4847-ba6d-77b6c089ba50\") " pod="openstack-operators/placement-operator-controller-manager-589c58c6c-r65hz" Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.293863 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-klp7r\" (UniqueName: \"kubernetes.io/projected/f80d2e7d-10a3-492f-a045-b48126a02490-kube-api-access-klp7r\") pod \"swift-operator-controller-manager-84d6b4b759-6f74g\" (UID: \"f80d2e7d-10a3-492f-a045-b48126a02490\") " pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-6f74g" Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.293894 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/dc61bc9b-923b-41cb-b724-62c72caf0586-cert\") pod \"openstack-baremetal-operator-controller-manager-5869cb545-wgfjr\" (UID: \"dc61bc9b-923b-41cb-b724-62c72caf0586\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-wgfjr" Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.293934 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dwm7p\" (UniqueName: \"kubernetes.io/projected/c540bfce-7653-4976-8d6e-36b479c3c711-kube-api-access-dwm7p\") pod \"ovn-operator-controller-manager-9976ff44c-gl72q\" (UID: \"c540bfce-7653-4976-8d6e-36b479c3c711\") " pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-gl72q" Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.293960 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jq7ws\" (UniqueName: \"kubernetes.io/projected/dc61bc9b-923b-41cb-b724-62c72caf0586-kube-api-access-jq7ws\") pod \"openstack-baremetal-operator-controller-manager-5869cb545-wgfjr\" (UID: \"dc61bc9b-923b-41cb-b724-62c72caf0586\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-wgfjr" Oct 02 02:00:53 crc kubenswrapper[4885]: E1002 02:00:53.294391 4885 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 02 02:00:53 crc kubenswrapper[4885]: E1002 02:00:53.317579 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/dc61bc9b-923b-41cb-b724-62c72caf0586-cert podName:dc61bc9b-923b-41cb-b724-62c72caf0586 nodeName:}" failed. No retries permitted until 2025-10-02 02:00:53.81752192 +0000 UTC m=+842.629269319 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/dc61bc9b-923b-41cb-b724-62c72caf0586-cert") pod "openstack-baremetal-operator-controller-manager-5869cb545-wgfjr" (UID: "dc61bc9b-923b-41cb-b724-62c72caf0586") : secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.322004 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-88c7-nz599" Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.329780 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dwm7p\" (UniqueName: \"kubernetes.io/projected/c540bfce-7653-4976-8d6e-36b479c3c711-kube-api-access-dwm7p\") pod \"ovn-operator-controller-manager-9976ff44c-gl72q\" (UID: \"c540bfce-7653-4976-8d6e-36b479c3c711\") " pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-gl72q" Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.343921 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-44vfh\" (UniqueName: \"kubernetes.io/projected/6e740063-fb49-4847-ba6d-77b6c089ba50-kube-api-access-44vfh\") pod \"placement-operator-controller-manager-589c58c6c-r65hz\" (UID: \"6e740063-fb49-4847-ba6d-77b6c089ba50\") " pod="openstack-operators/placement-operator-controller-manager-589c58c6c-r65hz" Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.350584 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jq7ws\" (UniqueName: \"kubernetes.io/projected/dc61bc9b-923b-41cb-b724-62c72caf0586-kube-api-access-jq7ws\") pod \"openstack-baremetal-operator-controller-manager-5869cb545-wgfjr\" (UID: \"dc61bc9b-923b-41cb-b724-62c72caf0586\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-wgfjr" Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.359661 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-clsdk" Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.406784 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-kcpkv" Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.408765 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-hwrj4" Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.412025 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v7rnh\" (UniqueName: \"kubernetes.io/projected/dd09a3bb-1217-4a3a-822b-2b2f04171271-kube-api-access-v7rnh\") pod \"telemetry-operator-controller-manager-b8d54b5d7-26g2l\" (UID: \"dd09a3bb-1217-4a3a-822b-2b2f04171271\") " pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-26g2l" Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.412116 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-klp7r\" (UniqueName: \"kubernetes.io/projected/f80d2e7d-10a3-492f-a045-b48126a02490-kube-api-access-klp7r\") pod \"swift-operator-controller-manager-84d6b4b759-6f74g\" (UID: \"f80d2e7d-10a3-492f-a045-b48126a02490\") " pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-6f74g" Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.412182 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d84qx\" (UniqueName: \"kubernetes.io/projected/74a07e3e-5061-43f9-b239-eb480bade999-kube-api-access-d84qx\") pod \"test-operator-controller-manager-85777745bb-94284\" (UID: \"74a07e3e-5061-43f9-b239-eb480bade999\") " pod="openstack-operators/test-operator-controller-manager-85777745bb-94284" Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.416123 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-gl72q" Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.428907 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6b9957f54f-mb6mf"] Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.434705 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-6b9957f54f-mb6mf" Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.437010 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-5xcjz" Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.443890 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v7rnh\" (UniqueName: \"kubernetes.io/projected/dd09a3bb-1217-4a3a-822b-2b2f04171271-kube-api-access-v7rnh\") pod \"telemetry-operator-controller-manager-b8d54b5d7-26g2l\" (UID: \"dd09a3bb-1217-4a3a-822b-2b2f04171271\") " pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-26g2l" Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.445507 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-r65hz" Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.454623 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6b9957f54f-mb6mf"] Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.479776 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-557f5d867b-ndfjx"] Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.480794 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-557f5d867b-ndfjx" Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.483437 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.483638 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-64zff" Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.483836 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-klp7r\" (UniqueName: \"kubernetes.io/projected/f80d2e7d-10a3-492f-a045-b48126a02490-kube-api-access-klp7r\") pod \"swift-operator-controller-manager-84d6b4b759-6f74g\" (UID: \"f80d2e7d-10a3-492f-a045-b48126a02490\") " pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-6f74g" Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.494992 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-557f5d867b-ndfjx"] Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.495277 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-6f74g" Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.539014 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-26g2l" Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.540360 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-bngfl"] Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.541315 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gc8nr\" (UniqueName: \"kubernetes.io/projected/c5dd164b-699d-4f14-85d0-91f138a42c95-kube-api-access-gc8nr\") pod \"watcher-operator-controller-manager-6b9957f54f-mb6mf\" (UID: \"c5dd164b-699d-4f14-85d0-91f138a42c95\") " pod="openstack-operators/watcher-operator-controller-manager-6b9957f54f-mb6mf" Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.541634 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-bngfl" Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.555830 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-2mhd9" Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.558365 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-bngfl"] Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.584371 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/45e282ed-8195-40dc-86cf-1177f3ebeb65-cert\") pod \"infra-operator-controller-manager-9d6c5db85-9fjl5\" (UID: \"45e282ed-8195-40dc-86cf-1177f3ebeb65\") " pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-9fjl5" Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.584475 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d84qx\" (UniqueName: \"kubernetes.io/projected/74a07e3e-5061-43f9-b239-eb480bade999-kube-api-access-d84qx\") pod \"test-operator-controller-manager-85777745bb-94284\" (UID: \"74a07e3e-5061-43f9-b239-eb480bade999\") " pod="openstack-operators/test-operator-controller-manager-85777745bb-94284" Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.610789 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/45e282ed-8195-40dc-86cf-1177f3ebeb65-cert\") pod \"infra-operator-controller-manager-9d6c5db85-9fjl5\" (UID: \"45e282ed-8195-40dc-86cf-1177f3ebeb65\") " pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-9fjl5" Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.625737 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d84qx\" (UniqueName: \"kubernetes.io/projected/74a07e3e-5061-43f9-b239-eb480bade999-kube-api-access-d84qx\") pod \"test-operator-controller-manager-85777745bb-94284\" (UID: \"74a07e3e-5061-43f9-b239-eb480bade999\") " pod="openstack-operators/test-operator-controller-manager-85777745bb-94284" Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.633338 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-85777745bb-94284" Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.683647 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-84f4f7b77b-mw6h7"] Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.688544 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gc8nr\" (UniqueName: \"kubernetes.io/projected/c5dd164b-699d-4f14-85d0-91f138a42c95-kube-api-access-gc8nr\") pod \"watcher-operator-controller-manager-6b9957f54f-mb6mf\" (UID: \"c5dd164b-699d-4f14-85d0-91f138a42c95\") " pod="openstack-operators/watcher-operator-controller-manager-6b9957f54f-mb6mf" Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.688604 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rsjth\" (UniqueName: \"kubernetes.io/projected/d2768afc-17ee-4141-aa68-7366faaa0145-kube-api-access-rsjth\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-bngfl\" (UID: \"d2768afc-17ee-4141-aa68-7366faaa0145\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-bngfl" Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.688625 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8s55t\" (UniqueName: \"kubernetes.io/projected/8b18e6b1-82f2-46cd-99c8-85af816df20f-kube-api-access-8s55t\") pod \"openstack-operator-controller-manager-557f5d867b-ndfjx\" (UID: \"8b18e6b1-82f2-46cd-99c8-85af816df20f\") " pod="openstack-operators/openstack-operator-controller-manager-557f5d867b-ndfjx" Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.688640 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8b18e6b1-82f2-46cd-99c8-85af816df20f-cert\") pod \"openstack-operator-controller-manager-557f5d867b-ndfjx\" (UID: \"8b18e6b1-82f2-46cd-99c8-85af816df20f\") " pod="openstack-operators/openstack-operator-controller-manager-557f5d867b-ndfjx" Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.694818 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-6ff8b75857-wnxfk"] Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.720359 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gc8nr\" (UniqueName: \"kubernetes.io/projected/c5dd164b-699d-4f14-85d0-91f138a42c95-kube-api-access-gc8nr\") pod \"watcher-operator-controller-manager-6b9957f54f-mb6mf\" (UID: \"c5dd164b-699d-4f14-85d0-91f138a42c95\") " pod="openstack-operators/watcher-operator-controller-manager-6b9957f54f-mb6mf" Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.764823 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-9fjl5" Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.789416 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-6b9957f54f-mb6mf" Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.789890 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rsjth\" (UniqueName: \"kubernetes.io/projected/d2768afc-17ee-4141-aa68-7366faaa0145-kube-api-access-rsjth\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-bngfl\" (UID: \"d2768afc-17ee-4141-aa68-7366faaa0145\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-bngfl" Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.789925 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8s55t\" (UniqueName: \"kubernetes.io/projected/8b18e6b1-82f2-46cd-99c8-85af816df20f-kube-api-access-8s55t\") pod \"openstack-operator-controller-manager-557f5d867b-ndfjx\" (UID: \"8b18e6b1-82f2-46cd-99c8-85af816df20f\") " pod="openstack-operators/openstack-operator-controller-manager-557f5d867b-ndfjx" Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.789951 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8b18e6b1-82f2-46cd-99c8-85af816df20f-cert\") pod \"openstack-operator-controller-manager-557f5d867b-ndfjx\" (UID: \"8b18e6b1-82f2-46cd-99c8-85af816df20f\") " pod="openstack-operators/openstack-operator-controller-manager-557f5d867b-ndfjx" Oct 02 02:00:53 crc kubenswrapper[4885]: E1002 02:00:53.790118 4885 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Oct 02 02:00:53 crc kubenswrapper[4885]: E1002 02:00:53.790171 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8b18e6b1-82f2-46cd-99c8-85af816df20f-cert podName:8b18e6b1-82f2-46cd-99c8-85af816df20f nodeName:}" failed. No retries permitted until 2025-10-02 02:00:54.290152029 +0000 UTC m=+843.101899428 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/8b18e6b1-82f2-46cd-99c8-85af816df20f-cert") pod "openstack-operator-controller-manager-557f5d867b-ndfjx" (UID: "8b18e6b1-82f2-46cd-99c8-85af816df20f") : secret "webhook-server-cert" not found Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.814520 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rsjth\" (UniqueName: \"kubernetes.io/projected/d2768afc-17ee-4141-aa68-7366faaa0145-kube-api-access-rsjth\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-bngfl\" (UID: \"d2768afc-17ee-4141-aa68-7366faaa0145\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-bngfl" Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.818144 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8s55t\" (UniqueName: \"kubernetes.io/projected/8b18e6b1-82f2-46cd-99c8-85af816df20f-kube-api-access-8s55t\") pod \"openstack-operator-controller-manager-557f5d867b-ndfjx\" (UID: \"8b18e6b1-82f2-46cd-99c8-85af816df20f\") " pod="openstack-operators/openstack-operator-controller-manager-557f5d867b-ndfjx" Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.891790 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/dc61bc9b-923b-41cb-b724-62c72caf0586-cert\") pod \"openstack-baremetal-operator-controller-manager-5869cb545-wgfjr\" (UID: \"dc61bc9b-923b-41cb-b724-62c72caf0586\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-wgfjr" Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.898041 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/dc61bc9b-923b-41cb-b724-62c72caf0586-cert\") pod \"openstack-baremetal-operator-controller-manager-5869cb545-wgfjr\" (UID: \"dc61bc9b-923b-41cb-b724-62c72caf0586\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-wgfjr" Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.913067 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-84958c4d49-2dsv9"] Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.913125 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-644bddb6d8-kw92z"] Oct 02 02:00:53 crc kubenswrapper[4885]: I1002 02:00:53.926728 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-bngfl" Oct 02 02:00:54 crc kubenswrapper[4885]: I1002 02:00:54.036666 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-wgfjr" Oct 02 02:00:54 crc kubenswrapper[4885]: I1002 02:00:54.119935 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5cd4858477-mjtzp"] Oct 02 02:00:54 crc kubenswrapper[4885]: I1002 02:00:54.300117 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8b18e6b1-82f2-46cd-99c8-85af816df20f-cert\") pod \"openstack-operator-controller-manager-557f5d867b-ndfjx\" (UID: \"8b18e6b1-82f2-46cd-99c8-85af816df20f\") " pod="openstack-operators/openstack-operator-controller-manager-557f5d867b-ndfjx" Oct 02 02:00:54 crc kubenswrapper[4885]: E1002 02:00:54.300389 4885 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Oct 02 02:00:54 crc kubenswrapper[4885]: E1002 02:00:54.300501 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8b18e6b1-82f2-46cd-99c8-85af816df20f-cert podName:8b18e6b1-82f2-46cd-99c8-85af816df20f nodeName:}" failed. No retries permitted until 2025-10-02 02:00:55.300483633 +0000 UTC m=+844.112231042 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/8b18e6b1-82f2-46cd-99c8-85af816df20f-cert") pod "openstack-operator-controller-manager-557f5d867b-ndfjx" (UID: "8b18e6b1-82f2-46cd-99c8-85af816df20f") : secret "webhook-server-cert" not found Oct 02 02:00:54 crc kubenswrapper[4885]: I1002 02:00:54.330130 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5d889d78cf-862d6"] Oct 02 02:00:54 crc kubenswrapper[4885]: W1002 02:00:54.338810 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod78117b64_3b3a_4892_90de_0dea12fa2602.slice/crio-051873ecadf12b54c699bb7db462f0cdd95140cecb9e790843c2d1653354d737 WatchSource:0}: Error finding container 051873ecadf12b54c699bb7db462f0cdd95140cecb9e790843c2d1653354d737: Status 404 returned error can't find the container with id 051873ecadf12b54c699bb7db462f0cdd95140cecb9e790843c2d1653354d737 Oct 02 02:00:54 crc kubenswrapper[4885]: I1002 02:00:54.344765 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-849d5b9b84-clsdk"] Oct 02 02:00:54 crc kubenswrapper[4885]: I1002 02:00:54.438596 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-clsdk" event={"ID":"ae73c7b6-6b7e-4ca9-8a4e-0694ab34d9a9","Type":"ContainerStarted","Data":"5f3e564595e888db5875eb88fe5366c8d8d3e4e37267e2af159a2dbf37024272"} Oct 02 02:00:54 crc kubenswrapper[4885]: I1002 02:00:54.440635 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-kw92z" event={"ID":"e5874982-d78d-4b7c-bf0a-19a7572dddad","Type":"ContainerStarted","Data":"0632a89b2c3543fd79e6ba9693b18c19569bed7a7fd8d002b36251cc8ced6639"} Oct 02 02:00:54 crc kubenswrapper[4885]: I1002 02:00:54.442162 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5cd4858477-mjtzp" event={"ID":"bbd60151-4081-46a8-85c0-0783604392a1","Type":"ContainerStarted","Data":"59331f41d4d89b641c97e2ee9b70942f9ca78ed1413cdbd500078ec6d04ef7f6"} Oct 02 02:00:54 crc kubenswrapper[4885]: I1002 02:00:54.442988 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-2dsv9" event={"ID":"ff10db28-28ca-4f65-8c24-2ed813d7cc5c","Type":"ContainerStarted","Data":"6c2be7167313f94944d03bdf32d4502d2eaf8684075666942bdbcf42c7be7834"} Oct 02 02:00:54 crc kubenswrapper[4885]: I1002 02:00:54.444016 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-862d6" event={"ID":"78117b64-3b3a-4892-90de-0dea12fa2602","Type":"ContainerStarted","Data":"051873ecadf12b54c699bb7db462f0cdd95140cecb9e790843c2d1653354d737"} Oct 02 02:00:54 crc kubenswrapper[4885]: I1002 02:00:54.445651 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-mw6h7" event={"ID":"cc08d74a-adca-4ae7-98c6-16c7037bc0ca","Type":"ContainerStarted","Data":"2e55ba36a7fa4f60828e2bec0923e0636f082ba5b1e91b6a660806aca9629ff1"} Oct 02 02:00:54 crc kubenswrapper[4885]: I1002 02:00:54.446608 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-wnxfk" event={"ID":"8113ab96-dde5-4f9f-b922-687e7a17afa7","Type":"ContainerStarted","Data":"14c3db0c0375ee44750f48df5b30022663d8212be698feb71a91b6cb5d2e032a"} Oct 02 02:00:54 crc kubenswrapper[4885]: I1002 02:00:54.483912 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-9f4696d94-pv94t"] Oct 02 02:00:54 crc kubenswrapper[4885]: W1002 02:00:54.490871 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1e64a13b_65af_4460_80a8_4a4659edd8aa.slice/crio-3f629935070c1cee4a3fd4e913cbdd6c7be15fa86749e2a41f780a6e1e47e42b WatchSource:0}: Error finding container 3f629935070c1cee4a3fd4e913cbdd6c7be15fa86749e2a41f780a6e1e47e42b: Status 404 returned error can't find the container with id 3f629935070c1cee4a3fd4e913cbdd6c7be15fa86749e2a41f780a6e1e47e42b Oct 02 02:00:54 crc kubenswrapper[4885]: I1002 02:00:54.498664 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-5bd55b4bff-pww46"] Oct 02 02:00:54 crc kubenswrapper[4885]: I1002 02:00:54.510292 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-7b787867f4-kcpkv"] Oct 02 02:00:54 crc kubenswrapper[4885]: W1002 02:00:54.516875 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd2263844_1435_4352_b773_04e8d10f35a0.slice/crio-9a26b554bc0c6992cfc805993f86242172dcf2ccf5cf487f33f58f2b02ce0433 WatchSource:0}: Error finding container 9a26b554bc0c6992cfc805993f86242172dcf2ccf5cf487f33f58f2b02ce0433: Status 404 returned error can't find the container with id 9a26b554bc0c6992cfc805993f86242172dcf2ccf5cf487f33f58f2b02ce0433 Oct 02 02:00:54 crc kubenswrapper[4885]: W1002 02:00:54.548228 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d724768_b5a0_4a55_951a_d350127061cb.slice/crio-14799bc1e71bec4510416d5d147cb01e7ad6ed32ccf8b71a614c44057b5ac492 WatchSource:0}: Error finding container 14799bc1e71bec4510416d5d147cb01e7ad6ed32ccf8b71a614c44057b5ac492: Status 404 returned error can't find the container with id 14799bc1e71bec4510416d5d147cb01e7ad6ed32ccf8b71a614c44057b5ac492 Oct 02 02:00:54 crc kubenswrapper[4885]: I1002 02:00:54.556743 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-88c7-nz599"] Oct 02 02:00:54 crc kubenswrapper[4885]: I1002 02:00:54.566328 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-6d68dbc695-9zf8p"] Oct 02 02:00:54 crc kubenswrapper[4885]: I1002 02:00:54.571402 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-9976ff44c-gl72q"] Oct 02 02:00:54 crc kubenswrapper[4885]: I1002 02:00:54.886831 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-bngfl"] Oct 02 02:00:54 crc kubenswrapper[4885]: I1002 02:00:54.896882 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-84d6b4b759-6f74g"] Oct 02 02:00:54 crc kubenswrapper[4885]: I1002 02:00:54.901947 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-64cd67b5cb-hwrj4"] Oct 02 02:00:54 crc kubenswrapper[4885]: I1002 02:00:54.904655 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-9d6c5db85-9fjl5"] Oct 02 02:00:54 crc kubenswrapper[4885]: I1002 02:00:54.911048 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-85777745bb-94284"] Oct 02 02:00:54 crc kubenswrapper[4885]: W1002 02:00:54.920159 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6e740063_fb49_4847_ba6d_77b6c089ba50.slice/crio-6e942b005ee5b00867441f2246f0bedc337052382fca8038fe0e4672f9bda825 WatchSource:0}: Error finding container 6e942b005ee5b00867441f2246f0bedc337052382fca8038fe0e4672f9bda825: Status 404 returned error can't find the container with id 6e942b005ee5b00867441f2246f0bedc337052382fca8038fe0e4672f9bda825 Oct 02 02:00:54 crc kubenswrapper[4885]: E1002 02:00:54.921972 4885 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:f61fdfbfd12027ce6b4e7ad553ec0582f080de0cfb472de6dc04ad3078bb17e3,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-d84qx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-85777745bb-94284_openstack-operators(74a07e3e-5061-43f9-b239-eb480bade999): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 02 02:00:54 crc kubenswrapper[4885]: E1002 02:00:54.925082 4885 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-rsjth,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-5f97d8c699-bngfl_openstack-operators(d2768afc-17ee-4141-aa68-7366faaa0145): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 02 02:00:54 crc kubenswrapper[4885]: E1002 02:00:54.925277 4885 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/infra-operator@sha256:3f96f0843934236c261db73dacb50fc12a288890562ee4ebdc9ec22360937cd3,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{600 -3} {} 600m DecimalSI},memory: {{2147483648 0} {} 2Gi BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{536870912 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7fqlj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod infra-operator-controller-manager-9d6c5db85-9fjl5_openstack-operators(45e282ed-8195-40dc-86cf-1177f3ebeb65): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 02 02:00:54 crc kubenswrapper[4885]: E1002 02:00:54.925440 4885 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:a6b3408d79df6b6d4a467e49defaa4a9d9c088c94d0605a4fee0030c9ccc84d2,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-44vfh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-589c58c6c-r65hz_openstack-operators(6e740063-fb49-4847-ba6d-77b6c089ba50): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 02 02:00:54 crc kubenswrapper[4885]: I1002 02:00:54.927356 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-589c58c6c-r65hz"] Oct 02 02:00:54 crc kubenswrapper[4885]: E1002 02:00:54.927587 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-bngfl" podUID="d2768afc-17ee-4141-aa68-7366faaa0145" Oct 02 02:00:54 crc kubenswrapper[4885]: I1002 02:00:54.938531 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6b9957f54f-mb6mf"] Oct 02 02:00:54 crc kubenswrapper[4885]: I1002 02:00:54.944184 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-26g2l"] Oct 02 02:00:54 crc kubenswrapper[4885]: E1002 02:00:54.959608 4885 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:09c2f519ea218f6038b7be039b8e6ac33ee93b217b9be0d2d18a5e7f94faae06,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-gc8nr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-6b9957f54f-mb6mf_openstack-operators(c5dd164b-699d-4f14-85d0-91f138a42c95): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 02 02:00:54 crc kubenswrapper[4885]: I1002 02:00:54.978222 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-wgfjr"] Oct 02 02:00:55 crc kubenswrapper[4885]: W1002 02:00:55.035241 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddc61bc9b_923b_41cb_b724_62c72caf0586.slice/crio-4b6f7482ceb316988401cf1820399e5ff2f3609ea0d4923608e27e1249a66c0f WatchSource:0}: Error finding container 4b6f7482ceb316988401cf1820399e5ff2f3609ea0d4923608e27e1249a66c0f: Status 404 returned error can't find the container with id 4b6f7482ceb316988401cf1820399e5ff2f3609ea0d4923608e27e1249a66c0f Oct 02 02:00:55 crc kubenswrapper[4885]: E1002 02:00:55.047236 4885 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:e7cfed051c1cf801e651fd4035070e38698039f284ac0b2a0332769fdbb4a9c8,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-baremetal-operator-agent:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_ANSIBLEEE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-ansibleee-runner:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_EVALUATOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-evaluator:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-listener:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_NOTIFIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-notifier:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_APACHE_IMAGE_URL_DEFAULT,Value:registry.redhat.io/ubi9/httpd-24:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_KEYSTONE_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-keystone-listener:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_IPMI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-ipmi:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_MYSQLD_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/mysqld-exporter:v0.15.1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_NOTIFICATION_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-notification:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_SGCORE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/sg-core:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_BACKUP_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-backup:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_VOLUME_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-volume:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_BACKENDBIND9_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-backend-bind9:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-central:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_MDNS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-mdns:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_PRODUCER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-producer:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_UNBOUND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-unbound:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_FRR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-frr:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_ISCSID_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-iscsid:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_KEPLER_IMAGE_URL_DEFAULT,Value:quay.io/sustainable_computing_io/kepler:release-0.7.12,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_LOGROTATE_CROND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cron:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_MULTIPATHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-multipathd:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_DHCP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_METADATA_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_OVN_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-ovn-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_SRIOV_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-sriov-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NODE_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/node-exporter:v1.5.0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_OVN_BGP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-bgp-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_PODMAN_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/navidys/prometheus-podman-exporter:v1.10.1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_GLANCE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-glance-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_CFNAPI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api-cfn:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-engine:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HORIZON_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_MEMCACHED_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-memcached:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_REDIS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-redis:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-conductor:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_INSPECTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-inspector:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_NEUTRON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-neutron-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PXE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-pxe:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PYTHON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/ironic-python-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KEYSTONE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-keystone:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KSM_IMAGE_URL_DEFAULT,Value:registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SHARE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-share:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MARIADB_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-mariadb:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NET_UTILS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-netutils:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NEUTRON_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-compute:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-conductor:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_NOVNC_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-novncproxy:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HEALTHMANAGER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-health-manager:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HOUSEKEEPING_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-housekeeping:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_RSYSLOG_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rsyslog:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_CLIENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-openstackclient:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_MUST_GATHER_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-must-gather:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_NETWORK_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OS_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/edpm-hardened-uefi:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_OVS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-base:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NORTHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-northd:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_SB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-sb-db-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_PLACEMENT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-placement-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_RABBITMQ_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_ACCOUNT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-account:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-container:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_OBJECT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-object:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_PROXY_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-proxy-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_TEST_TEMPEST_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_APPLIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-applier:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_DECISION_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-decision-engine:current-podified,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jq7ws,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-baremetal-operator-controller-manager-5869cb545-wgfjr_openstack-operators(dc61bc9b-923b-41cb-b724-62c72caf0586): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 02 02:00:55 crc kubenswrapper[4885]: E1002 02:00:55.182504 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-9fjl5" podUID="45e282ed-8195-40dc-86cf-1177f3ebeb65" Oct 02 02:00:55 crc kubenswrapper[4885]: E1002 02:00:55.271096 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/watcher-operator-controller-manager-6b9957f54f-mb6mf" podUID="c5dd164b-699d-4f14-85d0-91f138a42c95" Oct 02 02:00:55 crc kubenswrapper[4885]: E1002 02:00:55.296576 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/test-operator-controller-manager-85777745bb-94284" podUID="74a07e3e-5061-43f9-b239-eb480bade999" Oct 02 02:00:55 crc kubenswrapper[4885]: E1002 02:00:55.296758 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-r65hz" podUID="6e740063-fb49-4847-ba6d-77b6c089ba50" Oct 02 02:00:55 crc kubenswrapper[4885]: E1002 02:00:55.313005 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-wgfjr" podUID="dc61bc9b-923b-41cb-b724-62c72caf0586" Oct 02 02:00:55 crc kubenswrapper[4885]: I1002 02:00:55.315015 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8b18e6b1-82f2-46cd-99c8-85af816df20f-cert\") pod \"openstack-operator-controller-manager-557f5d867b-ndfjx\" (UID: \"8b18e6b1-82f2-46cd-99c8-85af816df20f\") " pod="openstack-operators/openstack-operator-controller-manager-557f5d867b-ndfjx" Oct 02 02:00:55 crc kubenswrapper[4885]: I1002 02:00:55.321630 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8b18e6b1-82f2-46cd-99c8-85af816df20f-cert\") pod \"openstack-operator-controller-manager-557f5d867b-ndfjx\" (UID: \"8b18e6b1-82f2-46cd-99c8-85af816df20f\") " pod="openstack-operators/openstack-operator-controller-manager-557f5d867b-ndfjx" Oct 02 02:00:55 crc kubenswrapper[4885]: I1002 02:00:55.341574 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-557f5d867b-ndfjx" Oct 02 02:00:55 crc kubenswrapper[4885]: I1002 02:00:55.504710 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-26g2l" event={"ID":"dd09a3bb-1217-4a3a-822b-2b2f04171271","Type":"ContainerStarted","Data":"024ac76e9f05060214eb22fb1f232bc7045f06673548aece77606b180f47618a"} Oct 02 02:00:55 crc kubenswrapper[4885]: I1002 02:00:55.506847 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-9fjl5" event={"ID":"45e282ed-8195-40dc-86cf-1177f3ebeb65","Type":"ContainerStarted","Data":"dc64986a37651f5586ebe83dd95806c43d8fbabf0073712a317988d0f53db7f7"} Oct 02 02:00:55 crc kubenswrapper[4885]: I1002 02:00:55.506882 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-9fjl5" event={"ID":"45e282ed-8195-40dc-86cf-1177f3ebeb65","Type":"ContainerStarted","Data":"eadfa2c84193e9358cbb9624081c1069dbde4e4174bf9fe320e90a53eeca6310"} Oct 02 02:00:55 crc kubenswrapper[4885]: I1002 02:00:55.510798 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6b9957f54f-mb6mf" event={"ID":"c5dd164b-699d-4f14-85d0-91f138a42c95","Type":"ContainerStarted","Data":"0a21ad99962d69081ab0c4e73ca0083c9a6316caf422da0aa905fa2c7a663ae9"} Oct 02 02:00:55 crc kubenswrapper[4885]: I1002 02:00:55.510827 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6b9957f54f-mb6mf" event={"ID":"c5dd164b-699d-4f14-85d0-91f138a42c95","Type":"ContainerStarted","Data":"f527d972d9ab8b5e420ab4fa99c7653f2d0e94e49200065ec20c22ac5aaa4b1c"} Oct 02 02:00:55 crc kubenswrapper[4885]: E1002 02:00:55.511393 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/infra-operator@sha256:3f96f0843934236c261db73dacb50fc12a288890562ee4ebdc9ec22360937cd3\\\"\"" pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-9fjl5" podUID="45e282ed-8195-40dc-86cf-1177f3ebeb65" Oct 02 02:00:55 crc kubenswrapper[4885]: E1002 02:00:55.512062 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:09c2f519ea218f6038b7be039b8e6ac33ee93b217b9be0d2d18a5e7f94faae06\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-6b9957f54f-mb6mf" podUID="c5dd164b-699d-4f14-85d0-91f138a42c95" Oct 02 02:00:55 crc kubenswrapper[4885]: I1002 02:00:55.517489 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-6f74g" event={"ID":"f80d2e7d-10a3-492f-a045-b48126a02490","Type":"ContainerStarted","Data":"02c0b81f110cb6bd32b3ba639b735f16fa7354f2545f546d4095b7dfdca8b165"} Oct 02 02:00:55 crc kubenswrapper[4885]: I1002 02:00:55.530273 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-r65hz" event={"ID":"6e740063-fb49-4847-ba6d-77b6c089ba50","Type":"ContainerStarted","Data":"dade59ede8297e47dc21503d4a8738875dd9c2785d250befe5252b066a3d9827"} Oct 02 02:00:55 crc kubenswrapper[4885]: I1002 02:00:55.530318 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-r65hz" event={"ID":"6e740063-fb49-4847-ba6d-77b6c089ba50","Type":"ContainerStarted","Data":"6e942b005ee5b00867441f2246f0bedc337052382fca8038fe0e4672f9bda825"} Oct 02 02:00:55 crc kubenswrapper[4885]: E1002 02:00:55.531573 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:a6b3408d79df6b6d4a467e49defaa4a9d9c088c94d0605a4fee0030c9ccc84d2\\\"\"" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-r65hz" podUID="6e740063-fb49-4847-ba6d-77b6c089ba50" Oct 02 02:00:55 crc kubenswrapper[4885]: I1002 02:00:55.532056 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-bngfl" event={"ID":"d2768afc-17ee-4141-aa68-7366faaa0145","Type":"ContainerStarted","Data":"2e38ebb8f1438221a59c1f507c0afe2c8b1f77dbfe04cbe8cd5190a3233deaa9"} Oct 02 02:00:55 crc kubenswrapper[4885]: E1002 02:00:55.532925 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-bngfl" podUID="d2768afc-17ee-4141-aa68-7366faaa0145" Oct 02 02:00:55 crc kubenswrapper[4885]: I1002 02:00:55.533453 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-9zf8p" event={"ID":"bcb6c44c-c620-465c-8db3-6d770deda522","Type":"ContainerStarted","Data":"fb2b54e8621b98b2a445f0d9e7ab3af998cc54f94592b3b7d1e230cbc629d61c"} Oct 02 02:00:55 crc kubenswrapper[4885]: I1002 02:00:55.535276 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-hwrj4" event={"ID":"e6673002-4d33-4c32-8eb4-f1727339b76d","Type":"ContainerStarted","Data":"d56622be82a12e3ce8157b98eaccf5547781b9bea8ecbf85ad7ba9e7fb17b5be"} Oct 02 02:00:55 crc kubenswrapper[4885]: I1002 02:00:55.537861 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-88c7-nz599" event={"ID":"0f1b84a4-59d6-45fe-9c2f-0bdff82fde6e","Type":"ContainerStarted","Data":"a5899c8ff54268b7a6a289e2cdf6c71e620a5691ea5a6f9765d180b12a7245bf"} Oct 02 02:00:55 crc kubenswrapper[4885]: I1002 02:00:55.558093 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-gl72q" event={"ID":"c540bfce-7653-4976-8d6e-36b479c3c711","Type":"ContainerStarted","Data":"cb43bb1d8a97c6a5a2db51b24938272d158d21fa3679071e4a189bd90c7ce4a0"} Oct 02 02:00:55 crc kubenswrapper[4885]: I1002 02:00:55.564236 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-pww46" event={"ID":"d2263844-1435-4352-b773-04e8d10f35a0","Type":"ContainerStarted","Data":"9a26b554bc0c6992cfc805993f86242172dcf2ccf5cf487f33f58f2b02ce0433"} Oct 02 02:00:55 crc kubenswrapper[4885]: I1002 02:00:55.566864 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-pv94t" event={"ID":"1e64a13b-65af-4460-80a8-4a4659edd8aa","Type":"ContainerStarted","Data":"3f629935070c1cee4a3fd4e913cbdd6c7be15fa86749e2a41f780a6e1e47e42b"} Oct 02 02:00:55 crc kubenswrapper[4885]: I1002 02:00:55.569798 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-kcpkv" event={"ID":"9d724768-b5a0-4a55-951a-d350127061cb","Type":"ContainerStarted","Data":"14799bc1e71bec4510416d5d147cb01e7ad6ed32ccf8b71a614c44057b5ac492"} Oct 02 02:00:55 crc kubenswrapper[4885]: I1002 02:00:55.594553 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-wgfjr" event={"ID":"dc61bc9b-923b-41cb-b724-62c72caf0586","Type":"ContainerStarted","Data":"970af3fc12789b457776d6622cd1d22dacdb95022e780912772eeaec2494ec0b"} Oct 02 02:00:55 crc kubenswrapper[4885]: I1002 02:00:55.594621 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-wgfjr" event={"ID":"dc61bc9b-923b-41cb-b724-62c72caf0586","Type":"ContainerStarted","Data":"4b6f7482ceb316988401cf1820399e5ff2f3609ea0d4923608e27e1249a66c0f"} Oct 02 02:00:55 crc kubenswrapper[4885]: E1002 02:00:55.595586 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:e7cfed051c1cf801e651fd4035070e38698039f284ac0b2a0332769fdbb4a9c8\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-wgfjr" podUID="dc61bc9b-923b-41cb-b724-62c72caf0586" Oct 02 02:00:55 crc kubenswrapper[4885]: I1002 02:00:55.603101 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-85777745bb-94284" event={"ID":"74a07e3e-5061-43f9-b239-eb480bade999","Type":"ContainerStarted","Data":"a880d6c862a6e05e9836f58a821e9ed8c302c6a0b3ba130cda265a6ef8f2a75c"} Oct 02 02:00:55 crc kubenswrapper[4885]: I1002 02:00:55.603132 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-85777745bb-94284" event={"ID":"74a07e3e-5061-43f9-b239-eb480bade999","Type":"ContainerStarted","Data":"9395fec2b6bea6ff2b878b9bedcad261c0369849c6dd9cb1b1bfc2c5fad0eff5"} Oct 02 02:00:55 crc kubenswrapper[4885]: E1002 02:00:55.605780 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:f61fdfbfd12027ce6b4e7ad553ec0582f080de0cfb472de6dc04ad3078bb17e3\\\"\"" pod="openstack-operators/test-operator-controller-manager-85777745bb-94284" podUID="74a07e3e-5061-43f9-b239-eb480bade999" Oct 02 02:00:56 crc kubenswrapper[4885]: I1002 02:00:56.009736 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-557f5d867b-ndfjx"] Oct 02 02:00:56 crc kubenswrapper[4885]: E1002 02:00:56.615982 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-bngfl" podUID="d2768afc-17ee-4141-aa68-7366faaa0145" Oct 02 02:00:56 crc kubenswrapper[4885]: E1002 02:00:56.616421 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:a6b3408d79df6b6d4a467e49defaa4a9d9c088c94d0605a4fee0030c9ccc84d2\\\"\"" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-r65hz" podUID="6e740063-fb49-4847-ba6d-77b6c089ba50" Oct 02 02:00:56 crc kubenswrapper[4885]: E1002 02:00:56.616464 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/infra-operator@sha256:3f96f0843934236c261db73dacb50fc12a288890562ee4ebdc9ec22360937cd3\\\"\"" pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-9fjl5" podUID="45e282ed-8195-40dc-86cf-1177f3ebeb65" Oct 02 02:00:56 crc kubenswrapper[4885]: E1002 02:00:56.616514 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:f61fdfbfd12027ce6b4e7ad553ec0582f080de0cfb472de6dc04ad3078bb17e3\\\"\"" pod="openstack-operators/test-operator-controller-manager-85777745bb-94284" podUID="74a07e3e-5061-43f9-b239-eb480bade999" Oct 02 02:00:56 crc kubenswrapper[4885]: E1002 02:00:56.616725 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:e7cfed051c1cf801e651fd4035070e38698039f284ac0b2a0332769fdbb4a9c8\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-wgfjr" podUID="dc61bc9b-923b-41cb-b724-62c72caf0586" Oct 02 02:00:56 crc kubenswrapper[4885]: E1002 02:00:56.617730 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:09c2f519ea218f6038b7be039b8e6ac33ee93b217b9be0d2d18a5e7f94faae06\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-6b9957f54f-mb6mf" podUID="c5dd164b-699d-4f14-85d0-91f138a42c95" Oct 02 02:00:57 crc kubenswrapper[4885]: W1002 02:00:57.424388 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8b18e6b1_82f2_46cd_99c8_85af816df20f.slice/crio-38b48a633a8379ea5be8943d4980a2dc11682e079406267a6fa824defc691b07 WatchSource:0}: Error finding container 38b48a633a8379ea5be8943d4980a2dc11682e079406267a6fa824defc691b07: Status 404 returned error can't find the container with id 38b48a633a8379ea5be8943d4980a2dc11682e079406267a6fa824defc691b07 Oct 02 02:00:57 crc kubenswrapper[4885]: I1002 02:00:57.618528 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-557f5d867b-ndfjx" event={"ID":"8b18e6b1-82f2-46cd-99c8-85af816df20f","Type":"ContainerStarted","Data":"38b48a633a8379ea5be8943d4980a2dc11682e079406267a6fa824defc691b07"} Oct 02 02:01:07 crc kubenswrapper[4885]: I1002 02:01:07.695500 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-wnxfk" event={"ID":"8113ab96-dde5-4f9f-b922-687e7a17afa7","Type":"ContainerStarted","Data":"fb1e4d6746c65e763f5eb0dd4439f47ebcf70c6784396d60076dcdca60730105"} Oct 02 02:01:07 crc kubenswrapper[4885]: I1002 02:01:07.698709 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-88c7-nz599" event={"ID":"0f1b84a4-59d6-45fe-9c2f-0bdff82fde6e","Type":"ContainerStarted","Data":"df366ed11122d8190e4340bcf52daf980ce20c849b2d3b64fd8f39d6b22b4b0b"} Oct 02 02:01:07 crc kubenswrapper[4885]: I1002 02:01:07.707935 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-557f5d867b-ndfjx" event={"ID":"8b18e6b1-82f2-46cd-99c8-85af816df20f","Type":"ContainerStarted","Data":"071ec5a4ebeacec5f08456984c78dc3b10bce4742e0bf5c17e72c46beba65c63"} Oct 02 02:01:07 crc kubenswrapper[4885]: I1002 02:01:07.735386 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-862d6" event={"ID":"78117b64-3b3a-4892-90de-0dea12fa2602","Type":"ContainerStarted","Data":"d1ac915e00a0c79067de608558145ff3378d2ac160860ebbdad8934df6920990"} Oct 02 02:01:07 crc kubenswrapper[4885]: I1002 02:01:07.737438 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-pww46" event={"ID":"d2263844-1435-4352-b773-04e8d10f35a0","Type":"ContainerStarted","Data":"57aba7fba950b9bf1f5a73d0852d0774cbb3044340bcaef3672b4ae85590f3b3"} Oct 02 02:01:07 crc kubenswrapper[4885]: I1002 02:01:07.739794 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-2dsv9" event={"ID":"ff10db28-28ca-4f65-8c24-2ed813d7cc5c","Type":"ContainerStarted","Data":"16e2fc5f4298554c888999383b9d5d6e22f151226ac020399ad31529cd5cf66e"} Oct 02 02:01:07 crc kubenswrapper[4885]: I1002 02:01:07.751918 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-9zf8p" event={"ID":"bcb6c44c-c620-465c-8db3-6d770deda522","Type":"ContainerStarted","Data":"607e2bfd7791e62808f20a4811bb95d13fca1a4bc34b135491270a3680ae97a7"} Oct 02 02:01:07 crc kubenswrapper[4885]: I1002 02:01:07.754024 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-hwrj4" event={"ID":"e6673002-4d33-4c32-8eb4-f1727339b76d","Type":"ContainerStarted","Data":"bae52ad19ecffd8bdaade876d51c21cc221f06e181b087f1cdf07bdf986ab026"} Oct 02 02:01:07 crc kubenswrapper[4885]: I1002 02:01:07.755736 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5cd4858477-mjtzp" event={"ID":"bbd60151-4081-46a8-85c0-0783604392a1","Type":"ContainerStarted","Data":"09283eb09d90852ac1e11f07e6284ea143fd63978a6d619097051eb5c80e4413"} Oct 02 02:01:07 crc kubenswrapper[4885]: I1002 02:01:07.758619 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-gl72q" event={"ID":"c540bfce-7653-4976-8d6e-36b479c3c711","Type":"ContainerStarted","Data":"0e27ead30b33d15b24ce6b437a0ebefb105b4b642e8427027f8a7ff8757bf3d0"} Oct 02 02:01:07 crc kubenswrapper[4885]: I1002 02:01:07.769527 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-6f74g" event={"ID":"f80d2e7d-10a3-492f-a045-b48126a02490","Type":"ContainerStarted","Data":"f77de1bc1553e888923e9b49b909795ac8e0f919604f5239448c066a86bc9b67"} Oct 02 02:01:08 crc kubenswrapper[4885]: I1002 02:01:08.785422 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-gl72q" event={"ID":"c540bfce-7653-4976-8d6e-36b479c3c711","Type":"ContainerStarted","Data":"2ed837b9ca40e4f35f3932dbe37cd5e935e378998cbed1f8db6b34b828bacd1d"} Oct 02 02:01:08 crc kubenswrapper[4885]: I1002 02:01:08.785836 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-gl72q" Oct 02 02:01:08 crc kubenswrapper[4885]: I1002 02:01:08.794688 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-6f74g" event={"ID":"f80d2e7d-10a3-492f-a045-b48126a02490","Type":"ContainerStarted","Data":"3aa57b798115cb7f26f27938cb53fe9865ada07222012ebdec40d064a9f04de7"} Oct 02 02:01:08 crc kubenswrapper[4885]: I1002 02:01:08.794819 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-6f74g" Oct 02 02:01:08 crc kubenswrapper[4885]: I1002 02:01:08.801708 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-862d6" event={"ID":"78117b64-3b3a-4892-90de-0dea12fa2602","Type":"ContainerStarted","Data":"4d9e56e4eaf72ff9b2fbcdc36208dec2967a1c554cdccd6c0c120d824aee0797"} Oct 02 02:01:08 crc kubenswrapper[4885]: I1002 02:01:08.802464 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-862d6" Oct 02 02:01:08 crc kubenswrapper[4885]: I1002 02:01:08.804010 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-557f5d867b-ndfjx" event={"ID":"8b18e6b1-82f2-46cd-99c8-85af816df20f","Type":"ContainerStarted","Data":"61568a0dbb5ef88e91eb2a83d8a8e2d6f4d7712e309b206a753b5a06362e1930"} Oct 02 02:01:08 crc kubenswrapper[4885]: I1002 02:01:08.804513 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-557f5d867b-ndfjx" Oct 02 02:01:08 crc kubenswrapper[4885]: I1002 02:01:08.816829 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-pww46" event={"ID":"d2263844-1435-4352-b773-04e8d10f35a0","Type":"ContainerStarted","Data":"e679c028ba0dc590ca46d8a1ce4e08e1b8dfb30f9a5aeab6beda0d2906feb2df"} Oct 02 02:01:08 crc kubenswrapper[4885]: I1002 02:01:08.817458 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-pww46" Oct 02 02:01:08 crc kubenswrapper[4885]: I1002 02:01:08.821616 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-hwrj4" event={"ID":"e6673002-4d33-4c32-8eb4-f1727339b76d","Type":"ContainerStarted","Data":"8133de009425ef8b9cd7b3b09fb1bd746b4a3f8a144e4e5221a19d8fddabdf57"} Oct 02 02:01:08 crc kubenswrapper[4885]: I1002 02:01:08.822031 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-hwrj4" Oct 02 02:01:08 crc kubenswrapper[4885]: I1002 02:01:08.830458 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5cd4858477-mjtzp" event={"ID":"bbd60151-4081-46a8-85c0-0783604392a1","Type":"ContainerStarted","Data":"f82432ad47a7b04996ac1f9a2a5d1e7cb72a79561d0cbb05ebfbc03b1f1d822f"} Oct 02 02:01:08 crc kubenswrapper[4885]: I1002 02:01:08.830561 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-5cd4858477-mjtzp" Oct 02 02:01:08 crc kubenswrapper[4885]: I1002 02:01:08.832084 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-2dsv9" event={"ID":"ff10db28-28ca-4f65-8c24-2ed813d7cc5c","Type":"ContainerStarted","Data":"2ad9769bbbbfff33d5f289d131a9f3a6b4915f030dedc7245ed64073ceb1e31b"} Oct 02 02:01:08 crc kubenswrapper[4885]: I1002 02:01:08.832453 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-2dsv9" Oct 02 02:01:08 crc kubenswrapper[4885]: I1002 02:01:08.833906 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-kw92z" event={"ID":"e5874982-d78d-4b7c-bf0a-19a7572dddad","Type":"ContainerStarted","Data":"9341bb07257e78e9b5c273eb6414b39c3f6fc83123b246038a656a105f831bc4"} Oct 02 02:01:08 crc kubenswrapper[4885]: I1002 02:01:08.833927 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-kw92z" event={"ID":"e5874982-d78d-4b7c-bf0a-19a7572dddad","Type":"ContainerStarted","Data":"01e2ff01813e9022d48f94170305904c2298a41f9f86efacee2b2d8352d67531"} Oct 02 02:01:08 crc kubenswrapper[4885]: I1002 02:01:08.834243 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-kw92z" Oct 02 02:01:08 crc kubenswrapper[4885]: I1002 02:01:08.838820 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-gl72q" podStartSLOduration=4.489130744 podStartE2EDuration="16.838795576s" podCreationTimestamp="2025-10-02 02:00:52 +0000 UTC" firstStartedPulling="2025-10-02 02:00:54.581369147 +0000 UTC m=+843.393116546" lastFinishedPulling="2025-10-02 02:01:06.931033969 +0000 UTC m=+855.742781378" observedRunningTime="2025-10-02 02:01:08.813429125 +0000 UTC m=+857.625176524" watchObservedRunningTime="2025-10-02 02:01:08.838795576 +0000 UTC m=+857.650542975" Oct 02 02:01:08 crc kubenswrapper[4885]: I1002 02:01:08.840574 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-88c7-nz599" event={"ID":"0f1b84a4-59d6-45fe-9c2f-0bdff82fde6e","Type":"ContainerStarted","Data":"94e7fa72ced5d388ba74d4407414e2d7f62b295d966f8c4d0baccd9d412b3f76"} Oct 02 02:01:08 crc kubenswrapper[4885]: I1002 02:01:08.841284 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-88c7-nz599" Oct 02 02:01:08 crc kubenswrapper[4885]: I1002 02:01:08.853444 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-pv94t" event={"ID":"1e64a13b-65af-4460-80a8-4a4659edd8aa","Type":"ContainerStarted","Data":"a2b7393f2adb1d67e5baad9d91b5ea476755eaaf455e15b6679a6a9e87e6acc3"} Oct 02 02:01:08 crc kubenswrapper[4885]: I1002 02:01:08.853484 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-pv94t" event={"ID":"1e64a13b-65af-4460-80a8-4a4659edd8aa","Type":"ContainerStarted","Data":"b7ce1d6aea51f9a75b164e8d743423b0c6210e6d2b6e00465b51f8b32aaebcd7"} Oct 02 02:01:08 crc kubenswrapper[4885]: I1002 02:01:08.853680 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-pv94t" Oct 02 02:01:08 crc kubenswrapper[4885]: I1002 02:01:08.865816 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-kcpkv" event={"ID":"9d724768-b5a0-4a55-951a-d350127061cb","Type":"ContainerStarted","Data":"ee61b8761b453e5020110b7d205bd1ebe8a946b5a8e04256a06cf98c289f2355"} Oct 02 02:01:08 crc kubenswrapper[4885]: I1002 02:01:08.865951 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-kcpkv" event={"ID":"9d724768-b5a0-4a55-951a-d350127061cb","Type":"ContainerStarted","Data":"16a4f4e8fae995793e61607e219d9d1e55173449af1f9b469d6f60fdeb0a2211"} Oct 02 02:01:08 crc kubenswrapper[4885]: I1002 02:01:08.866524 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-kcpkv" Oct 02 02:01:08 crc kubenswrapper[4885]: I1002 02:01:08.868359 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-clsdk" event={"ID":"ae73c7b6-6b7e-4ca9-8a4e-0694ab34d9a9","Type":"ContainerStarted","Data":"ec8e80c277039070e84637e3559d361d470f97bbb033c6bbb6294ae1986a2124"} Oct 02 02:01:08 crc kubenswrapper[4885]: I1002 02:01:08.868465 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-clsdk" event={"ID":"ae73c7b6-6b7e-4ca9-8a4e-0694ab34d9a9","Type":"ContainerStarted","Data":"8397b844c67e06f2dddf33d20dad7cf410ecd3096b3b61bf7f6227aafc04959f"} Oct 02 02:01:08 crc kubenswrapper[4885]: I1002 02:01:08.868938 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-clsdk" Oct 02 02:01:08 crc kubenswrapper[4885]: I1002 02:01:08.870387 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-26g2l" event={"ID":"dd09a3bb-1217-4a3a-822b-2b2f04171271","Type":"ContainerStarted","Data":"971ec05ef4d1293219d9aa9ec57cfa6233b6899df6b304a185575459c89d377e"} Oct 02 02:01:08 crc kubenswrapper[4885]: I1002 02:01:08.870410 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-26g2l" event={"ID":"dd09a3bb-1217-4a3a-822b-2b2f04171271","Type":"ContainerStarted","Data":"be22b52748e719bddd9afea8dd074324c22fcbbc4f9736547b01b8bd9c4d683d"} Oct 02 02:01:08 crc kubenswrapper[4885]: I1002 02:01:08.870829 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-26g2l" Oct 02 02:01:08 crc kubenswrapper[4885]: I1002 02:01:08.886585 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-9zf8p" event={"ID":"bcb6c44c-c620-465c-8db3-6d770deda522","Type":"ContainerStarted","Data":"5839ef70024212e6fbe08bb9eabbc7181e2f2f6935e492ffe9c6449a04536b38"} Oct 02 02:01:08 crc kubenswrapper[4885]: I1002 02:01:08.886888 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-9zf8p" Oct 02 02:01:08 crc kubenswrapper[4885]: I1002 02:01:08.889737 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-557f5d867b-ndfjx" podStartSLOduration=15.889718256 podStartE2EDuration="15.889718256s" podCreationTimestamp="2025-10-02 02:00:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:01:08.881410133 +0000 UTC m=+857.693157532" watchObservedRunningTime="2025-10-02 02:01:08.889718256 +0000 UTC m=+857.701465665" Oct 02 02:01:08 crc kubenswrapper[4885]: I1002 02:01:08.891026 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-6f74g" podStartSLOduration=4.898592397 podStartE2EDuration="16.891016796s" podCreationTimestamp="2025-10-02 02:00:52 +0000 UTC" firstStartedPulling="2025-10-02 02:00:54.910474645 +0000 UTC m=+843.722222044" lastFinishedPulling="2025-10-02 02:01:06.902899044 +0000 UTC m=+855.714646443" observedRunningTime="2025-10-02 02:01:08.841433297 +0000 UTC m=+857.653180696" watchObservedRunningTime="2025-10-02 02:01:08.891016796 +0000 UTC m=+857.702764205" Oct 02 02:01:08 crc kubenswrapper[4885]: I1002 02:01:08.892453 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-wnxfk" event={"ID":"8113ab96-dde5-4f9f-b922-687e7a17afa7","Type":"ContainerStarted","Data":"0b141f690d59f7227088ffad8139e9c5ab247b894d7786ebdb913fcada4049de"} Oct 02 02:01:08 crc kubenswrapper[4885]: I1002 02:01:08.892695 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-wnxfk" Oct 02 02:01:08 crc kubenswrapper[4885]: I1002 02:01:08.910711 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-862d6" podStartSLOduration=4.313905633 podStartE2EDuration="16.910693594s" podCreationTimestamp="2025-10-02 02:00:52 +0000 UTC" firstStartedPulling="2025-10-02 02:00:54.341150729 +0000 UTC m=+843.152898128" lastFinishedPulling="2025-10-02 02:01:06.93793868 +0000 UTC m=+855.749686089" observedRunningTime="2025-10-02 02:01:08.909570851 +0000 UTC m=+857.721318250" watchObservedRunningTime="2025-10-02 02:01:08.910693594 +0000 UTC m=+857.722440983" Oct 02 02:01:08 crc kubenswrapper[4885]: I1002 02:01:08.922593 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-mw6h7" event={"ID":"cc08d74a-adca-4ae7-98c6-16c7037bc0ca","Type":"ContainerStarted","Data":"9ff0aaa552c411b2b8306cba2161e76810e62d1bd68dab9647cbb5a70906e9d7"} Oct 02 02:01:08 crc kubenswrapper[4885]: I1002 02:01:08.922642 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-mw6h7" event={"ID":"cc08d74a-adca-4ae7-98c6-16c7037bc0ca","Type":"ContainerStarted","Data":"c10318d26370a0a3043fe17f215fd152180bc85cf6a12aafae91f06fdda25e37"} Oct 02 02:01:08 crc kubenswrapper[4885]: I1002 02:01:08.923241 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-mw6h7" Oct 02 02:01:08 crc kubenswrapper[4885]: I1002 02:01:08.944898 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-hwrj4" podStartSLOduration=4.915593765 podStartE2EDuration="16.944878305s" podCreationTimestamp="2025-10-02 02:00:52 +0000 UTC" firstStartedPulling="2025-10-02 02:00:54.918795258 +0000 UTC m=+843.730542657" lastFinishedPulling="2025-10-02 02:01:06.948079798 +0000 UTC m=+855.759827197" observedRunningTime="2025-10-02 02:01:08.939526542 +0000 UTC m=+857.751273941" watchObservedRunningTime="2025-10-02 02:01:08.944878305 +0000 UTC m=+857.756625694" Oct 02 02:01:08 crc kubenswrapper[4885]: I1002 02:01:08.960925 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-26g2l" podStartSLOduration=3.9413693690000002 podStartE2EDuration="15.960910173s" podCreationTimestamp="2025-10-02 02:00:53 +0000 UTC" firstStartedPulling="2025-10-02 02:00:54.937473406 +0000 UTC m=+843.749220805" lastFinishedPulling="2025-10-02 02:01:06.9570142 +0000 UTC m=+855.768761609" observedRunningTime="2025-10-02 02:01:08.959378216 +0000 UTC m=+857.771125615" watchObservedRunningTime="2025-10-02 02:01:08.960910173 +0000 UTC m=+857.772657572" Oct 02 02:01:08 crc kubenswrapper[4885]: I1002 02:01:08.981941 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-9zf8p" podStartSLOduration=4.65725182 podStartE2EDuration="16.981930572s" podCreationTimestamp="2025-10-02 02:00:52 +0000 UTC" firstStartedPulling="2025-10-02 02:00:54.577713156 +0000 UTC m=+843.389460545" lastFinishedPulling="2025-10-02 02:01:06.902391888 +0000 UTC m=+855.714139297" observedRunningTime="2025-10-02 02:01:08.980699544 +0000 UTC m=+857.792446943" watchObservedRunningTime="2025-10-02 02:01:08.981930572 +0000 UTC m=+857.793677971" Oct 02 02:01:09 crc kubenswrapper[4885]: I1002 02:01:09.000722 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-88c7-nz599" podStartSLOduration=4.668251085 podStartE2EDuration="17.000708624s" podCreationTimestamp="2025-10-02 02:00:52 +0000 UTC" firstStartedPulling="2025-10-02 02:00:54.572946981 +0000 UTC m=+843.384694380" lastFinishedPulling="2025-10-02 02:01:06.90540449 +0000 UTC m=+855.717151919" observedRunningTime="2025-10-02 02:01:08.996810734 +0000 UTC m=+857.808558133" watchObservedRunningTime="2025-10-02 02:01:09.000708624 +0000 UTC m=+857.812456023" Oct 02 02:01:09 crc kubenswrapper[4885]: I1002 02:01:09.021361 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-wnxfk" podStartSLOduration=3.843727816 podStartE2EDuration="17.021345192s" podCreationTimestamp="2025-10-02 02:00:52 +0000 UTC" firstStartedPulling="2025-10-02 02:00:53.738148389 +0000 UTC m=+842.549895788" lastFinishedPulling="2025-10-02 02:01:06.915765755 +0000 UTC m=+855.727513164" observedRunningTime="2025-10-02 02:01:09.018509185 +0000 UTC m=+857.830256584" watchObservedRunningTime="2025-10-02 02:01:09.021345192 +0000 UTC m=+857.833092591" Oct 02 02:01:09 crc kubenswrapper[4885]: I1002 02:01:09.043590 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-pww46" podStartSLOduration=4.648649161 podStartE2EDuration="17.043573068s" podCreationTimestamp="2025-10-02 02:00:52 +0000 UTC" firstStartedPulling="2025-10-02 02:00:54.519220649 +0000 UTC m=+843.330968048" lastFinishedPulling="2025-10-02 02:01:06.914144546 +0000 UTC m=+855.725891955" observedRunningTime="2025-10-02 02:01:09.038793612 +0000 UTC m=+857.850541011" watchObservedRunningTime="2025-10-02 02:01:09.043573068 +0000 UTC m=+857.855320477" Oct 02 02:01:09 crc kubenswrapper[4885]: I1002 02:01:09.064374 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-2dsv9" podStartSLOduration=4.078210401 podStartE2EDuration="17.06435904s" podCreationTimestamp="2025-10-02 02:00:52 +0000 UTC" firstStartedPulling="2025-10-02 02:00:53.935831925 +0000 UTC m=+842.747579324" lastFinishedPulling="2025-10-02 02:01:06.921980524 +0000 UTC m=+855.733727963" observedRunningTime="2025-10-02 02:01:09.05352491 +0000 UTC m=+857.865272299" watchObservedRunningTime="2025-10-02 02:01:09.06435904 +0000 UTC m=+857.876106439" Oct 02 02:01:09 crc kubenswrapper[4885]: I1002 02:01:09.090489 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-pv94t" podStartSLOduration=4.6132455839999995 podStartE2EDuration="17.090469974s" podCreationTimestamp="2025-10-02 02:00:52 +0000 UTC" firstStartedPulling="2025-10-02 02:00:54.49589861 +0000 UTC m=+843.307646009" lastFinishedPulling="2025-10-02 02:01:06.97312299 +0000 UTC m=+855.784870399" observedRunningTime="2025-10-02 02:01:09.086558495 +0000 UTC m=+857.898305894" watchObservedRunningTime="2025-10-02 02:01:09.090469974 +0000 UTC m=+857.902217373" Oct 02 02:01:09 crc kubenswrapper[4885]: I1002 02:01:09.091464 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-kcpkv" podStartSLOduration=4.681508608 podStartE2EDuration="17.091456524s" podCreationTimestamp="2025-10-02 02:00:52 +0000 UTC" firstStartedPulling="2025-10-02 02:00:54.55876088 +0000 UTC m=+843.370508279" lastFinishedPulling="2025-10-02 02:01:06.968708766 +0000 UTC m=+855.780456195" observedRunningTime="2025-10-02 02:01:09.07224816 +0000 UTC m=+857.883995559" watchObservedRunningTime="2025-10-02 02:01:09.091456524 +0000 UTC m=+857.903203923" Oct 02 02:01:09 crc kubenswrapper[4885]: I1002 02:01:09.115598 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-kw92z" podStartSLOduration=4.071888899 podStartE2EDuration="17.115587039s" podCreationTimestamp="2025-10-02 02:00:52 +0000 UTC" firstStartedPulling="2025-10-02 02:00:53.925050727 +0000 UTC m=+842.736798126" lastFinishedPulling="2025-10-02 02:01:06.968748857 +0000 UTC m=+855.780496266" observedRunningTime="2025-10-02 02:01:09.114893228 +0000 UTC m=+857.926640627" watchObservedRunningTime="2025-10-02 02:01:09.115587039 +0000 UTC m=+857.927334438" Oct 02 02:01:09 crc kubenswrapper[4885]: I1002 02:01:09.130295 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-5cd4858477-mjtzp" podStartSLOduration=4.374263509 podStartE2EDuration="17.130278286s" podCreationTimestamp="2025-10-02 02:00:52 +0000 UTC" firstStartedPulling="2025-10-02 02:00:54.147429793 +0000 UTC m=+842.959177192" lastFinishedPulling="2025-10-02 02:01:06.90344454 +0000 UTC m=+855.715191969" observedRunningTime="2025-10-02 02:01:09.128230073 +0000 UTC m=+857.939977472" watchObservedRunningTime="2025-10-02 02:01:09.130278286 +0000 UTC m=+857.942025685" Oct 02 02:01:09 crc kubenswrapper[4885]: I1002 02:01:09.148894 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-clsdk" podStartSLOduration=4.584215757 podStartE2EDuration="17.148879272s" podCreationTimestamp="2025-10-02 02:00:52 +0000 UTC" firstStartedPulling="2025-10-02 02:00:54.350519203 +0000 UTC m=+843.162266602" lastFinishedPulling="2025-10-02 02:01:06.915182708 +0000 UTC m=+855.726930117" observedRunningTime="2025-10-02 02:01:09.143836028 +0000 UTC m=+857.955583427" watchObservedRunningTime="2025-10-02 02:01:09.148879272 +0000 UTC m=+857.960626671" Oct 02 02:01:09 crc kubenswrapper[4885]: I1002 02:01:09.158539 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-mw6h7" podStartSLOduration=3.93356175 podStartE2EDuration="17.158526045s" podCreationTimestamp="2025-10-02 02:00:52 +0000 UTC" firstStartedPulling="2025-10-02 02:00:53.727895628 +0000 UTC m=+842.539643027" lastFinishedPulling="2025-10-02 02:01:06.952859923 +0000 UTC m=+855.764607322" observedRunningTime="2025-10-02 02:01:09.155138712 +0000 UTC m=+857.966886111" watchObservedRunningTime="2025-10-02 02:01:09.158526045 +0000 UTC m=+857.970273444" Oct 02 02:01:10 crc kubenswrapper[4885]: I1002 02:01:10.944667 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-bngfl" event={"ID":"d2768afc-17ee-4141-aa68-7366faaa0145","Type":"ContainerStarted","Data":"228adbeb21a8c207f17fb96463357e5c2a9a14066ffead62ac8338d90e18664c"} Oct 02 02:01:10 crc kubenswrapper[4885]: I1002 02:01:10.965369 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-bngfl" podStartSLOduration=2.79806448 podStartE2EDuration="17.965354651s" podCreationTimestamp="2025-10-02 02:00:53 +0000 UTC" firstStartedPulling="2025-10-02 02:00:54.924966655 +0000 UTC m=+843.736714054" lastFinishedPulling="2025-10-02 02:01:10.092256806 +0000 UTC m=+858.904004225" observedRunningTime="2025-10-02 02:01:10.962065151 +0000 UTC m=+859.773812540" watchObservedRunningTime="2025-10-02 02:01:10.965354651 +0000 UTC m=+859.777102050" Oct 02 02:01:11 crc kubenswrapper[4885]: I1002 02:01:11.954424 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-wgfjr" event={"ID":"dc61bc9b-923b-41cb-b724-62c72caf0586","Type":"ContainerStarted","Data":"43ea3491807ad4786f7a290bba7314bd28af6d8f274a99403419c37b24a8a512"} Oct 02 02:01:11 crc kubenswrapper[4885]: I1002 02:01:11.954650 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-wgfjr" Oct 02 02:01:11 crc kubenswrapper[4885]: I1002 02:01:11.960561 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-85777745bb-94284" event={"ID":"74a07e3e-5061-43f9-b239-eb480bade999","Type":"ContainerStarted","Data":"ff17fe14b3cb7e6df3da75f1c3090f87ce3945dab5b9264eb0ae11f21e4a5a03"} Oct 02 02:01:11 crc kubenswrapper[4885]: I1002 02:01:11.961457 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-85777745bb-94284" Oct 02 02:01:11 crc kubenswrapper[4885]: I1002 02:01:11.986224 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-wgfjr" podStartSLOduration=3.3299410910000002 podStartE2EDuration="19.986200982s" podCreationTimestamp="2025-10-02 02:00:52 +0000 UTC" firstStartedPulling="2025-10-02 02:00:55.046856468 +0000 UTC m=+843.858603857" lastFinishedPulling="2025-10-02 02:01:11.703116319 +0000 UTC m=+860.514863748" observedRunningTime="2025-10-02 02:01:11.982021845 +0000 UTC m=+860.793769284" watchObservedRunningTime="2025-10-02 02:01:11.986200982 +0000 UTC m=+860.797948391" Oct 02 02:01:13 crc kubenswrapper[4885]: I1002 02:01:13.043114 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-kw92z" Oct 02 02:01:13 crc kubenswrapper[4885]: I1002 02:01:13.053667 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-wnxfk" Oct 02 02:01:13 crc kubenswrapper[4885]: I1002 02:01:13.062051 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-mw6h7" Oct 02 02:01:13 crc kubenswrapper[4885]: I1002 02:01:13.068725 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-85777745bb-94284" podStartSLOduration=3.283492061 podStartE2EDuration="20.06870413s" podCreationTimestamp="2025-10-02 02:00:53 +0000 UTC" firstStartedPulling="2025-10-02 02:00:54.92182113 +0000 UTC m=+843.733568529" lastFinishedPulling="2025-10-02 02:01:11.707033169 +0000 UTC m=+860.518780598" observedRunningTime="2025-10-02 02:01:12.00191628 +0000 UTC m=+860.813663689" watchObservedRunningTime="2025-10-02 02:01:13.06870413 +0000 UTC m=+861.880451569" Oct 02 02:01:13 crc kubenswrapper[4885]: I1002 02:01:13.081972 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-2dsv9" Oct 02 02:01:13 crc kubenswrapper[4885]: I1002 02:01:13.096958 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-862d6" Oct 02 02:01:13 crc kubenswrapper[4885]: I1002 02:01:13.124245 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-pv94t" Oct 02 02:01:13 crc kubenswrapper[4885]: I1002 02:01:13.182765 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-5cd4858477-mjtzp" Oct 02 02:01:13 crc kubenswrapper[4885]: I1002 02:01:13.218224 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-pww46" Oct 02 02:01:13 crc kubenswrapper[4885]: I1002 02:01:13.298821 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-9zf8p" Oct 02 02:01:13 crc kubenswrapper[4885]: I1002 02:01:13.332098 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-88c7-nz599" Oct 02 02:01:13 crc kubenswrapper[4885]: I1002 02:01:13.367743 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-clsdk" Oct 02 02:01:13 crc kubenswrapper[4885]: I1002 02:01:13.411186 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-kcpkv" Oct 02 02:01:13 crc kubenswrapper[4885]: I1002 02:01:13.413601 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-hwrj4" Oct 02 02:01:13 crc kubenswrapper[4885]: I1002 02:01:13.419452 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-gl72q" Oct 02 02:01:13 crc kubenswrapper[4885]: I1002 02:01:13.499177 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-6f74g" Oct 02 02:01:13 crc kubenswrapper[4885]: I1002 02:01:13.547617 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-26g2l" Oct 02 02:01:13 crc kubenswrapper[4885]: I1002 02:01:13.986773 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6b9957f54f-mb6mf" event={"ID":"c5dd164b-699d-4f14-85d0-91f138a42c95","Type":"ContainerStarted","Data":"9423af76bd73e49b61ae1046dd3fd35ffd6a1f0e2d9d30bd503fa5d187df6c74"} Oct 02 02:01:13 crc kubenswrapper[4885]: I1002 02:01:13.987200 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-6b9957f54f-mb6mf" Oct 02 02:01:13 crc kubenswrapper[4885]: I1002 02:01:13.989083 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-r65hz" event={"ID":"6e740063-fb49-4847-ba6d-77b6c089ba50","Type":"ContainerStarted","Data":"ef1a92fb115ac08b6d39ee3d242ad655ed079a02cf113a7e7ecfc60d75c5b1ee"} Oct 02 02:01:13 crc kubenswrapper[4885]: I1002 02:01:13.989453 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-r65hz" Oct 02 02:01:14 crc kubenswrapper[4885]: I1002 02:01:14.005745 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-6b9957f54f-mb6mf" podStartSLOduration=2.283832533 podStartE2EDuration="21.00572426s" podCreationTimestamp="2025-10-02 02:00:53 +0000 UTC" firstStartedPulling="2025-10-02 02:00:54.959443373 +0000 UTC m=+843.771190772" lastFinishedPulling="2025-10-02 02:01:13.6813351 +0000 UTC m=+862.493082499" observedRunningTime="2025-10-02 02:01:14.000481711 +0000 UTC m=+862.812229120" watchObservedRunningTime="2025-10-02 02:01:14.00572426 +0000 UTC m=+862.817471659" Oct 02 02:01:14 crc kubenswrapper[4885]: I1002 02:01:14.021530 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-r65hz" podStartSLOduration=3.25179384 podStartE2EDuration="22.021513521s" podCreationTimestamp="2025-10-02 02:00:52 +0000 UTC" firstStartedPulling="2025-10-02 02:00:54.925363897 +0000 UTC m=+843.737111296" lastFinishedPulling="2025-10-02 02:01:13.695083578 +0000 UTC m=+862.506830977" observedRunningTime="2025-10-02 02:01:14.018774197 +0000 UTC m=+862.830521596" watchObservedRunningTime="2025-10-02 02:01:14.021513521 +0000 UTC m=+862.833260920" Oct 02 02:01:15 crc kubenswrapper[4885]: I1002 02:01:15.352895 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-557f5d867b-ndfjx" Oct 02 02:01:16 crc kubenswrapper[4885]: I1002 02:01:16.038690 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-9fjl5" event={"ID":"45e282ed-8195-40dc-86cf-1177f3ebeb65","Type":"ContainerStarted","Data":"b097da9f0240711d9eaeb61e76903f1160c4a4a39dd189526a6edb13239f7316"} Oct 02 02:01:16 crc kubenswrapper[4885]: I1002 02:01:16.038996 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-9fjl5" Oct 02 02:01:16 crc kubenswrapper[4885]: I1002 02:01:16.063827 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-9fjl5" podStartSLOduration=4.061732794 podStartE2EDuration="24.063802981s" podCreationTimestamp="2025-10-02 02:00:52 +0000 UTC" firstStartedPulling="2025-10-02 02:00:54.925171962 +0000 UTC m=+843.736919361" lastFinishedPulling="2025-10-02 02:01:14.927242109 +0000 UTC m=+863.738989548" observedRunningTime="2025-10-02 02:01:16.061952245 +0000 UTC m=+864.873699684" watchObservedRunningTime="2025-10-02 02:01:16.063802981 +0000 UTC m=+864.875550410" Oct 02 02:01:23 crc kubenswrapper[4885]: I1002 02:01:23.448782 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-r65hz" Oct 02 02:01:23 crc kubenswrapper[4885]: I1002 02:01:23.639041 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-85777745bb-94284" Oct 02 02:01:23 crc kubenswrapper[4885]: I1002 02:01:23.777718 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-9fjl5" Oct 02 02:01:23 crc kubenswrapper[4885]: I1002 02:01:23.793999 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-6b9957f54f-mb6mf" Oct 02 02:01:24 crc kubenswrapper[4885]: I1002 02:01:24.056776 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-wgfjr" Oct 02 02:01:39 crc kubenswrapper[4885]: I1002 02:01:39.547510 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-tvzhw"] Oct 02 02:01:39 crc kubenswrapper[4885]: I1002 02:01:39.550046 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-tvzhw" Oct 02 02:01:39 crc kubenswrapper[4885]: I1002 02:01:39.555856 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-c56kr" Oct 02 02:01:39 crc kubenswrapper[4885]: I1002 02:01:39.556103 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Oct 02 02:01:39 crc kubenswrapper[4885]: I1002 02:01:39.556324 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Oct 02 02:01:39 crc kubenswrapper[4885]: I1002 02:01:39.556538 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Oct 02 02:01:39 crc kubenswrapper[4885]: I1002 02:01:39.571662 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-tvzhw"] Oct 02 02:01:39 crc kubenswrapper[4885]: I1002 02:01:39.608131 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10deeddc-732a-4154-a74c-b28abc841bc1-config\") pod \"dnsmasq-dns-675f4bcbfc-tvzhw\" (UID: \"10deeddc-732a-4154-a74c-b28abc841bc1\") " pod="openstack/dnsmasq-dns-675f4bcbfc-tvzhw" Oct 02 02:01:39 crc kubenswrapper[4885]: I1002 02:01:39.608253 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gpvqj\" (UniqueName: \"kubernetes.io/projected/10deeddc-732a-4154-a74c-b28abc841bc1-kube-api-access-gpvqj\") pod \"dnsmasq-dns-675f4bcbfc-tvzhw\" (UID: \"10deeddc-732a-4154-a74c-b28abc841bc1\") " pod="openstack/dnsmasq-dns-675f4bcbfc-tvzhw" Oct 02 02:01:39 crc kubenswrapper[4885]: I1002 02:01:39.623563 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-8ljrn"] Oct 02 02:01:39 crc kubenswrapper[4885]: I1002 02:01:39.625095 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-8ljrn" Oct 02 02:01:39 crc kubenswrapper[4885]: I1002 02:01:39.626719 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Oct 02 02:01:39 crc kubenswrapper[4885]: I1002 02:01:39.633092 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-8ljrn"] Oct 02 02:01:39 crc kubenswrapper[4885]: I1002 02:01:39.709677 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5w66h\" (UniqueName: \"kubernetes.io/projected/f541a783-9596-4275-b87e-cf289b223f9d-kube-api-access-5w66h\") pod \"dnsmasq-dns-78dd6ddcc-8ljrn\" (UID: \"f541a783-9596-4275-b87e-cf289b223f9d\") " pod="openstack/dnsmasq-dns-78dd6ddcc-8ljrn" Oct 02 02:01:39 crc kubenswrapper[4885]: I1002 02:01:39.709759 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gpvqj\" (UniqueName: \"kubernetes.io/projected/10deeddc-732a-4154-a74c-b28abc841bc1-kube-api-access-gpvqj\") pod \"dnsmasq-dns-675f4bcbfc-tvzhw\" (UID: \"10deeddc-732a-4154-a74c-b28abc841bc1\") " pod="openstack/dnsmasq-dns-675f4bcbfc-tvzhw" Oct 02 02:01:39 crc kubenswrapper[4885]: I1002 02:01:39.709935 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f541a783-9596-4275-b87e-cf289b223f9d-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-8ljrn\" (UID: \"f541a783-9596-4275-b87e-cf289b223f9d\") " pod="openstack/dnsmasq-dns-78dd6ddcc-8ljrn" Oct 02 02:01:39 crc kubenswrapper[4885]: I1002 02:01:39.709994 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f541a783-9596-4275-b87e-cf289b223f9d-config\") pod \"dnsmasq-dns-78dd6ddcc-8ljrn\" (UID: \"f541a783-9596-4275-b87e-cf289b223f9d\") " pod="openstack/dnsmasq-dns-78dd6ddcc-8ljrn" Oct 02 02:01:39 crc kubenswrapper[4885]: I1002 02:01:39.710049 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10deeddc-732a-4154-a74c-b28abc841bc1-config\") pod \"dnsmasq-dns-675f4bcbfc-tvzhw\" (UID: \"10deeddc-732a-4154-a74c-b28abc841bc1\") " pod="openstack/dnsmasq-dns-675f4bcbfc-tvzhw" Oct 02 02:01:39 crc kubenswrapper[4885]: I1002 02:01:39.711082 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10deeddc-732a-4154-a74c-b28abc841bc1-config\") pod \"dnsmasq-dns-675f4bcbfc-tvzhw\" (UID: \"10deeddc-732a-4154-a74c-b28abc841bc1\") " pod="openstack/dnsmasq-dns-675f4bcbfc-tvzhw" Oct 02 02:01:39 crc kubenswrapper[4885]: I1002 02:01:39.726771 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gpvqj\" (UniqueName: \"kubernetes.io/projected/10deeddc-732a-4154-a74c-b28abc841bc1-kube-api-access-gpvqj\") pod \"dnsmasq-dns-675f4bcbfc-tvzhw\" (UID: \"10deeddc-732a-4154-a74c-b28abc841bc1\") " pod="openstack/dnsmasq-dns-675f4bcbfc-tvzhw" Oct 02 02:01:39 crc kubenswrapper[4885]: I1002 02:01:39.811444 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5w66h\" (UniqueName: \"kubernetes.io/projected/f541a783-9596-4275-b87e-cf289b223f9d-kube-api-access-5w66h\") pod \"dnsmasq-dns-78dd6ddcc-8ljrn\" (UID: \"f541a783-9596-4275-b87e-cf289b223f9d\") " pod="openstack/dnsmasq-dns-78dd6ddcc-8ljrn" Oct 02 02:01:39 crc kubenswrapper[4885]: I1002 02:01:39.811545 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f541a783-9596-4275-b87e-cf289b223f9d-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-8ljrn\" (UID: \"f541a783-9596-4275-b87e-cf289b223f9d\") " pod="openstack/dnsmasq-dns-78dd6ddcc-8ljrn" Oct 02 02:01:39 crc kubenswrapper[4885]: I1002 02:01:39.811572 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f541a783-9596-4275-b87e-cf289b223f9d-config\") pod \"dnsmasq-dns-78dd6ddcc-8ljrn\" (UID: \"f541a783-9596-4275-b87e-cf289b223f9d\") " pod="openstack/dnsmasq-dns-78dd6ddcc-8ljrn" Oct 02 02:01:39 crc kubenswrapper[4885]: I1002 02:01:39.812654 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f541a783-9596-4275-b87e-cf289b223f9d-config\") pod \"dnsmasq-dns-78dd6ddcc-8ljrn\" (UID: \"f541a783-9596-4275-b87e-cf289b223f9d\") " pod="openstack/dnsmasq-dns-78dd6ddcc-8ljrn" Oct 02 02:01:39 crc kubenswrapper[4885]: I1002 02:01:39.812787 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f541a783-9596-4275-b87e-cf289b223f9d-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-8ljrn\" (UID: \"f541a783-9596-4275-b87e-cf289b223f9d\") " pod="openstack/dnsmasq-dns-78dd6ddcc-8ljrn" Oct 02 02:01:39 crc kubenswrapper[4885]: I1002 02:01:39.830390 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5w66h\" (UniqueName: \"kubernetes.io/projected/f541a783-9596-4275-b87e-cf289b223f9d-kube-api-access-5w66h\") pod \"dnsmasq-dns-78dd6ddcc-8ljrn\" (UID: \"f541a783-9596-4275-b87e-cf289b223f9d\") " pod="openstack/dnsmasq-dns-78dd6ddcc-8ljrn" Oct 02 02:01:39 crc kubenswrapper[4885]: I1002 02:01:39.874237 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-tvzhw" Oct 02 02:01:39 crc kubenswrapper[4885]: I1002 02:01:39.955826 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-8ljrn" Oct 02 02:01:40 crc kubenswrapper[4885]: I1002 02:01:40.196478 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-8ljrn"] Oct 02 02:01:40 crc kubenswrapper[4885]: I1002 02:01:40.205319 4885 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 02:01:40 crc kubenswrapper[4885]: I1002 02:01:40.295453 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-8ljrn" event={"ID":"f541a783-9596-4275-b87e-cf289b223f9d","Type":"ContainerStarted","Data":"6daa8fbbc09a7b09ecc20214ccff27d937a0cf770c37df7a629593e92a06b7be"} Oct 02 02:01:40 crc kubenswrapper[4885]: I1002 02:01:40.341858 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-tvzhw"] Oct 02 02:01:40 crc kubenswrapper[4885]: W1002 02:01:40.355301 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod10deeddc_732a_4154_a74c_b28abc841bc1.slice/crio-ec967b0b3cc64faa97c93ccaee1a4381e0d8c51f24e760935bc1cfa152d554bf WatchSource:0}: Error finding container ec967b0b3cc64faa97c93ccaee1a4381e0d8c51f24e760935bc1cfa152d554bf: Status 404 returned error can't find the container with id ec967b0b3cc64faa97c93ccaee1a4381e0d8c51f24e760935bc1cfa152d554bf Oct 02 02:01:41 crc kubenswrapper[4885]: I1002 02:01:41.310500 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-tvzhw" event={"ID":"10deeddc-732a-4154-a74c-b28abc841bc1","Type":"ContainerStarted","Data":"ec967b0b3cc64faa97c93ccaee1a4381e0d8c51f24e760935bc1cfa152d554bf"} Oct 02 02:01:42 crc kubenswrapper[4885]: I1002 02:01:42.606424 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-tvzhw"] Oct 02 02:01:42 crc kubenswrapper[4885]: I1002 02:01:42.641616 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-5wrwj"] Oct 02 02:01:42 crc kubenswrapper[4885]: I1002 02:01:42.642949 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-5wrwj" Oct 02 02:01:42 crc kubenswrapper[4885]: I1002 02:01:42.652417 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-5wrwj"] Oct 02 02:01:42 crc kubenswrapper[4885]: I1002 02:01:42.659395 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/32a76623-c272-48c1-85e6-2170cbb017cc-config\") pod \"dnsmasq-dns-5ccc8479f9-5wrwj\" (UID: \"32a76623-c272-48c1-85e6-2170cbb017cc\") " pod="openstack/dnsmasq-dns-5ccc8479f9-5wrwj" Oct 02 02:01:42 crc kubenswrapper[4885]: I1002 02:01:42.659435 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gf6c6\" (UniqueName: \"kubernetes.io/projected/32a76623-c272-48c1-85e6-2170cbb017cc-kube-api-access-gf6c6\") pod \"dnsmasq-dns-5ccc8479f9-5wrwj\" (UID: \"32a76623-c272-48c1-85e6-2170cbb017cc\") " pod="openstack/dnsmasq-dns-5ccc8479f9-5wrwj" Oct 02 02:01:42 crc kubenswrapper[4885]: I1002 02:01:42.659470 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/32a76623-c272-48c1-85e6-2170cbb017cc-dns-svc\") pod \"dnsmasq-dns-5ccc8479f9-5wrwj\" (UID: \"32a76623-c272-48c1-85e6-2170cbb017cc\") " pod="openstack/dnsmasq-dns-5ccc8479f9-5wrwj" Oct 02 02:01:42 crc kubenswrapper[4885]: I1002 02:01:42.760191 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/32a76623-c272-48c1-85e6-2170cbb017cc-dns-svc\") pod \"dnsmasq-dns-5ccc8479f9-5wrwj\" (UID: \"32a76623-c272-48c1-85e6-2170cbb017cc\") " pod="openstack/dnsmasq-dns-5ccc8479f9-5wrwj" Oct 02 02:01:42 crc kubenswrapper[4885]: I1002 02:01:42.760304 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/32a76623-c272-48c1-85e6-2170cbb017cc-config\") pod \"dnsmasq-dns-5ccc8479f9-5wrwj\" (UID: \"32a76623-c272-48c1-85e6-2170cbb017cc\") " pod="openstack/dnsmasq-dns-5ccc8479f9-5wrwj" Oct 02 02:01:42 crc kubenswrapper[4885]: I1002 02:01:42.760330 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gf6c6\" (UniqueName: \"kubernetes.io/projected/32a76623-c272-48c1-85e6-2170cbb017cc-kube-api-access-gf6c6\") pod \"dnsmasq-dns-5ccc8479f9-5wrwj\" (UID: \"32a76623-c272-48c1-85e6-2170cbb017cc\") " pod="openstack/dnsmasq-dns-5ccc8479f9-5wrwj" Oct 02 02:01:42 crc kubenswrapper[4885]: I1002 02:01:42.761561 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/32a76623-c272-48c1-85e6-2170cbb017cc-config\") pod \"dnsmasq-dns-5ccc8479f9-5wrwj\" (UID: \"32a76623-c272-48c1-85e6-2170cbb017cc\") " pod="openstack/dnsmasq-dns-5ccc8479f9-5wrwj" Oct 02 02:01:42 crc kubenswrapper[4885]: I1002 02:01:42.761870 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/32a76623-c272-48c1-85e6-2170cbb017cc-dns-svc\") pod \"dnsmasq-dns-5ccc8479f9-5wrwj\" (UID: \"32a76623-c272-48c1-85e6-2170cbb017cc\") " pod="openstack/dnsmasq-dns-5ccc8479f9-5wrwj" Oct 02 02:01:42 crc kubenswrapper[4885]: I1002 02:01:42.783598 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gf6c6\" (UniqueName: \"kubernetes.io/projected/32a76623-c272-48c1-85e6-2170cbb017cc-kube-api-access-gf6c6\") pod \"dnsmasq-dns-5ccc8479f9-5wrwj\" (UID: \"32a76623-c272-48c1-85e6-2170cbb017cc\") " pod="openstack/dnsmasq-dns-5ccc8479f9-5wrwj" Oct 02 02:01:42 crc kubenswrapper[4885]: I1002 02:01:42.865725 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-8ljrn"] Oct 02 02:01:42 crc kubenswrapper[4885]: I1002 02:01:42.889016 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-p9cqp"] Oct 02 02:01:42 crc kubenswrapper[4885]: I1002 02:01:42.890134 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-p9cqp" Oct 02 02:01:42 crc kubenswrapper[4885]: I1002 02:01:42.918511 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-p9cqp"] Oct 02 02:01:42 crc kubenswrapper[4885]: I1002 02:01:42.973316 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-5wrwj" Oct 02 02:01:43 crc kubenswrapper[4885]: I1002 02:01:43.064616 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2tgx5\" (UniqueName: \"kubernetes.io/projected/90626475-ee00-431f-8992-d5fc20728c6c-kube-api-access-2tgx5\") pod \"dnsmasq-dns-57d769cc4f-p9cqp\" (UID: \"90626475-ee00-431f-8992-d5fc20728c6c\") " pod="openstack/dnsmasq-dns-57d769cc4f-p9cqp" Oct 02 02:01:43 crc kubenswrapper[4885]: I1002 02:01:43.064674 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/90626475-ee00-431f-8992-d5fc20728c6c-config\") pod \"dnsmasq-dns-57d769cc4f-p9cqp\" (UID: \"90626475-ee00-431f-8992-d5fc20728c6c\") " pod="openstack/dnsmasq-dns-57d769cc4f-p9cqp" Oct 02 02:01:43 crc kubenswrapper[4885]: I1002 02:01:43.064721 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/90626475-ee00-431f-8992-d5fc20728c6c-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-p9cqp\" (UID: \"90626475-ee00-431f-8992-d5fc20728c6c\") " pod="openstack/dnsmasq-dns-57d769cc4f-p9cqp" Oct 02 02:01:43 crc kubenswrapper[4885]: I1002 02:01:43.166514 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2tgx5\" (UniqueName: \"kubernetes.io/projected/90626475-ee00-431f-8992-d5fc20728c6c-kube-api-access-2tgx5\") pod \"dnsmasq-dns-57d769cc4f-p9cqp\" (UID: \"90626475-ee00-431f-8992-d5fc20728c6c\") " pod="openstack/dnsmasq-dns-57d769cc4f-p9cqp" Oct 02 02:01:43 crc kubenswrapper[4885]: I1002 02:01:43.167318 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/90626475-ee00-431f-8992-d5fc20728c6c-config\") pod \"dnsmasq-dns-57d769cc4f-p9cqp\" (UID: \"90626475-ee00-431f-8992-d5fc20728c6c\") " pod="openstack/dnsmasq-dns-57d769cc4f-p9cqp" Oct 02 02:01:43 crc kubenswrapper[4885]: I1002 02:01:43.167440 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/90626475-ee00-431f-8992-d5fc20728c6c-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-p9cqp\" (UID: \"90626475-ee00-431f-8992-d5fc20728c6c\") " pod="openstack/dnsmasq-dns-57d769cc4f-p9cqp" Oct 02 02:01:43 crc kubenswrapper[4885]: I1002 02:01:43.169992 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/90626475-ee00-431f-8992-d5fc20728c6c-config\") pod \"dnsmasq-dns-57d769cc4f-p9cqp\" (UID: \"90626475-ee00-431f-8992-d5fc20728c6c\") " pod="openstack/dnsmasq-dns-57d769cc4f-p9cqp" Oct 02 02:01:43 crc kubenswrapper[4885]: I1002 02:01:43.170047 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/90626475-ee00-431f-8992-d5fc20728c6c-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-p9cqp\" (UID: \"90626475-ee00-431f-8992-d5fc20728c6c\") " pod="openstack/dnsmasq-dns-57d769cc4f-p9cqp" Oct 02 02:01:43 crc kubenswrapper[4885]: I1002 02:01:43.187724 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2tgx5\" (UniqueName: \"kubernetes.io/projected/90626475-ee00-431f-8992-d5fc20728c6c-kube-api-access-2tgx5\") pod \"dnsmasq-dns-57d769cc4f-p9cqp\" (UID: \"90626475-ee00-431f-8992-d5fc20728c6c\") " pod="openstack/dnsmasq-dns-57d769cc4f-p9cqp" Oct 02 02:01:43 crc kubenswrapper[4885]: I1002 02:01:43.213102 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-p9cqp" Oct 02 02:01:43 crc kubenswrapper[4885]: I1002 02:01:43.522830 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-5wrwj"] Oct 02 02:01:43 crc kubenswrapper[4885]: W1002 02:01:43.529177 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod32a76623_c272_48c1_85e6_2170cbb017cc.slice/crio-a75ce6ed97ce5e957c0426fdc15f2a177449241a425a66e15fb3b8edc6443ad6 WatchSource:0}: Error finding container a75ce6ed97ce5e957c0426fdc15f2a177449241a425a66e15fb3b8edc6443ad6: Status 404 returned error can't find the container with id a75ce6ed97ce5e957c0426fdc15f2a177449241a425a66e15fb3b8edc6443ad6 Oct 02 02:01:43 crc kubenswrapper[4885]: I1002 02:01:43.708880 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-p9cqp"] Oct 02 02:01:43 crc kubenswrapper[4885]: I1002 02:01:43.751433 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 02 02:01:43 crc kubenswrapper[4885]: I1002 02:01:43.755280 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:01:43 crc kubenswrapper[4885]: I1002 02:01:43.758389 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 02 02:01:43 crc kubenswrapper[4885]: I1002 02:01:43.758578 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 02 02:01:43 crc kubenswrapper[4885]: I1002 02:01:43.758908 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Oct 02 02:01:43 crc kubenswrapper[4885]: I1002 02:01:43.759033 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-mghd5" Oct 02 02:01:43 crc kubenswrapper[4885]: I1002 02:01:43.759138 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 02 02:01:43 crc kubenswrapper[4885]: I1002 02:01:43.759255 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Oct 02 02:01:43 crc kubenswrapper[4885]: I1002 02:01:43.759736 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 02 02:01:43 crc kubenswrapper[4885]: I1002 02:01:43.764009 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 02 02:01:43 crc kubenswrapper[4885]: I1002 02:01:43.879403 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/eea66006-67ba-45e3-bdf9-9d144d772386-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"eea66006-67ba-45e3-bdf9-9d144d772386\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:01:43 crc kubenswrapper[4885]: I1002 02:01:43.879446 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/eea66006-67ba-45e3-bdf9-9d144d772386-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"eea66006-67ba-45e3-bdf9-9d144d772386\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:01:43 crc kubenswrapper[4885]: I1002 02:01:43.879475 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/eea66006-67ba-45e3-bdf9-9d144d772386-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"eea66006-67ba-45e3-bdf9-9d144d772386\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:01:43 crc kubenswrapper[4885]: I1002 02:01:43.879557 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/eea66006-67ba-45e3-bdf9-9d144d772386-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"eea66006-67ba-45e3-bdf9-9d144d772386\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:01:43 crc kubenswrapper[4885]: I1002 02:01:43.879573 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9c4hd\" (UniqueName: \"kubernetes.io/projected/eea66006-67ba-45e3-bdf9-9d144d772386-kube-api-access-9c4hd\") pod \"rabbitmq-cell1-server-0\" (UID: \"eea66006-67ba-45e3-bdf9-9d144d772386\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:01:43 crc kubenswrapper[4885]: I1002 02:01:43.879599 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/eea66006-67ba-45e3-bdf9-9d144d772386-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"eea66006-67ba-45e3-bdf9-9d144d772386\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:01:43 crc kubenswrapper[4885]: I1002 02:01:43.879619 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/eea66006-67ba-45e3-bdf9-9d144d772386-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"eea66006-67ba-45e3-bdf9-9d144d772386\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:01:43 crc kubenswrapper[4885]: I1002 02:01:43.879636 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"eea66006-67ba-45e3-bdf9-9d144d772386\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:01:43 crc kubenswrapper[4885]: I1002 02:01:43.879653 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/eea66006-67ba-45e3-bdf9-9d144d772386-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"eea66006-67ba-45e3-bdf9-9d144d772386\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:01:43 crc kubenswrapper[4885]: I1002 02:01:43.879680 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/eea66006-67ba-45e3-bdf9-9d144d772386-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"eea66006-67ba-45e3-bdf9-9d144d772386\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:01:43 crc kubenswrapper[4885]: I1002 02:01:43.879698 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/eea66006-67ba-45e3-bdf9-9d144d772386-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"eea66006-67ba-45e3-bdf9-9d144d772386\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:01:43 crc kubenswrapper[4885]: I1002 02:01:43.981695 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/eea66006-67ba-45e3-bdf9-9d144d772386-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"eea66006-67ba-45e3-bdf9-9d144d772386\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:01:43 crc kubenswrapper[4885]: I1002 02:01:43.981734 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"eea66006-67ba-45e3-bdf9-9d144d772386\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:01:43 crc kubenswrapper[4885]: I1002 02:01:43.981760 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/eea66006-67ba-45e3-bdf9-9d144d772386-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"eea66006-67ba-45e3-bdf9-9d144d772386\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:01:43 crc kubenswrapper[4885]: I1002 02:01:43.981795 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/eea66006-67ba-45e3-bdf9-9d144d772386-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"eea66006-67ba-45e3-bdf9-9d144d772386\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:01:43 crc kubenswrapper[4885]: I1002 02:01:43.981813 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/eea66006-67ba-45e3-bdf9-9d144d772386-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"eea66006-67ba-45e3-bdf9-9d144d772386\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:01:43 crc kubenswrapper[4885]: I1002 02:01:43.981843 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/eea66006-67ba-45e3-bdf9-9d144d772386-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"eea66006-67ba-45e3-bdf9-9d144d772386\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:01:43 crc kubenswrapper[4885]: I1002 02:01:43.981863 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/eea66006-67ba-45e3-bdf9-9d144d772386-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"eea66006-67ba-45e3-bdf9-9d144d772386\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:01:43 crc kubenswrapper[4885]: I1002 02:01:43.981886 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/eea66006-67ba-45e3-bdf9-9d144d772386-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"eea66006-67ba-45e3-bdf9-9d144d772386\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:01:43 crc kubenswrapper[4885]: I1002 02:01:43.981934 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/eea66006-67ba-45e3-bdf9-9d144d772386-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"eea66006-67ba-45e3-bdf9-9d144d772386\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:01:43 crc kubenswrapper[4885]: I1002 02:01:43.981954 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9c4hd\" (UniqueName: \"kubernetes.io/projected/eea66006-67ba-45e3-bdf9-9d144d772386-kube-api-access-9c4hd\") pod \"rabbitmq-cell1-server-0\" (UID: \"eea66006-67ba-45e3-bdf9-9d144d772386\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:01:43 crc kubenswrapper[4885]: I1002 02:01:43.981978 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/eea66006-67ba-45e3-bdf9-9d144d772386-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"eea66006-67ba-45e3-bdf9-9d144d772386\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:01:43 crc kubenswrapper[4885]: I1002 02:01:43.983470 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/eea66006-67ba-45e3-bdf9-9d144d772386-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"eea66006-67ba-45e3-bdf9-9d144d772386\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:01:43 crc kubenswrapper[4885]: I1002 02:01:43.983493 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/eea66006-67ba-45e3-bdf9-9d144d772386-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"eea66006-67ba-45e3-bdf9-9d144d772386\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:01:43 crc kubenswrapper[4885]: I1002 02:01:43.984988 4885 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"eea66006-67ba-45e3-bdf9-9d144d772386\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:01:43 crc kubenswrapper[4885]: I1002 02:01:43.985421 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/eea66006-67ba-45e3-bdf9-9d144d772386-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"eea66006-67ba-45e3-bdf9-9d144d772386\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:01:43 crc kubenswrapper[4885]: I1002 02:01:43.985462 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/eea66006-67ba-45e3-bdf9-9d144d772386-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"eea66006-67ba-45e3-bdf9-9d144d772386\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:01:43 crc kubenswrapper[4885]: I1002 02:01:43.990034 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/eea66006-67ba-45e3-bdf9-9d144d772386-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"eea66006-67ba-45e3-bdf9-9d144d772386\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:01:43 crc kubenswrapper[4885]: I1002 02:01:43.990580 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/eea66006-67ba-45e3-bdf9-9d144d772386-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"eea66006-67ba-45e3-bdf9-9d144d772386\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:01:43 crc kubenswrapper[4885]: I1002 02:01:43.990659 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/eea66006-67ba-45e3-bdf9-9d144d772386-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"eea66006-67ba-45e3-bdf9-9d144d772386\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:01:43 crc kubenswrapper[4885]: I1002 02:01:43.993883 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/eea66006-67ba-45e3-bdf9-9d144d772386-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"eea66006-67ba-45e3-bdf9-9d144d772386\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:01:43 crc kubenswrapper[4885]: I1002 02:01:43.995761 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/eea66006-67ba-45e3-bdf9-9d144d772386-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"eea66006-67ba-45e3-bdf9-9d144d772386\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:01:44 crc kubenswrapper[4885]: I1002 02:01:44.008060 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"eea66006-67ba-45e3-bdf9-9d144d772386\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:01:44 crc kubenswrapper[4885]: I1002 02:01:44.008856 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9c4hd\" (UniqueName: \"kubernetes.io/projected/eea66006-67ba-45e3-bdf9-9d144d772386-kube-api-access-9c4hd\") pod \"rabbitmq-cell1-server-0\" (UID: \"eea66006-67ba-45e3-bdf9-9d144d772386\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:01:44 crc kubenswrapper[4885]: I1002 02:01:44.038144 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 02 02:01:44 crc kubenswrapper[4885]: I1002 02:01:44.039809 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 02 02:01:44 crc kubenswrapper[4885]: I1002 02:01:44.042912 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Oct 02 02:01:44 crc kubenswrapper[4885]: I1002 02:01:44.043112 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 02 02:01:44 crc kubenswrapper[4885]: I1002 02:01:44.043225 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Oct 02 02:01:44 crc kubenswrapper[4885]: I1002 02:01:44.043622 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-2stht" Oct 02 02:01:44 crc kubenswrapper[4885]: I1002 02:01:44.043766 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 02 02:01:44 crc kubenswrapper[4885]: I1002 02:01:44.043874 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 02 02:01:44 crc kubenswrapper[4885]: I1002 02:01:44.043967 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 02 02:01:44 crc kubenswrapper[4885]: I1002 02:01:44.068099 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 02 02:01:44 crc kubenswrapper[4885]: I1002 02:01:44.141542 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:01:44 crc kubenswrapper[4885]: I1002 02:01:44.185646 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/6692f52c-5ee9-4cd6-8674-0fb3f5fd6679-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"6692f52c-5ee9-4cd6-8674-0fb3f5fd6679\") " pod="openstack/rabbitmq-server-0" Oct 02 02:01:44 crc kubenswrapper[4885]: I1002 02:01:44.185702 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/6692f52c-5ee9-4cd6-8674-0fb3f5fd6679-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"6692f52c-5ee9-4cd6-8674-0fb3f5fd6679\") " pod="openstack/rabbitmq-server-0" Oct 02 02:01:44 crc kubenswrapper[4885]: I1002 02:01:44.185796 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"6692f52c-5ee9-4cd6-8674-0fb3f5fd6679\") " pod="openstack/rabbitmq-server-0" Oct 02 02:01:44 crc kubenswrapper[4885]: I1002 02:01:44.185820 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6692f52c-5ee9-4cd6-8674-0fb3f5fd6679-config-data\") pod \"rabbitmq-server-0\" (UID: \"6692f52c-5ee9-4cd6-8674-0fb3f5fd6679\") " pod="openstack/rabbitmq-server-0" Oct 02 02:01:44 crc kubenswrapper[4885]: I1002 02:01:44.185848 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/6692f52c-5ee9-4cd6-8674-0fb3f5fd6679-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"6692f52c-5ee9-4cd6-8674-0fb3f5fd6679\") " pod="openstack/rabbitmq-server-0" Oct 02 02:01:44 crc kubenswrapper[4885]: I1002 02:01:44.185879 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/6692f52c-5ee9-4cd6-8674-0fb3f5fd6679-server-conf\") pod \"rabbitmq-server-0\" (UID: \"6692f52c-5ee9-4cd6-8674-0fb3f5fd6679\") " pod="openstack/rabbitmq-server-0" Oct 02 02:01:44 crc kubenswrapper[4885]: I1002 02:01:44.185902 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tlkrh\" (UniqueName: \"kubernetes.io/projected/6692f52c-5ee9-4cd6-8674-0fb3f5fd6679-kube-api-access-tlkrh\") pod \"rabbitmq-server-0\" (UID: \"6692f52c-5ee9-4cd6-8674-0fb3f5fd6679\") " pod="openstack/rabbitmq-server-0" Oct 02 02:01:44 crc kubenswrapper[4885]: I1002 02:01:44.185928 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/6692f52c-5ee9-4cd6-8674-0fb3f5fd6679-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"6692f52c-5ee9-4cd6-8674-0fb3f5fd6679\") " pod="openstack/rabbitmq-server-0" Oct 02 02:01:44 crc kubenswrapper[4885]: I1002 02:01:44.185962 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/6692f52c-5ee9-4cd6-8674-0fb3f5fd6679-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"6692f52c-5ee9-4cd6-8674-0fb3f5fd6679\") " pod="openstack/rabbitmq-server-0" Oct 02 02:01:44 crc kubenswrapper[4885]: I1002 02:01:44.185998 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/6692f52c-5ee9-4cd6-8674-0fb3f5fd6679-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"6692f52c-5ee9-4cd6-8674-0fb3f5fd6679\") " pod="openstack/rabbitmq-server-0" Oct 02 02:01:44 crc kubenswrapper[4885]: I1002 02:01:44.186019 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/6692f52c-5ee9-4cd6-8674-0fb3f5fd6679-pod-info\") pod \"rabbitmq-server-0\" (UID: \"6692f52c-5ee9-4cd6-8674-0fb3f5fd6679\") " pod="openstack/rabbitmq-server-0" Oct 02 02:01:44 crc kubenswrapper[4885]: I1002 02:01:44.290094 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/6692f52c-5ee9-4cd6-8674-0fb3f5fd6679-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"6692f52c-5ee9-4cd6-8674-0fb3f5fd6679\") " pod="openstack/rabbitmq-server-0" Oct 02 02:01:44 crc kubenswrapper[4885]: I1002 02:01:44.290297 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/6692f52c-5ee9-4cd6-8674-0fb3f5fd6679-server-conf\") pod \"rabbitmq-server-0\" (UID: \"6692f52c-5ee9-4cd6-8674-0fb3f5fd6679\") " pod="openstack/rabbitmq-server-0" Oct 02 02:01:44 crc kubenswrapper[4885]: I1002 02:01:44.290326 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tlkrh\" (UniqueName: \"kubernetes.io/projected/6692f52c-5ee9-4cd6-8674-0fb3f5fd6679-kube-api-access-tlkrh\") pod \"rabbitmq-server-0\" (UID: \"6692f52c-5ee9-4cd6-8674-0fb3f5fd6679\") " pod="openstack/rabbitmq-server-0" Oct 02 02:01:44 crc kubenswrapper[4885]: I1002 02:01:44.290365 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/6692f52c-5ee9-4cd6-8674-0fb3f5fd6679-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"6692f52c-5ee9-4cd6-8674-0fb3f5fd6679\") " pod="openstack/rabbitmq-server-0" Oct 02 02:01:44 crc kubenswrapper[4885]: I1002 02:01:44.290428 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/6692f52c-5ee9-4cd6-8674-0fb3f5fd6679-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"6692f52c-5ee9-4cd6-8674-0fb3f5fd6679\") " pod="openstack/rabbitmq-server-0" Oct 02 02:01:44 crc kubenswrapper[4885]: I1002 02:01:44.290497 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/6692f52c-5ee9-4cd6-8674-0fb3f5fd6679-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"6692f52c-5ee9-4cd6-8674-0fb3f5fd6679\") " pod="openstack/rabbitmq-server-0" Oct 02 02:01:44 crc kubenswrapper[4885]: I1002 02:01:44.290513 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/6692f52c-5ee9-4cd6-8674-0fb3f5fd6679-pod-info\") pod \"rabbitmq-server-0\" (UID: \"6692f52c-5ee9-4cd6-8674-0fb3f5fd6679\") " pod="openstack/rabbitmq-server-0" Oct 02 02:01:44 crc kubenswrapper[4885]: I1002 02:01:44.290583 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/6692f52c-5ee9-4cd6-8674-0fb3f5fd6679-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"6692f52c-5ee9-4cd6-8674-0fb3f5fd6679\") " pod="openstack/rabbitmq-server-0" Oct 02 02:01:44 crc kubenswrapper[4885]: I1002 02:01:44.290611 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/6692f52c-5ee9-4cd6-8674-0fb3f5fd6679-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"6692f52c-5ee9-4cd6-8674-0fb3f5fd6679\") " pod="openstack/rabbitmq-server-0" Oct 02 02:01:44 crc kubenswrapper[4885]: I1002 02:01:44.290761 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"6692f52c-5ee9-4cd6-8674-0fb3f5fd6679\") " pod="openstack/rabbitmq-server-0" Oct 02 02:01:44 crc kubenswrapper[4885]: I1002 02:01:44.290777 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6692f52c-5ee9-4cd6-8674-0fb3f5fd6679-config-data\") pod \"rabbitmq-server-0\" (UID: \"6692f52c-5ee9-4cd6-8674-0fb3f5fd6679\") " pod="openstack/rabbitmq-server-0" Oct 02 02:01:44 crc kubenswrapper[4885]: I1002 02:01:44.291344 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/6692f52c-5ee9-4cd6-8674-0fb3f5fd6679-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"6692f52c-5ee9-4cd6-8674-0fb3f5fd6679\") " pod="openstack/rabbitmq-server-0" Oct 02 02:01:44 crc kubenswrapper[4885]: I1002 02:01:44.291691 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6692f52c-5ee9-4cd6-8674-0fb3f5fd6679-config-data\") pod \"rabbitmq-server-0\" (UID: \"6692f52c-5ee9-4cd6-8674-0fb3f5fd6679\") " pod="openstack/rabbitmq-server-0" Oct 02 02:01:44 crc kubenswrapper[4885]: I1002 02:01:44.291875 4885 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"6692f52c-5ee9-4cd6-8674-0fb3f5fd6679\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/rabbitmq-server-0" Oct 02 02:01:44 crc kubenswrapper[4885]: I1002 02:01:44.292596 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/6692f52c-5ee9-4cd6-8674-0fb3f5fd6679-server-conf\") pod \"rabbitmq-server-0\" (UID: \"6692f52c-5ee9-4cd6-8674-0fb3f5fd6679\") " pod="openstack/rabbitmq-server-0" Oct 02 02:01:44 crc kubenswrapper[4885]: I1002 02:01:44.295455 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/6692f52c-5ee9-4cd6-8674-0fb3f5fd6679-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"6692f52c-5ee9-4cd6-8674-0fb3f5fd6679\") " pod="openstack/rabbitmq-server-0" Oct 02 02:01:44 crc kubenswrapper[4885]: I1002 02:01:44.295679 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/6692f52c-5ee9-4cd6-8674-0fb3f5fd6679-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"6692f52c-5ee9-4cd6-8674-0fb3f5fd6679\") " pod="openstack/rabbitmq-server-0" Oct 02 02:01:44 crc kubenswrapper[4885]: I1002 02:01:44.297020 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/6692f52c-5ee9-4cd6-8674-0fb3f5fd6679-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"6692f52c-5ee9-4cd6-8674-0fb3f5fd6679\") " pod="openstack/rabbitmq-server-0" Oct 02 02:01:44 crc kubenswrapper[4885]: I1002 02:01:44.297803 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/6692f52c-5ee9-4cd6-8674-0fb3f5fd6679-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"6692f52c-5ee9-4cd6-8674-0fb3f5fd6679\") " pod="openstack/rabbitmq-server-0" Oct 02 02:01:44 crc kubenswrapper[4885]: I1002 02:01:44.306540 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/6692f52c-5ee9-4cd6-8674-0fb3f5fd6679-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"6692f52c-5ee9-4cd6-8674-0fb3f5fd6679\") " pod="openstack/rabbitmq-server-0" Oct 02 02:01:44 crc kubenswrapper[4885]: I1002 02:01:44.307142 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/6692f52c-5ee9-4cd6-8674-0fb3f5fd6679-pod-info\") pod \"rabbitmq-server-0\" (UID: \"6692f52c-5ee9-4cd6-8674-0fb3f5fd6679\") " pod="openstack/rabbitmq-server-0" Oct 02 02:01:44 crc kubenswrapper[4885]: I1002 02:01:44.314336 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tlkrh\" (UniqueName: \"kubernetes.io/projected/6692f52c-5ee9-4cd6-8674-0fb3f5fd6679-kube-api-access-tlkrh\") pod \"rabbitmq-server-0\" (UID: \"6692f52c-5ee9-4cd6-8674-0fb3f5fd6679\") " pod="openstack/rabbitmq-server-0" Oct 02 02:01:44 crc kubenswrapper[4885]: I1002 02:01:44.320793 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"6692f52c-5ee9-4cd6-8674-0fb3f5fd6679\") " pod="openstack/rabbitmq-server-0" Oct 02 02:01:44 crc kubenswrapper[4885]: I1002 02:01:44.343934 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-5wrwj" event={"ID":"32a76623-c272-48c1-85e6-2170cbb017cc","Type":"ContainerStarted","Data":"a75ce6ed97ce5e957c0426fdc15f2a177449241a425a66e15fb3b8edc6443ad6"} Oct 02 02:01:44 crc kubenswrapper[4885]: I1002 02:01:44.363771 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 02 02:01:46 crc kubenswrapper[4885]: I1002 02:01:46.877785 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 02 02:01:46 crc kubenswrapper[4885]: I1002 02:01:46.879953 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 02 02:01:46 crc kubenswrapper[4885]: I1002 02:01:46.882463 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Oct 02 02:01:46 crc kubenswrapper[4885]: I1002 02:01:46.882796 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Oct 02 02:01:46 crc kubenswrapper[4885]: I1002 02:01:46.883364 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-9pbhx" Oct 02 02:01:46 crc kubenswrapper[4885]: I1002 02:01:46.883595 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Oct 02 02:01:46 crc kubenswrapper[4885]: I1002 02:01:46.885876 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Oct 02 02:01:46 crc kubenswrapper[4885]: I1002 02:01:46.891365 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 02 02:01:46 crc kubenswrapper[4885]: I1002 02:01:46.899788 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Oct 02 02:01:47 crc kubenswrapper[4885]: I1002 02:01:47.004823 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Oct 02 02:01:47 crc kubenswrapper[4885]: I1002 02:01:47.006091 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 02 02:01:47 crc kubenswrapper[4885]: I1002 02:01:47.007760 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Oct 02 02:01:47 crc kubenswrapper[4885]: I1002 02:01:47.008516 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-f2rn5" Oct 02 02:01:47 crc kubenswrapper[4885]: I1002 02:01:47.008792 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Oct 02 02:01:47 crc kubenswrapper[4885]: I1002 02:01:47.010805 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Oct 02 02:01:47 crc kubenswrapper[4885]: I1002 02:01:47.031682 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 02 02:01:47 crc kubenswrapper[4885]: I1002 02:01:47.041181 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/87196135-8d40-47cb-b918-98aea04c7c17-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"87196135-8d40-47cb-b918-98aea04c7c17\") " pod="openstack/openstack-cell1-galera-0" Oct 02 02:01:47 crc kubenswrapper[4885]: I1002 02:01:47.041241 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-cell1-galera-0\" (UID: \"87196135-8d40-47cb-b918-98aea04c7c17\") " pod="openstack/openstack-cell1-galera-0" Oct 02 02:01:47 crc kubenswrapper[4885]: I1002 02:01:47.041320 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/87196135-8d40-47cb-b918-98aea04c7c17-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"87196135-8d40-47cb-b918-98aea04c7c17\") " pod="openstack/openstack-cell1-galera-0" Oct 02 02:01:47 crc kubenswrapper[4885]: I1002 02:01:47.041348 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/87196135-8d40-47cb-b918-98aea04c7c17-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"87196135-8d40-47cb-b918-98aea04c7c17\") " pod="openstack/openstack-cell1-galera-0" Oct 02 02:01:47 crc kubenswrapper[4885]: I1002 02:01:47.041434 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/87196135-8d40-47cb-b918-98aea04c7c17-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"87196135-8d40-47cb-b918-98aea04c7c17\") " pod="openstack/openstack-cell1-galera-0" Oct 02 02:01:47 crc kubenswrapper[4885]: I1002 02:01:47.041486 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hjpjc\" (UniqueName: \"kubernetes.io/projected/87196135-8d40-47cb-b918-98aea04c7c17-kube-api-access-hjpjc\") pod \"openstack-cell1-galera-0\" (UID: \"87196135-8d40-47cb-b918-98aea04c7c17\") " pod="openstack/openstack-cell1-galera-0" Oct 02 02:01:47 crc kubenswrapper[4885]: I1002 02:01:47.041523 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/87196135-8d40-47cb-b918-98aea04c7c17-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"87196135-8d40-47cb-b918-98aea04c7c17\") " pod="openstack/openstack-cell1-galera-0" Oct 02 02:01:47 crc kubenswrapper[4885]: I1002 02:01:47.041596 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/87196135-8d40-47cb-b918-98aea04c7c17-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"87196135-8d40-47cb-b918-98aea04c7c17\") " pod="openstack/openstack-cell1-galera-0" Oct 02 02:01:47 crc kubenswrapper[4885]: I1002 02:01:47.041618 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87196135-8d40-47cb-b918-98aea04c7c17-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"87196135-8d40-47cb-b918-98aea04c7c17\") " pod="openstack/openstack-cell1-galera-0" Oct 02 02:01:47 crc kubenswrapper[4885]: W1002 02:01:47.052639 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod90626475_ee00_431f_8992_d5fc20728c6c.slice/crio-b9202f38c3e6b86dff94c2d73561c7e9d89082b2a9efa38a64f9bfd18a8c4c84 WatchSource:0}: Error finding container b9202f38c3e6b86dff94c2d73561c7e9d89082b2a9efa38a64f9bfd18a8c4c84: Status 404 returned error can't find the container with id b9202f38c3e6b86dff94c2d73561c7e9d89082b2a9efa38a64f9bfd18a8c4c84 Oct 02 02:01:47 crc kubenswrapper[4885]: I1002 02:01:47.143907 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8427e298-965b-48f0-821f-bd078fa2e96f-operator-scripts\") pod \"openstack-galera-0\" (UID: \"8427e298-965b-48f0-821f-bd078fa2e96f\") " pod="openstack/openstack-galera-0" Oct 02 02:01:47 crc kubenswrapper[4885]: I1002 02:01:47.144015 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/87196135-8d40-47cb-b918-98aea04c7c17-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"87196135-8d40-47cb-b918-98aea04c7c17\") " pod="openstack/openstack-cell1-galera-0" Oct 02 02:01:47 crc kubenswrapper[4885]: I1002 02:01:47.144070 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-cell1-galera-0\" (UID: \"87196135-8d40-47cb-b918-98aea04c7c17\") " pod="openstack/openstack-cell1-galera-0" Oct 02 02:01:47 crc kubenswrapper[4885]: I1002 02:01:47.144103 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/8427e298-965b-48f0-821f-bd078fa2e96f-secrets\") pod \"openstack-galera-0\" (UID: \"8427e298-965b-48f0-821f-bd078fa2e96f\") " pod="openstack/openstack-galera-0" Oct 02 02:01:47 crc kubenswrapper[4885]: I1002 02:01:47.144151 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/8427e298-965b-48f0-821f-bd078fa2e96f-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"8427e298-965b-48f0-821f-bd078fa2e96f\") " pod="openstack/openstack-galera-0" Oct 02 02:01:47 crc kubenswrapper[4885]: I1002 02:01:47.144180 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cdjf7\" (UniqueName: \"kubernetes.io/projected/8427e298-965b-48f0-821f-bd078fa2e96f-kube-api-access-cdjf7\") pod \"openstack-galera-0\" (UID: \"8427e298-965b-48f0-821f-bd078fa2e96f\") " pod="openstack/openstack-galera-0" Oct 02 02:01:47 crc kubenswrapper[4885]: I1002 02:01:47.144252 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/87196135-8d40-47cb-b918-98aea04c7c17-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"87196135-8d40-47cb-b918-98aea04c7c17\") " pod="openstack/openstack-cell1-galera-0" Oct 02 02:01:47 crc kubenswrapper[4885]: I1002 02:01:47.144316 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/87196135-8d40-47cb-b918-98aea04c7c17-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"87196135-8d40-47cb-b918-98aea04c7c17\") " pod="openstack/openstack-cell1-galera-0" Oct 02 02:01:47 crc kubenswrapper[4885]: I1002 02:01:47.144368 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/87196135-8d40-47cb-b918-98aea04c7c17-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"87196135-8d40-47cb-b918-98aea04c7c17\") " pod="openstack/openstack-cell1-galera-0" Oct 02 02:01:47 crc kubenswrapper[4885]: I1002 02:01:47.144406 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8427e298-965b-48f0-821f-bd078fa2e96f-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"8427e298-965b-48f0-821f-bd078fa2e96f\") " pod="openstack/openstack-galera-0" Oct 02 02:01:47 crc kubenswrapper[4885]: I1002 02:01:47.144442 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hjpjc\" (UniqueName: \"kubernetes.io/projected/87196135-8d40-47cb-b918-98aea04c7c17-kube-api-access-hjpjc\") pod \"openstack-cell1-galera-0\" (UID: \"87196135-8d40-47cb-b918-98aea04c7c17\") " pod="openstack/openstack-cell1-galera-0" Oct 02 02:01:47 crc kubenswrapper[4885]: I1002 02:01:47.144480 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/87196135-8d40-47cb-b918-98aea04c7c17-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"87196135-8d40-47cb-b918-98aea04c7c17\") " pod="openstack/openstack-cell1-galera-0" Oct 02 02:01:47 crc kubenswrapper[4885]: I1002 02:01:47.144547 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/8427e298-965b-48f0-821f-bd078fa2e96f-kolla-config\") pod \"openstack-galera-0\" (UID: \"8427e298-965b-48f0-821f-bd078fa2e96f\") " pod="openstack/openstack-galera-0" Oct 02 02:01:47 crc kubenswrapper[4885]: I1002 02:01:47.144595 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/8427e298-965b-48f0-821f-bd078fa2e96f-config-data-default\") pod \"openstack-galera-0\" (UID: \"8427e298-965b-48f0-821f-bd078fa2e96f\") " pod="openstack/openstack-galera-0" Oct 02 02:01:47 crc kubenswrapper[4885]: I1002 02:01:47.144646 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87196135-8d40-47cb-b918-98aea04c7c17-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"87196135-8d40-47cb-b918-98aea04c7c17\") " pod="openstack/openstack-cell1-galera-0" Oct 02 02:01:47 crc kubenswrapper[4885]: I1002 02:01:47.144679 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/8427e298-965b-48f0-821f-bd078fa2e96f-config-data-generated\") pod \"openstack-galera-0\" (UID: \"8427e298-965b-48f0-821f-bd078fa2e96f\") " pod="openstack/openstack-galera-0" Oct 02 02:01:47 crc kubenswrapper[4885]: I1002 02:01:47.144714 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/87196135-8d40-47cb-b918-98aea04c7c17-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"87196135-8d40-47cb-b918-98aea04c7c17\") " pod="openstack/openstack-cell1-galera-0" Oct 02 02:01:47 crc kubenswrapper[4885]: I1002 02:01:47.144750 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-galera-0\" (UID: \"8427e298-965b-48f0-821f-bd078fa2e96f\") " pod="openstack/openstack-galera-0" Oct 02 02:01:47 crc kubenswrapper[4885]: I1002 02:01:47.146233 4885 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-cell1-galera-0\" (UID: \"87196135-8d40-47cb-b918-98aea04c7c17\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/openstack-cell1-galera-0" Oct 02 02:01:47 crc kubenswrapper[4885]: I1002 02:01:47.147609 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/87196135-8d40-47cb-b918-98aea04c7c17-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"87196135-8d40-47cb-b918-98aea04c7c17\") " pod="openstack/openstack-cell1-galera-0" Oct 02 02:01:47 crc kubenswrapper[4885]: I1002 02:01:47.150055 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/87196135-8d40-47cb-b918-98aea04c7c17-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"87196135-8d40-47cb-b918-98aea04c7c17\") " pod="openstack/openstack-cell1-galera-0" Oct 02 02:01:47 crc kubenswrapper[4885]: I1002 02:01:47.154646 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/87196135-8d40-47cb-b918-98aea04c7c17-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"87196135-8d40-47cb-b918-98aea04c7c17\") " pod="openstack/openstack-cell1-galera-0" Oct 02 02:01:47 crc kubenswrapper[4885]: I1002 02:01:47.154947 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87196135-8d40-47cb-b918-98aea04c7c17-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"87196135-8d40-47cb-b918-98aea04c7c17\") " pod="openstack/openstack-cell1-galera-0" Oct 02 02:01:47 crc kubenswrapper[4885]: I1002 02:01:47.158179 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/87196135-8d40-47cb-b918-98aea04c7c17-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"87196135-8d40-47cb-b918-98aea04c7c17\") " pod="openstack/openstack-cell1-galera-0" Oct 02 02:01:47 crc kubenswrapper[4885]: I1002 02:01:47.158505 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/87196135-8d40-47cb-b918-98aea04c7c17-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"87196135-8d40-47cb-b918-98aea04c7c17\") " pod="openstack/openstack-cell1-galera-0" Oct 02 02:01:47 crc kubenswrapper[4885]: I1002 02:01:47.158754 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/87196135-8d40-47cb-b918-98aea04c7c17-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"87196135-8d40-47cb-b918-98aea04c7c17\") " pod="openstack/openstack-cell1-galera-0" Oct 02 02:01:47 crc kubenswrapper[4885]: I1002 02:01:47.174316 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hjpjc\" (UniqueName: \"kubernetes.io/projected/87196135-8d40-47cb-b918-98aea04c7c17-kube-api-access-hjpjc\") pod \"openstack-cell1-galera-0\" (UID: \"87196135-8d40-47cb-b918-98aea04c7c17\") " pod="openstack/openstack-cell1-galera-0" Oct 02 02:01:47 crc kubenswrapper[4885]: I1002 02:01:47.183943 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-cell1-galera-0\" (UID: \"87196135-8d40-47cb-b918-98aea04c7c17\") " pod="openstack/openstack-cell1-galera-0" Oct 02 02:01:47 crc kubenswrapper[4885]: I1002 02:01:47.212721 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 02 02:01:47 crc kubenswrapper[4885]: I1002 02:01:47.245608 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/8427e298-965b-48f0-821f-bd078fa2e96f-config-data-generated\") pod \"openstack-galera-0\" (UID: \"8427e298-965b-48f0-821f-bd078fa2e96f\") " pod="openstack/openstack-galera-0" Oct 02 02:01:47 crc kubenswrapper[4885]: I1002 02:01:47.245870 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-galera-0\" (UID: \"8427e298-965b-48f0-821f-bd078fa2e96f\") " pod="openstack/openstack-galera-0" Oct 02 02:01:47 crc kubenswrapper[4885]: I1002 02:01:47.245995 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8427e298-965b-48f0-821f-bd078fa2e96f-operator-scripts\") pod \"openstack-galera-0\" (UID: \"8427e298-965b-48f0-821f-bd078fa2e96f\") " pod="openstack/openstack-galera-0" Oct 02 02:01:47 crc kubenswrapper[4885]: I1002 02:01:47.246144 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/8427e298-965b-48f0-821f-bd078fa2e96f-secrets\") pod \"openstack-galera-0\" (UID: \"8427e298-965b-48f0-821f-bd078fa2e96f\") " pod="openstack/openstack-galera-0" Oct 02 02:01:47 crc kubenswrapper[4885]: I1002 02:01:47.246242 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/8427e298-965b-48f0-821f-bd078fa2e96f-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"8427e298-965b-48f0-821f-bd078fa2e96f\") " pod="openstack/openstack-galera-0" Oct 02 02:01:47 crc kubenswrapper[4885]: I1002 02:01:47.246355 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cdjf7\" (UniqueName: \"kubernetes.io/projected/8427e298-965b-48f0-821f-bd078fa2e96f-kube-api-access-cdjf7\") pod \"openstack-galera-0\" (UID: \"8427e298-965b-48f0-821f-bd078fa2e96f\") " pod="openstack/openstack-galera-0" Oct 02 02:01:47 crc kubenswrapper[4885]: I1002 02:01:47.246468 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8427e298-965b-48f0-821f-bd078fa2e96f-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"8427e298-965b-48f0-821f-bd078fa2e96f\") " pod="openstack/openstack-galera-0" Oct 02 02:01:47 crc kubenswrapper[4885]: I1002 02:01:47.246593 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/8427e298-965b-48f0-821f-bd078fa2e96f-kolla-config\") pod \"openstack-galera-0\" (UID: \"8427e298-965b-48f0-821f-bd078fa2e96f\") " pod="openstack/openstack-galera-0" Oct 02 02:01:47 crc kubenswrapper[4885]: I1002 02:01:47.246694 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/8427e298-965b-48f0-821f-bd078fa2e96f-config-data-default\") pod \"openstack-galera-0\" (UID: \"8427e298-965b-48f0-821f-bd078fa2e96f\") " pod="openstack/openstack-galera-0" Oct 02 02:01:47 crc kubenswrapper[4885]: I1002 02:01:47.247844 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/8427e298-965b-48f0-821f-bd078fa2e96f-config-data-default\") pod \"openstack-galera-0\" (UID: \"8427e298-965b-48f0-821f-bd078fa2e96f\") " pod="openstack/openstack-galera-0" Oct 02 02:01:47 crc kubenswrapper[4885]: I1002 02:01:47.248230 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/8427e298-965b-48f0-821f-bd078fa2e96f-config-data-generated\") pod \"openstack-galera-0\" (UID: \"8427e298-965b-48f0-821f-bd078fa2e96f\") " pod="openstack/openstack-galera-0" Oct 02 02:01:47 crc kubenswrapper[4885]: I1002 02:01:47.248590 4885 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-galera-0\" (UID: \"8427e298-965b-48f0-821f-bd078fa2e96f\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/openstack-galera-0" Oct 02 02:01:47 crc kubenswrapper[4885]: I1002 02:01:47.260705 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/8427e298-965b-48f0-821f-bd078fa2e96f-kolla-config\") pod \"openstack-galera-0\" (UID: \"8427e298-965b-48f0-821f-bd078fa2e96f\") " pod="openstack/openstack-galera-0" Oct 02 02:01:47 crc kubenswrapper[4885]: I1002 02:01:47.262060 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8427e298-965b-48f0-821f-bd078fa2e96f-operator-scripts\") pod \"openstack-galera-0\" (UID: \"8427e298-965b-48f0-821f-bd078fa2e96f\") " pod="openstack/openstack-galera-0" Oct 02 02:01:47 crc kubenswrapper[4885]: I1002 02:01:47.262243 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/8427e298-965b-48f0-821f-bd078fa2e96f-secrets\") pod \"openstack-galera-0\" (UID: \"8427e298-965b-48f0-821f-bd078fa2e96f\") " pod="openstack/openstack-galera-0" Oct 02 02:01:47 crc kubenswrapper[4885]: I1002 02:01:47.295512 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-galera-0\" (UID: \"8427e298-965b-48f0-821f-bd078fa2e96f\") " pod="openstack/openstack-galera-0" Oct 02 02:01:47 crc kubenswrapper[4885]: I1002 02:01:47.298650 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/8427e298-965b-48f0-821f-bd078fa2e96f-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"8427e298-965b-48f0-821f-bd078fa2e96f\") " pod="openstack/openstack-galera-0" Oct 02 02:01:47 crc kubenswrapper[4885]: I1002 02:01:47.302648 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8427e298-965b-48f0-821f-bd078fa2e96f-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"8427e298-965b-48f0-821f-bd078fa2e96f\") " pod="openstack/openstack-galera-0" Oct 02 02:01:47 crc kubenswrapper[4885]: I1002 02:01:47.302817 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cdjf7\" (UniqueName: \"kubernetes.io/projected/8427e298-965b-48f0-821f-bd078fa2e96f-kube-api-access-cdjf7\") pod \"openstack-galera-0\" (UID: \"8427e298-965b-48f0-821f-bd078fa2e96f\") " pod="openstack/openstack-galera-0" Oct 02 02:01:47 crc kubenswrapper[4885]: I1002 02:01:47.325300 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 02 02:01:47 crc kubenswrapper[4885]: I1002 02:01:47.366444 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-p9cqp" event={"ID":"90626475-ee00-431f-8992-d5fc20728c6c","Type":"ContainerStarted","Data":"b9202f38c3e6b86dff94c2d73561c7e9d89082b2a9efa38a64f9bfd18a8c4c84"} Oct 02 02:01:47 crc kubenswrapper[4885]: I1002 02:01:47.450278 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Oct 02 02:01:47 crc kubenswrapper[4885]: I1002 02:01:47.451926 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 02 02:01:47 crc kubenswrapper[4885]: I1002 02:01:47.456617 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-rwv97" Oct 02 02:01:47 crc kubenswrapper[4885]: I1002 02:01:47.456637 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Oct 02 02:01:47 crc kubenswrapper[4885]: I1002 02:01:47.460176 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Oct 02 02:01:47 crc kubenswrapper[4885]: I1002 02:01:47.470321 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 02 02:01:47 crc kubenswrapper[4885]: I1002 02:01:47.551221 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d798d658-f60b-4e23-8753-b3de580e44b0-combined-ca-bundle\") pod \"memcached-0\" (UID: \"d798d658-f60b-4e23-8753-b3de580e44b0\") " pod="openstack/memcached-0" Oct 02 02:01:47 crc kubenswrapper[4885]: I1002 02:01:47.551330 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d798d658-f60b-4e23-8753-b3de580e44b0-kolla-config\") pod \"memcached-0\" (UID: \"d798d658-f60b-4e23-8753-b3de580e44b0\") " pod="openstack/memcached-0" Oct 02 02:01:47 crc kubenswrapper[4885]: I1002 02:01:47.551347 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l42rj\" (UniqueName: \"kubernetes.io/projected/d798d658-f60b-4e23-8753-b3de580e44b0-kube-api-access-l42rj\") pod \"memcached-0\" (UID: \"d798d658-f60b-4e23-8753-b3de580e44b0\") " pod="openstack/memcached-0" Oct 02 02:01:47 crc kubenswrapper[4885]: I1002 02:01:47.551369 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/d798d658-f60b-4e23-8753-b3de580e44b0-memcached-tls-certs\") pod \"memcached-0\" (UID: \"d798d658-f60b-4e23-8753-b3de580e44b0\") " pod="openstack/memcached-0" Oct 02 02:01:47 crc kubenswrapper[4885]: I1002 02:01:47.551403 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d798d658-f60b-4e23-8753-b3de580e44b0-config-data\") pod \"memcached-0\" (UID: \"d798d658-f60b-4e23-8753-b3de580e44b0\") " pod="openstack/memcached-0" Oct 02 02:01:47 crc kubenswrapper[4885]: I1002 02:01:47.652646 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d798d658-f60b-4e23-8753-b3de580e44b0-combined-ca-bundle\") pod \"memcached-0\" (UID: \"d798d658-f60b-4e23-8753-b3de580e44b0\") " pod="openstack/memcached-0" Oct 02 02:01:47 crc kubenswrapper[4885]: I1002 02:01:47.652756 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d798d658-f60b-4e23-8753-b3de580e44b0-kolla-config\") pod \"memcached-0\" (UID: \"d798d658-f60b-4e23-8753-b3de580e44b0\") " pod="openstack/memcached-0" Oct 02 02:01:47 crc kubenswrapper[4885]: I1002 02:01:47.652784 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l42rj\" (UniqueName: \"kubernetes.io/projected/d798d658-f60b-4e23-8753-b3de580e44b0-kube-api-access-l42rj\") pod \"memcached-0\" (UID: \"d798d658-f60b-4e23-8753-b3de580e44b0\") " pod="openstack/memcached-0" Oct 02 02:01:47 crc kubenswrapper[4885]: I1002 02:01:47.652811 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/d798d658-f60b-4e23-8753-b3de580e44b0-memcached-tls-certs\") pod \"memcached-0\" (UID: \"d798d658-f60b-4e23-8753-b3de580e44b0\") " pod="openstack/memcached-0" Oct 02 02:01:47 crc kubenswrapper[4885]: I1002 02:01:47.652852 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d798d658-f60b-4e23-8753-b3de580e44b0-config-data\") pod \"memcached-0\" (UID: \"d798d658-f60b-4e23-8753-b3de580e44b0\") " pod="openstack/memcached-0" Oct 02 02:01:47 crc kubenswrapper[4885]: I1002 02:01:47.653745 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d798d658-f60b-4e23-8753-b3de580e44b0-kolla-config\") pod \"memcached-0\" (UID: \"d798d658-f60b-4e23-8753-b3de580e44b0\") " pod="openstack/memcached-0" Oct 02 02:01:47 crc kubenswrapper[4885]: I1002 02:01:47.653790 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d798d658-f60b-4e23-8753-b3de580e44b0-config-data\") pod \"memcached-0\" (UID: \"d798d658-f60b-4e23-8753-b3de580e44b0\") " pod="openstack/memcached-0" Oct 02 02:01:47 crc kubenswrapper[4885]: I1002 02:01:47.656984 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d798d658-f60b-4e23-8753-b3de580e44b0-combined-ca-bundle\") pod \"memcached-0\" (UID: \"d798d658-f60b-4e23-8753-b3de580e44b0\") " pod="openstack/memcached-0" Oct 02 02:01:47 crc kubenswrapper[4885]: I1002 02:01:47.657689 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/d798d658-f60b-4e23-8753-b3de580e44b0-memcached-tls-certs\") pod \"memcached-0\" (UID: \"d798d658-f60b-4e23-8753-b3de580e44b0\") " pod="openstack/memcached-0" Oct 02 02:01:47 crc kubenswrapper[4885]: I1002 02:01:47.679947 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l42rj\" (UniqueName: \"kubernetes.io/projected/d798d658-f60b-4e23-8753-b3de580e44b0-kube-api-access-l42rj\") pod \"memcached-0\" (UID: \"d798d658-f60b-4e23-8753-b3de580e44b0\") " pod="openstack/memcached-0" Oct 02 02:01:47 crc kubenswrapper[4885]: I1002 02:01:47.769813 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 02 02:01:49 crc kubenswrapper[4885]: I1002 02:01:49.292190 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 02 02:01:49 crc kubenswrapper[4885]: I1002 02:01:49.297057 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 02 02:01:49 crc kubenswrapper[4885]: I1002 02:01:49.301682 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-nwgqt" Oct 02 02:01:49 crc kubenswrapper[4885]: I1002 02:01:49.302785 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 02 02:01:49 crc kubenswrapper[4885]: I1002 02:01:49.376980 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2ngf5\" (UniqueName: \"kubernetes.io/projected/faff5405-00c6-4d75-bbca-77f9230c2811-kube-api-access-2ngf5\") pod \"kube-state-metrics-0\" (UID: \"faff5405-00c6-4d75-bbca-77f9230c2811\") " pod="openstack/kube-state-metrics-0" Oct 02 02:01:49 crc kubenswrapper[4885]: I1002 02:01:49.478916 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2ngf5\" (UniqueName: \"kubernetes.io/projected/faff5405-00c6-4d75-bbca-77f9230c2811-kube-api-access-2ngf5\") pod \"kube-state-metrics-0\" (UID: \"faff5405-00c6-4d75-bbca-77f9230c2811\") " pod="openstack/kube-state-metrics-0" Oct 02 02:01:49 crc kubenswrapper[4885]: I1002 02:01:49.495102 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2ngf5\" (UniqueName: \"kubernetes.io/projected/faff5405-00c6-4d75-bbca-77f9230c2811-kube-api-access-2ngf5\") pod \"kube-state-metrics-0\" (UID: \"faff5405-00c6-4d75-bbca-77f9230c2811\") " pod="openstack/kube-state-metrics-0" Oct 02 02:01:49 crc kubenswrapper[4885]: I1002 02:01:49.616580 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 02 02:01:52 crc kubenswrapper[4885]: I1002 02:01:52.175899 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-fgvpt"] Oct 02 02:01:52 crc kubenswrapper[4885]: I1002 02:01:52.177370 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-fgvpt" Oct 02 02:01:52 crc kubenswrapper[4885]: I1002 02:01:52.182752 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-mn8r7" Oct 02 02:01:52 crc kubenswrapper[4885]: I1002 02:01:52.183643 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Oct 02 02:01:52 crc kubenswrapper[4885]: I1002 02:01:52.183923 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Oct 02 02:01:52 crc kubenswrapper[4885]: I1002 02:01:52.188300 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-fgvpt"] Oct 02 02:01:52 crc kubenswrapper[4885]: I1002 02:01:52.192074 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-q92rb"] Oct 02 02:01:52 crc kubenswrapper[4885]: I1002 02:01:52.194045 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-q92rb" Oct 02 02:01:52 crc kubenswrapper[4885]: I1002 02:01:52.217828 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-q92rb"] Oct 02 02:01:52 crc kubenswrapper[4885]: I1002 02:01:52.231464 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/924fb321-4073-449a-a546-811b046ed26a-combined-ca-bundle\") pod \"ovn-controller-fgvpt\" (UID: \"924fb321-4073-449a-a546-811b046ed26a\") " pod="openstack/ovn-controller-fgvpt" Oct 02 02:01:52 crc kubenswrapper[4885]: I1002 02:01:52.231529 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9ml5g\" (UniqueName: \"kubernetes.io/projected/924fb321-4073-449a-a546-811b046ed26a-kube-api-access-9ml5g\") pod \"ovn-controller-fgvpt\" (UID: \"924fb321-4073-449a-a546-811b046ed26a\") " pod="openstack/ovn-controller-fgvpt" Oct 02 02:01:52 crc kubenswrapper[4885]: I1002 02:01:52.231566 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/924fb321-4073-449a-a546-811b046ed26a-var-run-ovn\") pod \"ovn-controller-fgvpt\" (UID: \"924fb321-4073-449a-a546-811b046ed26a\") " pod="openstack/ovn-controller-fgvpt" Oct 02 02:01:52 crc kubenswrapper[4885]: I1002 02:01:52.231594 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/924fb321-4073-449a-a546-811b046ed26a-ovn-controller-tls-certs\") pod \"ovn-controller-fgvpt\" (UID: \"924fb321-4073-449a-a546-811b046ed26a\") " pod="openstack/ovn-controller-fgvpt" Oct 02 02:01:52 crc kubenswrapper[4885]: I1002 02:01:52.231638 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/924fb321-4073-449a-a546-811b046ed26a-var-log-ovn\") pod \"ovn-controller-fgvpt\" (UID: \"924fb321-4073-449a-a546-811b046ed26a\") " pod="openstack/ovn-controller-fgvpt" Oct 02 02:01:52 crc kubenswrapper[4885]: I1002 02:01:52.231667 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/924fb321-4073-449a-a546-811b046ed26a-var-run\") pod \"ovn-controller-fgvpt\" (UID: \"924fb321-4073-449a-a546-811b046ed26a\") " pod="openstack/ovn-controller-fgvpt" Oct 02 02:01:52 crc kubenswrapper[4885]: I1002 02:01:52.231691 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/924fb321-4073-449a-a546-811b046ed26a-scripts\") pod \"ovn-controller-fgvpt\" (UID: \"924fb321-4073-449a-a546-811b046ed26a\") " pod="openstack/ovn-controller-fgvpt" Oct 02 02:01:52 crc kubenswrapper[4885]: I1002 02:01:52.332949 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/b65df3c4-cc33-43da-9bf1-8f46a480f7eb-var-lib\") pod \"ovn-controller-ovs-q92rb\" (UID: \"b65df3c4-cc33-43da-9bf1-8f46a480f7eb\") " pod="openstack/ovn-controller-ovs-q92rb" Oct 02 02:01:52 crc kubenswrapper[4885]: I1002 02:01:52.333001 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nngt4\" (UniqueName: \"kubernetes.io/projected/b65df3c4-cc33-43da-9bf1-8f46a480f7eb-kube-api-access-nngt4\") pod \"ovn-controller-ovs-q92rb\" (UID: \"b65df3c4-cc33-43da-9bf1-8f46a480f7eb\") " pod="openstack/ovn-controller-ovs-q92rb" Oct 02 02:01:52 crc kubenswrapper[4885]: I1002 02:01:52.333055 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/924fb321-4073-449a-a546-811b046ed26a-combined-ca-bundle\") pod \"ovn-controller-fgvpt\" (UID: \"924fb321-4073-449a-a546-811b046ed26a\") " pod="openstack/ovn-controller-fgvpt" Oct 02 02:01:52 crc kubenswrapper[4885]: I1002 02:01:52.333077 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b65df3c4-cc33-43da-9bf1-8f46a480f7eb-var-run\") pod \"ovn-controller-ovs-q92rb\" (UID: \"b65df3c4-cc33-43da-9bf1-8f46a480f7eb\") " pod="openstack/ovn-controller-ovs-q92rb" Oct 02 02:01:52 crc kubenswrapper[4885]: I1002 02:01:52.333098 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9ml5g\" (UniqueName: \"kubernetes.io/projected/924fb321-4073-449a-a546-811b046ed26a-kube-api-access-9ml5g\") pod \"ovn-controller-fgvpt\" (UID: \"924fb321-4073-449a-a546-811b046ed26a\") " pod="openstack/ovn-controller-fgvpt" Oct 02 02:01:52 crc kubenswrapper[4885]: I1002 02:01:52.333117 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/924fb321-4073-449a-a546-811b046ed26a-var-run-ovn\") pod \"ovn-controller-fgvpt\" (UID: \"924fb321-4073-449a-a546-811b046ed26a\") " pod="openstack/ovn-controller-fgvpt" Oct 02 02:01:52 crc kubenswrapper[4885]: I1002 02:01:52.333136 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/924fb321-4073-449a-a546-811b046ed26a-ovn-controller-tls-certs\") pod \"ovn-controller-fgvpt\" (UID: \"924fb321-4073-449a-a546-811b046ed26a\") " pod="openstack/ovn-controller-fgvpt" Oct 02 02:01:52 crc kubenswrapper[4885]: I1002 02:01:52.333161 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/b65df3c4-cc33-43da-9bf1-8f46a480f7eb-var-log\") pod \"ovn-controller-ovs-q92rb\" (UID: \"b65df3c4-cc33-43da-9bf1-8f46a480f7eb\") " pod="openstack/ovn-controller-ovs-q92rb" Oct 02 02:01:52 crc kubenswrapper[4885]: I1002 02:01:52.333182 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/924fb321-4073-449a-a546-811b046ed26a-var-log-ovn\") pod \"ovn-controller-fgvpt\" (UID: \"924fb321-4073-449a-a546-811b046ed26a\") " pod="openstack/ovn-controller-fgvpt" Oct 02 02:01:52 crc kubenswrapper[4885]: I1002 02:01:52.333203 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/924fb321-4073-449a-a546-811b046ed26a-var-run\") pod \"ovn-controller-fgvpt\" (UID: \"924fb321-4073-449a-a546-811b046ed26a\") " pod="openstack/ovn-controller-fgvpt" Oct 02 02:01:52 crc kubenswrapper[4885]: I1002 02:01:52.333222 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/924fb321-4073-449a-a546-811b046ed26a-scripts\") pod \"ovn-controller-fgvpt\" (UID: \"924fb321-4073-449a-a546-811b046ed26a\") " pod="openstack/ovn-controller-fgvpt" Oct 02 02:01:52 crc kubenswrapper[4885]: I1002 02:01:52.333243 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b65df3c4-cc33-43da-9bf1-8f46a480f7eb-scripts\") pod \"ovn-controller-ovs-q92rb\" (UID: \"b65df3c4-cc33-43da-9bf1-8f46a480f7eb\") " pod="openstack/ovn-controller-ovs-q92rb" Oct 02 02:01:52 crc kubenswrapper[4885]: I1002 02:01:52.333299 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/b65df3c4-cc33-43da-9bf1-8f46a480f7eb-etc-ovs\") pod \"ovn-controller-ovs-q92rb\" (UID: \"b65df3c4-cc33-43da-9bf1-8f46a480f7eb\") " pod="openstack/ovn-controller-ovs-q92rb" Oct 02 02:01:52 crc kubenswrapper[4885]: I1002 02:01:52.335570 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/924fb321-4073-449a-a546-811b046ed26a-var-run-ovn\") pod \"ovn-controller-fgvpt\" (UID: \"924fb321-4073-449a-a546-811b046ed26a\") " pod="openstack/ovn-controller-fgvpt" Oct 02 02:01:52 crc kubenswrapper[4885]: I1002 02:01:52.335656 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/924fb321-4073-449a-a546-811b046ed26a-var-log-ovn\") pod \"ovn-controller-fgvpt\" (UID: \"924fb321-4073-449a-a546-811b046ed26a\") " pod="openstack/ovn-controller-fgvpt" Oct 02 02:01:52 crc kubenswrapper[4885]: I1002 02:01:52.336866 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/924fb321-4073-449a-a546-811b046ed26a-scripts\") pod \"ovn-controller-fgvpt\" (UID: \"924fb321-4073-449a-a546-811b046ed26a\") " pod="openstack/ovn-controller-fgvpt" Oct 02 02:01:52 crc kubenswrapper[4885]: I1002 02:01:52.337157 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/924fb321-4073-449a-a546-811b046ed26a-var-run\") pod \"ovn-controller-fgvpt\" (UID: \"924fb321-4073-449a-a546-811b046ed26a\") " pod="openstack/ovn-controller-fgvpt" Oct 02 02:01:52 crc kubenswrapper[4885]: I1002 02:01:52.340841 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/924fb321-4073-449a-a546-811b046ed26a-combined-ca-bundle\") pod \"ovn-controller-fgvpt\" (UID: \"924fb321-4073-449a-a546-811b046ed26a\") " pod="openstack/ovn-controller-fgvpt" Oct 02 02:01:52 crc kubenswrapper[4885]: I1002 02:01:52.341849 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/924fb321-4073-449a-a546-811b046ed26a-ovn-controller-tls-certs\") pod \"ovn-controller-fgvpt\" (UID: \"924fb321-4073-449a-a546-811b046ed26a\") " pod="openstack/ovn-controller-fgvpt" Oct 02 02:01:52 crc kubenswrapper[4885]: I1002 02:01:52.348728 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9ml5g\" (UniqueName: \"kubernetes.io/projected/924fb321-4073-449a-a546-811b046ed26a-kube-api-access-9ml5g\") pod \"ovn-controller-fgvpt\" (UID: \"924fb321-4073-449a-a546-811b046ed26a\") " pod="openstack/ovn-controller-fgvpt" Oct 02 02:01:52 crc kubenswrapper[4885]: I1002 02:01:52.436290 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/b65df3c4-cc33-43da-9bf1-8f46a480f7eb-var-log\") pod \"ovn-controller-ovs-q92rb\" (UID: \"b65df3c4-cc33-43da-9bf1-8f46a480f7eb\") " pod="openstack/ovn-controller-ovs-q92rb" Oct 02 02:01:52 crc kubenswrapper[4885]: I1002 02:01:52.436374 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b65df3c4-cc33-43da-9bf1-8f46a480f7eb-scripts\") pod \"ovn-controller-ovs-q92rb\" (UID: \"b65df3c4-cc33-43da-9bf1-8f46a480f7eb\") " pod="openstack/ovn-controller-ovs-q92rb" Oct 02 02:01:52 crc kubenswrapper[4885]: I1002 02:01:52.436399 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/b65df3c4-cc33-43da-9bf1-8f46a480f7eb-etc-ovs\") pod \"ovn-controller-ovs-q92rb\" (UID: \"b65df3c4-cc33-43da-9bf1-8f46a480f7eb\") " pod="openstack/ovn-controller-ovs-q92rb" Oct 02 02:01:52 crc kubenswrapper[4885]: I1002 02:01:52.436433 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/b65df3c4-cc33-43da-9bf1-8f46a480f7eb-var-lib\") pod \"ovn-controller-ovs-q92rb\" (UID: \"b65df3c4-cc33-43da-9bf1-8f46a480f7eb\") " pod="openstack/ovn-controller-ovs-q92rb" Oct 02 02:01:52 crc kubenswrapper[4885]: I1002 02:01:52.436465 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nngt4\" (UniqueName: \"kubernetes.io/projected/b65df3c4-cc33-43da-9bf1-8f46a480f7eb-kube-api-access-nngt4\") pod \"ovn-controller-ovs-q92rb\" (UID: \"b65df3c4-cc33-43da-9bf1-8f46a480f7eb\") " pod="openstack/ovn-controller-ovs-q92rb" Oct 02 02:01:52 crc kubenswrapper[4885]: I1002 02:01:52.436528 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b65df3c4-cc33-43da-9bf1-8f46a480f7eb-var-run\") pod \"ovn-controller-ovs-q92rb\" (UID: \"b65df3c4-cc33-43da-9bf1-8f46a480f7eb\") " pod="openstack/ovn-controller-ovs-q92rb" Oct 02 02:01:52 crc kubenswrapper[4885]: I1002 02:01:52.436798 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b65df3c4-cc33-43da-9bf1-8f46a480f7eb-var-run\") pod \"ovn-controller-ovs-q92rb\" (UID: \"b65df3c4-cc33-43da-9bf1-8f46a480f7eb\") " pod="openstack/ovn-controller-ovs-q92rb" Oct 02 02:01:52 crc kubenswrapper[4885]: I1002 02:01:52.436941 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/b65df3c4-cc33-43da-9bf1-8f46a480f7eb-var-log\") pod \"ovn-controller-ovs-q92rb\" (UID: \"b65df3c4-cc33-43da-9bf1-8f46a480f7eb\") " pod="openstack/ovn-controller-ovs-q92rb" Oct 02 02:01:52 crc kubenswrapper[4885]: I1002 02:01:52.437007 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/b65df3c4-cc33-43da-9bf1-8f46a480f7eb-etc-ovs\") pod \"ovn-controller-ovs-q92rb\" (UID: \"b65df3c4-cc33-43da-9bf1-8f46a480f7eb\") " pod="openstack/ovn-controller-ovs-q92rb" Oct 02 02:01:52 crc kubenswrapper[4885]: I1002 02:01:52.437081 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/b65df3c4-cc33-43da-9bf1-8f46a480f7eb-var-lib\") pod \"ovn-controller-ovs-q92rb\" (UID: \"b65df3c4-cc33-43da-9bf1-8f46a480f7eb\") " pod="openstack/ovn-controller-ovs-q92rb" Oct 02 02:01:52 crc kubenswrapper[4885]: I1002 02:01:52.439725 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b65df3c4-cc33-43da-9bf1-8f46a480f7eb-scripts\") pod \"ovn-controller-ovs-q92rb\" (UID: \"b65df3c4-cc33-43da-9bf1-8f46a480f7eb\") " pod="openstack/ovn-controller-ovs-q92rb" Oct 02 02:01:52 crc kubenswrapper[4885]: I1002 02:01:52.464039 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nngt4\" (UniqueName: \"kubernetes.io/projected/b65df3c4-cc33-43da-9bf1-8f46a480f7eb-kube-api-access-nngt4\") pod \"ovn-controller-ovs-q92rb\" (UID: \"b65df3c4-cc33-43da-9bf1-8f46a480f7eb\") " pod="openstack/ovn-controller-ovs-q92rb" Oct 02 02:01:52 crc kubenswrapper[4885]: I1002 02:01:52.495206 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-fgvpt" Oct 02 02:01:52 crc kubenswrapper[4885]: I1002 02:01:52.516883 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-q92rb" Oct 02 02:01:53 crc kubenswrapper[4885]: I1002 02:01:53.514043 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 02 02:01:53 crc kubenswrapper[4885]: I1002 02:01:53.518160 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 02 02:01:53 crc kubenswrapper[4885]: I1002 02:01:53.520372 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Oct 02 02:01:53 crc kubenswrapper[4885]: I1002 02:01:53.521855 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Oct 02 02:01:53 crc kubenswrapper[4885]: I1002 02:01:53.522731 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Oct 02 02:01:53 crc kubenswrapper[4885]: I1002 02:01:53.523022 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-gfs4v" Oct 02 02:01:53 crc kubenswrapper[4885]: I1002 02:01:53.523220 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Oct 02 02:01:53 crc kubenswrapper[4885]: I1002 02:01:53.526521 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 02 02:01:53 crc kubenswrapper[4885]: I1002 02:01:53.550348 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/079e125d-c938-4962-8f41-09b3109fb0ae-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"079e125d-c938-4962-8f41-09b3109fb0ae\") " pod="openstack/ovsdbserver-nb-0" Oct 02 02:01:53 crc kubenswrapper[4885]: I1002 02:01:53.550410 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/079e125d-c938-4962-8f41-09b3109fb0ae-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"079e125d-c938-4962-8f41-09b3109fb0ae\") " pod="openstack/ovsdbserver-nb-0" Oct 02 02:01:53 crc kubenswrapper[4885]: I1002 02:01:53.550434 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-nb-0\" (UID: \"079e125d-c938-4962-8f41-09b3109fb0ae\") " pod="openstack/ovsdbserver-nb-0" Oct 02 02:01:53 crc kubenswrapper[4885]: I1002 02:01:53.550683 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/079e125d-c938-4962-8f41-09b3109fb0ae-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"079e125d-c938-4962-8f41-09b3109fb0ae\") " pod="openstack/ovsdbserver-nb-0" Oct 02 02:01:53 crc kubenswrapper[4885]: I1002 02:01:53.550773 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/079e125d-c938-4962-8f41-09b3109fb0ae-config\") pod \"ovsdbserver-nb-0\" (UID: \"079e125d-c938-4962-8f41-09b3109fb0ae\") " pod="openstack/ovsdbserver-nb-0" Oct 02 02:01:53 crc kubenswrapper[4885]: I1002 02:01:53.550827 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/079e125d-c938-4962-8f41-09b3109fb0ae-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"079e125d-c938-4962-8f41-09b3109fb0ae\") " pod="openstack/ovsdbserver-nb-0" Oct 02 02:01:53 crc kubenswrapper[4885]: I1002 02:01:53.550854 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vgltz\" (UniqueName: \"kubernetes.io/projected/079e125d-c938-4962-8f41-09b3109fb0ae-kube-api-access-vgltz\") pod \"ovsdbserver-nb-0\" (UID: \"079e125d-c938-4962-8f41-09b3109fb0ae\") " pod="openstack/ovsdbserver-nb-0" Oct 02 02:01:53 crc kubenswrapper[4885]: I1002 02:01:53.551055 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/079e125d-c938-4962-8f41-09b3109fb0ae-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"079e125d-c938-4962-8f41-09b3109fb0ae\") " pod="openstack/ovsdbserver-nb-0" Oct 02 02:01:53 crc kubenswrapper[4885]: I1002 02:01:53.652923 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/079e125d-c938-4962-8f41-09b3109fb0ae-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"079e125d-c938-4962-8f41-09b3109fb0ae\") " pod="openstack/ovsdbserver-nb-0" Oct 02 02:01:53 crc kubenswrapper[4885]: I1002 02:01:53.652983 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-nb-0\" (UID: \"079e125d-c938-4962-8f41-09b3109fb0ae\") " pod="openstack/ovsdbserver-nb-0" Oct 02 02:01:53 crc kubenswrapper[4885]: I1002 02:01:53.653013 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/079e125d-c938-4962-8f41-09b3109fb0ae-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"079e125d-c938-4962-8f41-09b3109fb0ae\") " pod="openstack/ovsdbserver-nb-0" Oct 02 02:01:53 crc kubenswrapper[4885]: I1002 02:01:53.653040 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/079e125d-c938-4962-8f41-09b3109fb0ae-config\") pod \"ovsdbserver-nb-0\" (UID: \"079e125d-c938-4962-8f41-09b3109fb0ae\") " pod="openstack/ovsdbserver-nb-0" Oct 02 02:01:53 crc kubenswrapper[4885]: I1002 02:01:53.653072 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/079e125d-c938-4962-8f41-09b3109fb0ae-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"079e125d-c938-4962-8f41-09b3109fb0ae\") " pod="openstack/ovsdbserver-nb-0" Oct 02 02:01:53 crc kubenswrapper[4885]: I1002 02:01:53.653091 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vgltz\" (UniqueName: \"kubernetes.io/projected/079e125d-c938-4962-8f41-09b3109fb0ae-kube-api-access-vgltz\") pod \"ovsdbserver-nb-0\" (UID: \"079e125d-c938-4962-8f41-09b3109fb0ae\") " pod="openstack/ovsdbserver-nb-0" Oct 02 02:01:53 crc kubenswrapper[4885]: I1002 02:01:53.653122 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/079e125d-c938-4962-8f41-09b3109fb0ae-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"079e125d-c938-4962-8f41-09b3109fb0ae\") " pod="openstack/ovsdbserver-nb-0" Oct 02 02:01:53 crc kubenswrapper[4885]: I1002 02:01:53.653221 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/079e125d-c938-4962-8f41-09b3109fb0ae-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"079e125d-c938-4962-8f41-09b3109fb0ae\") " pod="openstack/ovsdbserver-nb-0" Oct 02 02:01:53 crc kubenswrapper[4885]: I1002 02:01:53.653822 4885 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-nb-0\" (UID: \"079e125d-c938-4962-8f41-09b3109fb0ae\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/ovsdbserver-nb-0" Oct 02 02:01:53 crc kubenswrapper[4885]: I1002 02:01:53.653962 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/079e125d-c938-4962-8f41-09b3109fb0ae-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"079e125d-c938-4962-8f41-09b3109fb0ae\") " pod="openstack/ovsdbserver-nb-0" Oct 02 02:01:53 crc kubenswrapper[4885]: I1002 02:01:53.654835 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/079e125d-c938-4962-8f41-09b3109fb0ae-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"079e125d-c938-4962-8f41-09b3109fb0ae\") " pod="openstack/ovsdbserver-nb-0" Oct 02 02:01:53 crc kubenswrapper[4885]: I1002 02:01:53.654894 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/079e125d-c938-4962-8f41-09b3109fb0ae-config\") pod \"ovsdbserver-nb-0\" (UID: \"079e125d-c938-4962-8f41-09b3109fb0ae\") " pod="openstack/ovsdbserver-nb-0" Oct 02 02:01:53 crc kubenswrapper[4885]: I1002 02:01:53.659898 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/079e125d-c938-4962-8f41-09b3109fb0ae-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"079e125d-c938-4962-8f41-09b3109fb0ae\") " pod="openstack/ovsdbserver-nb-0" Oct 02 02:01:53 crc kubenswrapper[4885]: I1002 02:01:53.663896 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/079e125d-c938-4962-8f41-09b3109fb0ae-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"079e125d-c938-4962-8f41-09b3109fb0ae\") " pod="openstack/ovsdbserver-nb-0" Oct 02 02:01:53 crc kubenswrapper[4885]: I1002 02:01:53.664504 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/079e125d-c938-4962-8f41-09b3109fb0ae-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"079e125d-c938-4962-8f41-09b3109fb0ae\") " pod="openstack/ovsdbserver-nb-0" Oct 02 02:01:53 crc kubenswrapper[4885]: I1002 02:01:53.669118 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vgltz\" (UniqueName: \"kubernetes.io/projected/079e125d-c938-4962-8f41-09b3109fb0ae-kube-api-access-vgltz\") pod \"ovsdbserver-nb-0\" (UID: \"079e125d-c938-4962-8f41-09b3109fb0ae\") " pod="openstack/ovsdbserver-nb-0" Oct 02 02:01:53 crc kubenswrapper[4885]: I1002 02:01:53.706386 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-nb-0\" (UID: \"079e125d-c938-4962-8f41-09b3109fb0ae\") " pod="openstack/ovsdbserver-nb-0" Oct 02 02:01:53 crc kubenswrapper[4885]: I1002 02:01:53.853987 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 02 02:01:56 crc kubenswrapper[4885]: E1002 02:01:56.037092 4885 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Oct 02 02:01:56 crc kubenswrapper[4885]: E1002 02:01:56.037434 4885 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gpvqj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-tvzhw_openstack(10deeddc-732a-4154-a74c-b28abc841bc1): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 02:01:56 crc kubenswrapper[4885]: E1002 02:01:56.038690 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-tvzhw" podUID="10deeddc-732a-4154-a74c-b28abc841bc1" Oct 02 02:01:56 crc kubenswrapper[4885]: E1002 02:01:56.223527 4885 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Oct 02 02:01:56 crc kubenswrapper[4885]: E1002 02:01:56.223998 4885 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5w66h,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-8ljrn_openstack(f541a783-9596-4275-b87e-cf289b223f9d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 02:01:56 crc kubenswrapper[4885]: E1002 02:01:56.225675 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-8ljrn" podUID="f541a783-9596-4275-b87e-cf289b223f9d" Oct 02 02:01:56 crc kubenswrapper[4885]: I1002 02:01:56.708393 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 02 02:01:56 crc kubenswrapper[4885]: I1002 02:01:56.710116 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 02 02:01:56 crc kubenswrapper[4885]: I1002 02:01:56.716185 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Oct 02 02:01:56 crc kubenswrapper[4885]: I1002 02:01:56.716550 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-r2sjw" Oct 02 02:01:56 crc kubenswrapper[4885]: I1002 02:01:56.716835 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Oct 02 02:01:56 crc kubenswrapper[4885]: I1002 02:01:56.717090 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Oct 02 02:01:56 crc kubenswrapper[4885]: I1002 02:01:56.726487 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 02 02:01:56 crc kubenswrapper[4885]: I1002 02:01:56.826071 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/33286960-496c-4f14-a334-c6dc765ca956-config\") pod \"ovsdbserver-sb-0\" (UID: \"33286960-496c-4f14-a334-c6dc765ca956\") " pod="openstack/ovsdbserver-sb-0" Oct 02 02:01:56 crc kubenswrapper[4885]: I1002 02:01:56.826349 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33286960-496c-4f14-a334-c6dc765ca956-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"33286960-496c-4f14-a334-c6dc765ca956\") " pod="openstack/ovsdbserver-sb-0" Oct 02 02:01:56 crc kubenswrapper[4885]: I1002 02:01:56.826370 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/33286960-496c-4f14-a334-c6dc765ca956-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"33286960-496c-4f14-a334-c6dc765ca956\") " pod="openstack/ovsdbserver-sb-0" Oct 02 02:01:56 crc kubenswrapper[4885]: I1002 02:01:56.826384 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-sb-0\" (UID: \"33286960-496c-4f14-a334-c6dc765ca956\") " pod="openstack/ovsdbserver-sb-0" Oct 02 02:01:56 crc kubenswrapper[4885]: I1002 02:01:56.826418 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n5dwt\" (UniqueName: \"kubernetes.io/projected/33286960-496c-4f14-a334-c6dc765ca956-kube-api-access-n5dwt\") pod \"ovsdbserver-sb-0\" (UID: \"33286960-496c-4f14-a334-c6dc765ca956\") " pod="openstack/ovsdbserver-sb-0" Oct 02 02:01:56 crc kubenswrapper[4885]: I1002 02:01:56.826461 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/33286960-496c-4f14-a334-c6dc765ca956-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"33286960-496c-4f14-a334-c6dc765ca956\") " pod="openstack/ovsdbserver-sb-0" Oct 02 02:01:56 crc kubenswrapper[4885]: I1002 02:01:56.826483 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/33286960-496c-4f14-a334-c6dc765ca956-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"33286960-496c-4f14-a334-c6dc765ca956\") " pod="openstack/ovsdbserver-sb-0" Oct 02 02:01:56 crc kubenswrapper[4885]: I1002 02:01:56.826525 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/33286960-496c-4f14-a334-c6dc765ca956-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"33286960-496c-4f14-a334-c6dc765ca956\") " pod="openstack/ovsdbserver-sb-0" Oct 02 02:01:56 crc kubenswrapper[4885]: I1002 02:01:56.887518 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 02 02:01:56 crc kubenswrapper[4885]: I1002 02:01:56.895736 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 02 02:01:56 crc kubenswrapper[4885]: W1002 02:01:56.905958 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod87196135_8d40_47cb_b918_98aea04c7c17.slice/crio-5d1a28c374d6ffa71543ec3a874926563f5ebab19a669270010934e9f4cede69 WatchSource:0}: Error finding container 5d1a28c374d6ffa71543ec3a874926563f5ebab19a669270010934e9f4cede69: Status 404 returned error can't find the container with id 5d1a28c374d6ffa71543ec3a874926563f5ebab19a669270010934e9f4cede69 Oct 02 02:01:56 crc kubenswrapper[4885]: I1002 02:01:56.918897 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-8ljrn" Oct 02 02:01:56 crc kubenswrapper[4885]: I1002 02:01:56.928774 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-tvzhw" Oct 02 02:01:56 crc kubenswrapper[4885]: I1002 02:01:56.929116 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/33286960-496c-4f14-a334-c6dc765ca956-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"33286960-496c-4f14-a334-c6dc765ca956\") " pod="openstack/ovsdbserver-sb-0" Oct 02 02:01:56 crc kubenswrapper[4885]: I1002 02:01:56.929151 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/33286960-496c-4f14-a334-c6dc765ca956-config\") pod \"ovsdbserver-sb-0\" (UID: \"33286960-496c-4f14-a334-c6dc765ca956\") " pod="openstack/ovsdbserver-sb-0" Oct 02 02:01:56 crc kubenswrapper[4885]: I1002 02:01:56.929185 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33286960-496c-4f14-a334-c6dc765ca956-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"33286960-496c-4f14-a334-c6dc765ca956\") " pod="openstack/ovsdbserver-sb-0" Oct 02 02:01:56 crc kubenswrapper[4885]: I1002 02:01:56.929209 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/33286960-496c-4f14-a334-c6dc765ca956-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"33286960-496c-4f14-a334-c6dc765ca956\") " pod="openstack/ovsdbserver-sb-0" Oct 02 02:01:56 crc kubenswrapper[4885]: I1002 02:01:56.929226 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-sb-0\" (UID: \"33286960-496c-4f14-a334-c6dc765ca956\") " pod="openstack/ovsdbserver-sb-0" Oct 02 02:01:56 crc kubenswrapper[4885]: I1002 02:01:56.929251 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n5dwt\" (UniqueName: \"kubernetes.io/projected/33286960-496c-4f14-a334-c6dc765ca956-kube-api-access-n5dwt\") pod \"ovsdbserver-sb-0\" (UID: \"33286960-496c-4f14-a334-c6dc765ca956\") " pod="openstack/ovsdbserver-sb-0" Oct 02 02:01:56 crc kubenswrapper[4885]: I1002 02:01:56.929307 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/33286960-496c-4f14-a334-c6dc765ca956-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"33286960-496c-4f14-a334-c6dc765ca956\") " pod="openstack/ovsdbserver-sb-0" Oct 02 02:01:56 crc kubenswrapper[4885]: I1002 02:01:56.929332 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/33286960-496c-4f14-a334-c6dc765ca956-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"33286960-496c-4f14-a334-c6dc765ca956\") " pod="openstack/ovsdbserver-sb-0" Oct 02 02:01:56 crc kubenswrapper[4885]: I1002 02:01:56.930793 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/33286960-496c-4f14-a334-c6dc765ca956-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"33286960-496c-4f14-a334-c6dc765ca956\") " pod="openstack/ovsdbserver-sb-0" Oct 02 02:01:56 crc kubenswrapper[4885]: I1002 02:01:56.931851 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/33286960-496c-4f14-a334-c6dc765ca956-config\") pod \"ovsdbserver-sb-0\" (UID: \"33286960-496c-4f14-a334-c6dc765ca956\") " pod="openstack/ovsdbserver-sb-0" Oct 02 02:01:56 crc kubenswrapper[4885]: I1002 02:01:56.932328 4885 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-sb-0\" (UID: \"33286960-496c-4f14-a334-c6dc765ca956\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/ovsdbserver-sb-0" Oct 02 02:01:56 crc kubenswrapper[4885]: I1002 02:01:56.933301 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/33286960-496c-4f14-a334-c6dc765ca956-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"33286960-496c-4f14-a334-c6dc765ca956\") " pod="openstack/ovsdbserver-sb-0" Oct 02 02:01:56 crc kubenswrapper[4885]: I1002 02:01:56.943441 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33286960-496c-4f14-a334-c6dc765ca956-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"33286960-496c-4f14-a334-c6dc765ca956\") " pod="openstack/ovsdbserver-sb-0" Oct 02 02:01:56 crc kubenswrapper[4885]: I1002 02:01:56.943920 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/33286960-496c-4f14-a334-c6dc765ca956-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"33286960-496c-4f14-a334-c6dc765ca956\") " pod="openstack/ovsdbserver-sb-0" Oct 02 02:01:56 crc kubenswrapper[4885]: I1002 02:01:56.950818 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/33286960-496c-4f14-a334-c6dc765ca956-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"33286960-496c-4f14-a334-c6dc765ca956\") " pod="openstack/ovsdbserver-sb-0" Oct 02 02:01:56 crc kubenswrapper[4885]: I1002 02:01:56.953662 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-sb-0\" (UID: \"33286960-496c-4f14-a334-c6dc765ca956\") " pod="openstack/ovsdbserver-sb-0" Oct 02 02:01:56 crc kubenswrapper[4885]: I1002 02:01:56.956867 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n5dwt\" (UniqueName: \"kubernetes.io/projected/33286960-496c-4f14-a334-c6dc765ca956-kube-api-access-n5dwt\") pod \"ovsdbserver-sb-0\" (UID: \"33286960-496c-4f14-a334-c6dc765ca956\") " pod="openstack/ovsdbserver-sb-0" Oct 02 02:01:57 crc kubenswrapper[4885]: I1002 02:01:57.030374 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gpvqj\" (UniqueName: \"kubernetes.io/projected/10deeddc-732a-4154-a74c-b28abc841bc1-kube-api-access-gpvqj\") pod \"10deeddc-732a-4154-a74c-b28abc841bc1\" (UID: \"10deeddc-732a-4154-a74c-b28abc841bc1\") " Oct 02 02:01:57 crc kubenswrapper[4885]: I1002 02:01:57.030933 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5w66h\" (UniqueName: \"kubernetes.io/projected/f541a783-9596-4275-b87e-cf289b223f9d-kube-api-access-5w66h\") pod \"f541a783-9596-4275-b87e-cf289b223f9d\" (UID: \"f541a783-9596-4275-b87e-cf289b223f9d\") " Oct 02 02:01:57 crc kubenswrapper[4885]: I1002 02:01:57.031024 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f541a783-9596-4275-b87e-cf289b223f9d-dns-svc\") pod \"f541a783-9596-4275-b87e-cf289b223f9d\" (UID: \"f541a783-9596-4275-b87e-cf289b223f9d\") " Oct 02 02:01:57 crc kubenswrapper[4885]: I1002 02:01:57.031043 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f541a783-9596-4275-b87e-cf289b223f9d-config\") pod \"f541a783-9596-4275-b87e-cf289b223f9d\" (UID: \"f541a783-9596-4275-b87e-cf289b223f9d\") " Oct 02 02:01:57 crc kubenswrapper[4885]: I1002 02:01:57.031112 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10deeddc-732a-4154-a74c-b28abc841bc1-config\") pod \"10deeddc-732a-4154-a74c-b28abc841bc1\" (UID: \"10deeddc-732a-4154-a74c-b28abc841bc1\") " Oct 02 02:01:57 crc kubenswrapper[4885]: I1002 02:01:57.031490 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f541a783-9596-4275-b87e-cf289b223f9d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f541a783-9596-4275-b87e-cf289b223f9d" (UID: "f541a783-9596-4275-b87e-cf289b223f9d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:01:57 crc kubenswrapper[4885]: I1002 02:01:57.031687 4885 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f541a783-9596-4275-b87e-cf289b223f9d-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 02:01:57 crc kubenswrapper[4885]: I1002 02:01:57.031837 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f541a783-9596-4275-b87e-cf289b223f9d-config" (OuterVolumeSpecName: "config") pod "f541a783-9596-4275-b87e-cf289b223f9d" (UID: "f541a783-9596-4275-b87e-cf289b223f9d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:01:57 crc kubenswrapper[4885]: I1002 02:01:57.032239 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/10deeddc-732a-4154-a74c-b28abc841bc1-config" (OuterVolumeSpecName: "config") pod "10deeddc-732a-4154-a74c-b28abc841bc1" (UID: "10deeddc-732a-4154-a74c-b28abc841bc1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:01:57 crc kubenswrapper[4885]: I1002 02:01:57.034561 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/10deeddc-732a-4154-a74c-b28abc841bc1-kube-api-access-gpvqj" (OuterVolumeSpecName: "kube-api-access-gpvqj") pod "10deeddc-732a-4154-a74c-b28abc841bc1" (UID: "10deeddc-732a-4154-a74c-b28abc841bc1"). InnerVolumeSpecName "kube-api-access-gpvqj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:01:57 crc kubenswrapper[4885]: I1002 02:01:57.037448 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f541a783-9596-4275-b87e-cf289b223f9d-kube-api-access-5w66h" (OuterVolumeSpecName: "kube-api-access-5w66h") pod "f541a783-9596-4275-b87e-cf289b223f9d" (UID: "f541a783-9596-4275-b87e-cf289b223f9d"). InnerVolumeSpecName "kube-api-access-5w66h". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:01:57 crc kubenswrapper[4885]: I1002 02:01:57.070868 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 02 02:01:57 crc kubenswrapper[4885]: I1002 02:01:57.134629 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gpvqj\" (UniqueName: \"kubernetes.io/projected/10deeddc-732a-4154-a74c-b28abc841bc1-kube-api-access-gpvqj\") on node \"crc\" DevicePath \"\"" Oct 02 02:01:57 crc kubenswrapper[4885]: I1002 02:01:57.134900 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5w66h\" (UniqueName: \"kubernetes.io/projected/f541a783-9596-4275-b87e-cf289b223f9d-kube-api-access-5w66h\") on node \"crc\" DevicePath \"\"" Oct 02 02:01:57 crc kubenswrapper[4885]: I1002 02:01:57.134912 4885 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f541a783-9596-4275-b87e-cf289b223f9d-config\") on node \"crc\" DevicePath \"\"" Oct 02 02:01:57 crc kubenswrapper[4885]: I1002 02:01:57.134921 4885 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10deeddc-732a-4154-a74c-b28abc841bc1-config\") on node \"crc\" DevicePath \"\"" Oct 02 02:01:57 crc kubenswrapper[4885]: I1002 02:01:57.285815 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 02 02:01:57 crc kubenswrapper[4885]: W1002 02:01:57.287222 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfaff5405_00c6_4d75_bbca_77f9230c2811.slice/crio-49bd68dd50e5f1991581195877fe3325d00d374d02ade0c4582071dc275342dc WatchSource:0}: Error finding container 49bd68dd50e5f1991581195877fe3325d00d374d02ade0c4582071dc275342dc: Status 404 returned error can't find the container with id 49bd68dd50e5f1991581195877fe3325d00d374d02ade0c4582071dc275342dc Oct 02 02:01:57 crc kubenswrapper[4885]: I1002 02:01:57.293644 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 02 02:01:57 crc kubenswrapper[4885]: I1002 02:01:57.305379 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 02 02:01:57 crc kubenswrapper[4885]: W1002 02:01:57.316572 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd798d658_f60b_4e23_8753_b3de580e44b0.slice/crio-6802beb70c60eb3adc182dc85ba2f98fea9efaf69df87e783bb7930ff85ac161 WatchSource:0}: Error finding container 6802beb70c60eb3adc182dc85ba2f98fea9efaf69df87e783bb7930ff85ac161: Status 404 returned error can't find the container with id 6802beb70c60eb3adc182dc85ba2f98fea9efaf69df87e783bb7930ff85ac161 Oct 02 02:01:57 crc kubenswrapper[4885]: W1002 02:01:57.316863 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8427e298_965b_48f0_821f_bd078fa2e96f.slice/crio-5e9b7ac772a38fa33362f019dfa6067c07ede09e15cb8cbf5ec165ab3eb06fbf WatchSource:0}: Error finding container 5e9b7ac772a38fa33362f019dfa6067c07ede09e15cb8cbf5ec165ab3eb06fbf: Status 404 returned error can't find the container with id 5e9b7ac772a38fa33362f019dfa6067c07ede09e15cb8cbf5ec165ab3eb06fbf Oct 02 02:01:57 crc kubenswrapper[4885]: I1002 02:01:57.329511 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 02 02:01:57 crc kubenswrapper[4885]: I1002 02:01:57.338415 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-fgvpt"] Oct 02 02:01:57 crc kubenswrapper[4885]: I1002 02:01:57.403626 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 02 02:01:57 crc kubenswrapper[4885]: W1002 02:01:57.414992 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod079e125d_c938_4962_8f41_09b3109fb0ae.slice/crio-8dc9389425773f955b5ee09adb61ba0f4121fc12469908efe1d7e98bedc1d45b WatchSource:0}: Error finding container 8dc9389425773f955b5ee09adb61ba0f4121fc12469908efe1d7e98bedc1d45b: Status 404 returned error can't find the container with id 8dc9389425773f955b5ee09adb61ba0f4121fc12469908efe1d7e98bedc1d45b Oct 02 02:01:57 crc kubenswrapper[4885]: I1002 02:01:57.447473 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"079e125d-c938-4962-8f41-09b3109fb0ae","Type":"ContainerStarted","Data":"8dc9389425773f955b5ee09adb61ba0f4121fc12469908efe1d7e98bedc1d45b"} Oct 02 02:01:57 crc kubenswrapper[4885]: I1002 02:01:57.448593 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"6692f52c-5ee9-4cd6-8674-0fb3f5fd6679","Type":"ContainerStarted","Data":"212fc43e2e0d43911b7a078d43f599e155a6d9739da223da5377570980156bd5"} Oct 02 02:01:57 crc kubenswrapper[4885]: I1002 02:01:57.449452 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-8ljrn" Oct 02 02:01:57 crc kubenswrapper[4885]: I1002 02:01:57.449690 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-8ljrn" event={"ID":"f541a783-9596-4275-b87e-cf289b223f9d","Type":"ContainerDied","Data":"6daa8fbbc09a7b09ecc20214ccff27d937a0cf770c37df7a629593e92a06b7be"} Oct 02 02:01:57 crc kubenswrapper[4885]: I1002 02:01:57.450611 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"87196135-8d40-47cb-b918-98aea04c7c17","Type":"ContainerStarted","Data":"5d1a28c374d6ffa71543ec3a874926563f5ebab19a669270010934e9f4cede69"} Oct 02 02:01:57 crc kubenswrapper[4885]: I1002 02:01:57.452416 4885 generic.go:334] "Generic (PLEG): container finished" podID="32a76623-c272-48c1-85e6-2170cbb017cc" containerID="50afcde965796330d215a414642001d2420758eddba6a7fca8ef03537d7303ef" exitCode=0 Oct 02 02:01:57 crc kubenswrapper[4885]: I1002 02:01:57.452590 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-5wrwj" event={"ID":"32a76623-c272-48c1-85e6-2170cbb017cc","Type":"ContainerDied","Data":"50afcde965796330d215a414642001d2420758eddba6a7fca8ef03537d7303ef"} Oct 02 02:01:57 crc kubenswrapper[4885]: I1002 02:01:57.455309 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"eea66006-67ba-45e3-bdf9-9d144d772386","Type":"ContainerStarted","Data":"58d97479733782d79442e6a3ba3ed7515c1bff74cd50689b956caf317562f0b9"} Oct 02 02:01:57 crc kubenswrapper[4885]: I1002 02:01:57.459115 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"d798d658-f60b-4e23-8753-b3de580e44b0","Type":"ContainerStarted","Data":"6802beb70c60eb3adc182dc85ba2f98fea9efaf69df87e783bb7930ff85ac161"} Oct 02 02:01:57 crc kubenswrapper[4885]: I1002 02:01:57.463114 4885 generic.go:334] "Generic (PLEG): container finished" podID="90626475-ee00-431f-8992-d5fc20728c6c" containerID="35fff824d306ebab1e80c6e4cb5ff69baa9c27248d26050fa3683c38aa8b261c" exitCode=0 Oct 02 02:01:57 crc kubenswrapper[4885]: I1002 02:01:57.463582 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-p9cqp" event={"ID":"90626475-ee00-431f-8992-d5fc20728c6c","Type":"ContainerDied","Data":"35fff824d306ebab1e80c6e4cb5ff69baa9c27248d26050fa3683c38aa8b261c"} Oct 02 02:01:57 crc kubenswrapper[4885]: I1002 02:01:57.466064 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"8427e298-965b-48f0-821f-bd078fa2e96f","Type":"ContainerStarted","Data":"5e9b7ac772a38fa33362f019dfa6067c07ede09e15cb8cbf5ec165ab3eb06fbf"} Oct 02 02:01:57 crc kubenswrapper[4885]: I1002 02:01:57.467351 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-tvzhw" event={"ID":"10deeddc-732a-4154-a74c-b28abc841bc1","Type":"ContainerDied","Data":"ec967b0b3cc64faa97c93ccaee1a4381e0d8c51f24e760935bc1cfa152d554bf"} Oct 02 02:01:57 crc kubenswrapper[4885]: I1002 02:01:57.467418 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-tvzhw" Oct 02 02:01:57 crc kubenswrapper[4885]: I1002 02:01:57.469505 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-fgvpt" event={"ID":"924fb321-4073-449a-a546-811b046ed26a","Type":"ContainerStarted","Data":"d37a7bbfee8b19c651f50c38b3149c4ea4d01d99447fba8416ee675cdff3e9da"} Oct 02 02:01:57 crc kubenswrapper[4885]: I1002 02:01:57.484599 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"faff5405-00c6-4d75-bbca-77f9230c2811","Type":"ContainerStarted","Data":"49bd68dd50e5f1991581195877fe3325d00d374d02ade0c4582071dc275342dc"} Oct 02 02:01:57 crc kubenswrapper[4885]: I1002 02:01:57.524595 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-8ljrn"] Oct 02 02:01:57 crc kubenswrapper[4885]: I1002 02:01:57.531995 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-8ljrn"] Oct 02 02:01:57 crc kubenswrapper[4885]: W1002 02:01:57.537865 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb65df3c4_cc33_43da_9bf1_8f46a480f7eb.slice/crio-f272c3c34daf1d9f35c3a1fd0f9ded4336957a7a2faf83b86ad40e153ebd31c1 WatchSource:0}: Error finding container f272c3c34daf1d9f35c3a1fd0f9ded4336957a7a2faf83b86ad40e153ebd31c1: Status 404 returned error can't find the container with id f272c3c34daf1d9f35c3a1fd0f9ded4336957a7a2faf83b86ad40e153ebd31c1 Oct 02 02:01:57 crc kubenswrapper[4885]: I1002 02:01:57.543078 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-q92rb"] Oct 02 02:01:57 crc kubenswrapper[4885]: I1002 02:01:57.573925 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-tvzhw"] Oct 02 02:01:57 crc kubenswrapper[4885]: I1002 02:01:57.577990 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-tvzhw"] Oct 02 02:01:57 crc kubenswrapper[4885]: E1002 02:01:57.644554 4885 log.go:32] "CreateContainer in sandbox from runtime service failed" err=< Oct 02 02:01:57 crc kubenswrapper[4885]: rpc error: code = Unknown desc = container create failed: mount `/var/lib/kubelet/pods/32a76623-c272-48c1-85e6-2170cbb017cc/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Oct 02 02:01:57 crc kubenswrapper[4885]: > podSandboxID="a75ce6ed97ce5e957c0426fdc15f2a177449241a425a66e15fb3b8edc6443ad6" Oct 02 02:01:57 crc kubenswrapper[4885]: E1002 02:01:57.644690 4885 kuberuntime_manager.go:1274] "Unhandled Error" err=< Oct 02 02:01:57 crc kubenswrapper[4885]: container &Container{Name:dnsmasq-dns,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nfdh5dfhb6h64h676hc4h78h97h669h54chfbh696hb5h54bh5d4h6bh64h644h677h584h5cbh698h9dh5bbh5f8h5b8hcdh644h5c7h694hbfh589q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gf6c6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-5ccc8479f9-5wrwj_openstack(32a76623-c272-48c1-85e6-2170cbb017cc): CreateContainerError: container create failed: mount `/var/lib/kubelet/pods/32a76623-c272-48c1-85e6-2170cbb017cc/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Oct 02 02:01:57 crc kubenswrapper[4885]: > logger="UnhandledError" Oct 02 02:01:57 crc kubenswrapper[4885]: E1002 02:01:57.646308 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"dnsmasq-dns\" with CreateContainerError: \"container create failed: mount `/var/lib/kubelet/pods/32a76623-c272-48c1-85e6-2170cbb017cc/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory\\n\"" pod="openstack/dnsmasq-dns-5ccc8479f9-5wrwj" podUID="32a76623-c272-48c1-85e6-2170cbb017cc" Oct 02 02:01:58 crc kubenswrapper[4885]: I1002 02:01:58.066124 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="10deeddc-732a-4154-a74c-b28abc841bc1" path="/var/lib/kubelet/pods/10deeddc-732a-4154-a74c-b28abc841bc1/volumes" Oct 02 02:01:58 crc kubenswrapper[4885]: I1002 02:01:58.066609 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f541a783-9596-4275-b87e-cf289b223f9d" path="/var/lib/kubelet/pods/f541a783-9596-4275-b87e-cf289b223f9d/volumes" Oct 02 02:01:58 crc kubenswrapper[4885]: I1002 02:01:58.236382 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 02 02:01:58 crc kubenswrapper[4885]: W1002 02:01:58.245176 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod33286960_496c_4f14_a334_c6dc765ca956.slice/crio-501743a11ebc7670176846d8779b9e3309b56b4a2f70f7dc8d5cfddb1c5dd159 WatchSource:0}: Error finding container 501743a11ebc7670176846d8779b9e3309b56b4a2f70f7dc8d5cfddb1c5dd159: Status 404 returned error can't find the container with id 501743a11ebc7670176846d8779b9e3309b56b4a2f70f7dc8d5cfddb1c5dd159 Oct 02 02:01:58 crc kubenswrapper[4885]: I1002 02:01:58.496761 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"33286960-496c-4f14-a334-c6dc765ca956","Type":"ContainerStarted","Data":"501743a11ebc7670176846d8779b9e3309b56b4a2f70f7dc8d5cfddb1c5dd159"} Oct 02 02:01:58 crc kubenswrapper[4885]: I1002 02:01:58.499214 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-p9cqp" event={"ID":"90626475-ee00-431f-8992-d5fc20728c6c","Type":"ContainerStarted","Data":"3b8c253cf16ce810fd1834840e2d1116620e42562f6f50eea87012550966a7b3"} Oct 02 02:01:58 crc kubenswrapper[4885]: I1002 02:01:58.499313 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57d769cc4f-p9cqp" Oct 02 02:01:58 crc kubenswrapper[4885]: I1002 02:01:58.501930 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-q92rb" event={"ID":"b65df3c4-cc33-43da-9bf1-8f46a480f7eb","Type":"ContainerStarted","Data":"f272c3c34daf1d9f35c3a1fd0f9ded4336957a7a2faf83b86ad40e153ebd31c1"} Oct 02 02:01:58 crc kubenswrapper[4885]: I1002 02:01:58.515462 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57d769cc4f-p9cqp" podStartSLOduration=7.329005664 podStartE2EDuration="16.515447488s" podCreationTimestamp="2025-10-02 02:01:42 +0000 UTC" firstStartedPulling="2025-10-02 02:01:47.054693213 +0000 UTC m=+895.866440622" lastFinishedPulling="2025-10-02 02:01:56.241135047 +0000 UTC m=+905.052882446" observedRunningTime="2025-10-02 02:01:58.514631563 +0000 UTC m=+907.326378962" watchObservedRunningTime="2025-10-02 02:01:58.515447488 +0000 UTC m=+907.327194887" Oct 02 02:02:01 crc kubenswrapper[4885]: I1002 02:02:01.528877 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-5wrwj" event={"ID":"32a76623-c272-48c1-85e6-2170cbb017cc","Type":"ContainerStarted","Data":"e262a05c4ba0f70d0f1b9b0467184e6c4fde9a89c2ce17b8d4383a08864940d7"} Oct 02 02:02:01 crc kubenswrapper[4885]: I1002 02:02:01.529508 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5ccc8479f9-5wrwj" Oct 02 02:02:02 crc kubenswrapper[4885]: I1002 02:02:02.085574 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5ccc8479f9-5wrwj" podStartSLOduration=7.392065072 podStartE2EDuration="20.085557585s" podCreationTimestamp="2025-10-02 02:01:42 +0000 UTC" firstStartedPulling="2025-10-02 02:01:43.531014568 +0000 UTC m=+892.342761967" lastFinishedPulling="2025-10-02 02:01:56.224507081 +0000 UTC m=+905.036254480" observedRunningTime="2025-10-02 02:02:01.552181335 +0000 UTC m=+910.363928774" watchObservedRunningTime="2025-10-02 02:02:02.085557585 +0000 UTC m=+910.897304984" Oct 02 02:02:03 crc kubenswrapper[4885]: I1002 02:02:03.215324 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-57d769cc4f-p9cqp" Oct 02 02:02:03 crc kubenswrapper[4885]: I1002 02:02:03.264055 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-5wrwj"] Oct 02 02:02:03 crc kubenswrapper[4885]: I1002 02:02:03.541668 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5ccc8479f9-5wrwj" podUID="32a76623-c272-48c1-85e6-2170cbb017cc" containerName="dnsmasq-dns" containerID="cri-o://e262a05c4ba0f70d0f1b9b0467184e6c4fde9a89c2ce17b8d4383a08864940d7" gracePeriod=10 Oct 02 02:02:04 crc kubenswrapper[4885]: I1002 02:02:04.560108 4885 generic.go:334] "Generic (PLEG): container finished" podID="32a76623-c272-48c1-85e6-2170cbb017cc" containerID="e262a05c4ba0f70d0f1b9b0467184e6c4fde9a89c2ce17b8d4383a08864940d7" exitCode=0 Oct 02 02:02:04 crc kubenswrapper[4885]: I1002 02:02:04.560255 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-5wrwj" event={"ID":"32a76623-c272-48c1-85e6-2170cbb017cc","Type":"ContainerDied","Data":"e262a05c4ba0f70d0f1b9b0467184e6c4fde9a89c2ce17b8d4383a08864940d7"} Oct 02 02:02:12 crc kubenswrapper[4885]: I1002 02:02:12.979823 4885 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5ccc8479f9-5wrwj" podUID="32a76623-c272-48c1-85e6-2170cbb017cc" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.99:5353: i/o timeout" Oct 02 02:02:13 crc kubenswrapper[4885]: I1002 02:02:13.265947 4885 patch_prober.go:28] interesting pod/machine-config-daemon-rttx8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 02:02:13 crc kubenswrapper[4885]: I1002 02:02:13.266314 4885 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 02:02:16 crc kubenswrapper[4885]: E1002 02:02:16.230849 4885 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ovn-sb-db-server:current-podified" Oct 02 02:02:16 crc kubenswrapper[4885]: E1002 02:02:16.231041 4885 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ovsdbserver-sb,Image:quay.io/podified-antelope-centos9/openstack-ovn-sb-db-server:current-podified,Command:[/usr/bin/dumb-init],Args:[/usr/local/bin/container-scripts/setup.sh],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n58h589hdbhb4hf9h5c8h667h5d4h598h6dh58dh56ch648h677h58bh577h68ch555h89h647h5h67ch654hf9h567hdfhdch65hd9h679h685h675q,ValueFrom:nil,},EnvVar{Name:OVN_LOGDIR,Value:/tmp,ValueFrom:nil,},EnvVar{Name:OVN_RUNDIR,Value:/tmp,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovndbcluster-sb-etc-ovn,ReadOnly:false,MountPath:/etc/ovn,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdb-rundir,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-sb-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovndb.crt,SubPath:tls.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-sb-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/private/ovndb.key,SubPath:tls.key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-sb-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovndbca.crt,SubPath:ca.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-n5dwt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/pidof ovsdb-server],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/pidof ovsdb-server],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:&Lifecycle{PostStart:nil,PreStop:&LifecycleHandler{Exec:&ExecAction{Command:[/usr/local/bin/container-scripts/cleanup.sh],},HTTPGet:nil,TCPSocket:nil,Sleep:nil,},},TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/pidof ovsdb-server],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:20,TerminationGracePeriodSeconds:nil,},ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovsdbserver-sb-0_openstack(33286960-496c-4f14-a334-c6dc765ca956): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 02:02:16 crc kubenswrapper[4885]: I1002 02:02:16.350701 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-5wrwj" Oct 02 02:02:16 crc kubenswrapper[4885]: I1002 02:02:16.402853 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/32a76623-c272-48c1-85e6-2170cbb017cc-config\") pod \"32a76623-c272-48c1-85e6-2170cbb017cc\" (UID: \"32a76623-c272-48c1-85e6-2170cbb017cc\") " Oct 02 02:02:16 crc kubenswrapper[4885]: I1002 02:02:16.403237 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf6c6\" (UniqueName: \"kubernetes.io/projected/32a76623-c272-48c1-85e6-2170cbb017cc-kube-api-access-gf6c6\") pod \"32a76623-c272-48c1-85e6-2170cbb017cc\" (UID: \"32a76623-c272-48c1-85e6-2170cbb017cc\") " Oct 02 02:02:16 crc kubenswrapper[4885]: I1002 02:02:16.403286 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/32a76623-c272-48c1-85e6-2170cbb017cc-dns-svc\") pod \"32a76623-c272-48c1-85e6-2170cbb017cc\" (UID: \"32a76623-c272-48c1-85e6-2170cbb017cc\") " Oct 02 02:02:16 crc kubenswrapper[4885]: I1002 02:02:16.409417 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32a76623-c272-48c1-85e6-2170cbb017cc-kube-api-access-gf6c6" (OuterVolumeSpecName: "kube-api-access-gf6c6") pod "32a76623-c272-48c1-85e6-2170cbb017cc" (UID: "32a76623-c272-48c1-85e6-2170cbb017cc"). InnerVolumeSpecName "kube-api-access-gf6c6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:02:16 crc kubenswrapper[4885]: I1002 02:02:16.460332 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/32a76623-c272-48c1-85e6-2170cbb017cc-config" (OuterVolumeSpecName: "config") pod "32a76623-c272-48c1-85e6-2170cbb017cc" (UID: "32a76623-c272-48c1-85e6-2170cbb017cc"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:02:16 crc kubenswrapper[4885]: I1002 02:02:16.463170 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/32a76623-c272-48c1-85e6-2170cbb017cc-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "32a76623-c272-48c1-85e6-2170cbb017cc" (UID: "32a76623-c272-48c1-85e6-2170cbb017cc"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:02:16 crc kubenswrapper[4885]: I1002 02:02:16.505442 4885 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/32a76623-c272-48c1-85e6-2170cbb017cc-config\") on node \"crc\" DevicePath \"\"" Oct 02 02:02:16 crc kubenswrapper[4885]: I1002 02:02:16.505486 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf6c6\" (UniqueName: \"kubernetes.io/projected/32a76623-c272-48c1-85e6-2170cbb017cc-kube-api-access-gf6c6\") on node \"crc\" DevicePath \"\"" Oct 02 02:02:16 crc kubenswrapper[4885]: I1002 02:02:16.505495 4885 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/32a76623-c272-48c1-85e6-2170cbb017cc-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 02:02:16 crc kubenswrapper[4885]: I1002 02:02:16.679419 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-5wrwj" event={"ID":"32a76623-c272-48c1-85e6-2170cbb017cc","Type":"ContainerDied","Data":"a75ce6ed97ce5e957c0426fdc15f2a177449241a425a66e15fb3b8edc6443ad6"} Oct 02 02:02:16 crc kubenswrapper[4885]: I1002 02:02:16.679495 4885 scope.go:117] "RemoveContainer" containerID="e262a05c4ba0f70d0f1b9b0467184e6c4fde9a89c2ce17b8d4383a08864940d7" Oct 02 02:02:16 crc kubenswrapper[4885]: I1002 02:02:16.679506 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-5wrwj" Oct 02 02:02:16 crc kubenswrapper[4885]: I1002 02:02:16.722844 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-5wrwj"] Oct 02 02:02:16 crc kubenswrapper[4885]: I1002 02:02:16.732154 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-5wrwj"] Oct 02 02:02:16 crc kubenswrapper[4885]: I1002 02:02:16.734807 4885 scope.go:117] "RemoveContainer" containerID="50afcde965796330d215a414642001d2420758eddba6a7fca8ef03537d7303ef" Oct 02 02:02:17 crc kubenswrapper[4885]: E1002 02:02:17.654894 4885 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0" Oct 02 02:02:17 crc kubenswrapper[4885]: E1002 02:02:17.654966 4885 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0" Oct 02 02:02:17 crc kubenswrapper[4885]: E1002 02:02:17.655145 4885 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-state-metrics,Image:registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0,Command:[],Args:[--resources=pods --namespaces=openstack],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:http-metrics,HostPort:0,ContainerPort:8080,Protocol:TCP,HostIP:,},ContainerPort{Name:telemetry,HostPort:0,ContainerPort:8081,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-2ngf5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/livez,Port:{0 8080 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:*true,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod kube-state-metrics-0_openstack(faff5405-00c6-4d75-bbca-77f9230c2811): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 02 02:02:17 crc kubenswrapper[4885]: E1002 02:02:17.656342 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-state-metrics\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openstack/kube-state-metrics-0" podUID="faff5405-00c6-4d75-bbca-77f9230c2811" Oct 02 02:02:17 crc kubenswrapper[4885]: E1002 02:02:17.694239 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-state-metrics\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0\\\"\"" pod="openstack/kube-state-metrics-0" podUID="faff5405-00c6-4d75-bbca-77f9230c2811" Oct 02 02:02:17 crc kubenswrapper[4885]: I1002 02:02:17.981537 4885 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5ccc8479f9-5wrwj" podUID="32a76623-c272-48c1-85e6-2170cbb017cc" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.99:5353: i/o timeout" Oct 02 02:02:18 crc kubenswrapper[4885]: I1002 02:02:18.058037 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="32a76623-c272-48c1-85e6-2170cbb017cc" path="/var/lib/kubelet/pods/32a76623-c272-48c1-85e6-2170cbb017cc/volumes" Oct 02 02:02:18 crc kubenswrapper[4885]: I1002 02:02:18.702930 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-q92rb" event={"ID":"b65df3c4-cc33-43da-9bf1-8f46a480f7eb","Type":"ContainerStarted","Data":"aa3bb60d5146804dcc7739e159fcb0dcabf35d3ff9acca4382ba96d6d01ba8b0"} Oct 02 02:02:18 crc kubenswrapper[4885]: I1002 02:02:18.706639 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"8427e298-965b-48f0-821f-bd078fa2e96f","Type":"ContainerStarted","Data":"f068f8c5a1d031159cc0d6e4c94d1b871a16e9160385c1912db8f0e8cb668289"} Oct 02 02:02:18 crc kubenswrapper[4885]: I1002 02:02:18.708971 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"87196135-8d40-47cb-b918-98aea04c7c17","Type":"ContainerStarted","Data":"9d57810a37a1cca799c557440d8bd72c570370f072692434ad16f328610f95b0"} Oct 02 02:02:18 crc kubenswrapper[4885]: I1002 02:02:18.714378 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-fgvpt" event={"ID":"924fb321-4073-449a-a546-811b046ed26a","Type":"ContainerStarted","Data":"72eccd4618708438c887f259d424e770c6ed5f14bd6c667766116d585466c9a0"} Oct 02 02:02:18 crc kubenswrapper[4885]: I1002 02:02:18.714847 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-fgvpt" Oct 02 02:02:18 crc kubenswrapper[4885]: I1002 02:02:18.716168 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"d798d658-f60b-4e23-8753-b3de580e44b0","Type":"ContainerStarted","Data":"fdb0e73a516549fefaa9d0817bcdc20c8c1121b16851760bfb289be26d977302"} Oct 02 02:02:18 crc kubenswrapper[4885]: I1002 02:02:18.716235 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Oct 02 02:02:18 crc kubenswrapper[4885]: I1002 02:02:18.719570 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"079e125d-c938-4962-8f41-09b3109fb0ae","Type":"ContainerStarted","Data":"02dd4fb0a5d37bfd5357ed24a29c1ac7aac0cf404f4061488f40749c7375e43e"} Oct 02 02:02:18 crc kubenswrapper[4885]: I1002 02:02:18.792523 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-fgvpt" podStartSLOduration=7.902088279 podStartE2EDuration="26.792508393s" podCreationTimestamp="2025-10-02 02:01:52 +0000 UTC" firstStartedPulling="2025-10-02 02:01:57.340915391 +0000 UTC m=+906.152662790" lastFinishedPulling="2025-10-02 02:02:16.231335505 +0000 UTC m=+925.043082904" observedRunningTime="2025-10-02 02:02:18.787394017 +0000 UTC m=+927.599141416" watchObservedRunningTime="2025-10-02 02:02:18.792508393 +0000 UTC m=+927.604255792" Oct 02 02:02:18 crc kubenswrapper[4885]: I1002 02:02:18.841254 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=12.428967673 podStartE2EDuration="31.841237315s" podCreationTimestamp="2025-10-02 02:01:47 +0000 UTC" firstStartedPulling="2025-10-02 02:01:57.322544252 +0000 UTC m=+906.134291651" lastFinishedPulling="2025-10-02 02:02:16.734813884 +0000 UTC m=+925.546561293" observedRunningTime="2025-10-02 02:02:18.832436708 +0000 UTC m=+927.644184127" watchObservedRunningTime="2025-10-02 02:02:18.841237315 +0000 UTC m=+927.652984714" Oct 02 02:02:19 crc kubenswrapper[4885]: I1002 02:02:19.732592 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"6692f52c-5ee9-4cd6-8674-0fb3f5fd6679","Type":"ContainerStarted","Data":"a52ec36756e9b2deb9a3362cfb43912af369dd3918986e7fd40070a1385f68f0"} Oct 02 02:02:19 crc kubenswrapper[4885]: I1002 02:02:19.734362 4885 generic.go:334] "Generic (PLEG): container finished" podID="b65df3c4-cc33-43da-9bf1-8f46a480f7eb" containerID="aa3bb60d5146804dcc7739e159fcb0dcabf35d3ff9acca4382ba96d6d01ba8b0" exitCode=0 Oct 02 02:02:19 crc kubenswrapper[4885]: I1002 02:02:19.734464 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-q92rb" event={"ID":"b65df3c4-cc33-43da-9bf1-8f46a480f7eb","Type":"ContainerDied","Data":"aa3bb60d5146804dcc7739e159fcb0dcabf35d3ff9acca4382ba96d6d01ba8b0"} Oct 02 02:02:19 crc kubenswrapper[4885]: I1002 02:02:19.738583 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"eea66006-67ba-45e3-bdf9-9d144d772386","Type":"ContainerStarted","Data":"6f3eaeb38ec37be6825c9165f2dca00ced63746de7714d811aaa270ddf582da9"} Oct 02 02:02:21 crc kubenswrapper[4885]: E1002 02:02:21.680800 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovsdbserver-sb\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ovsdbserver-sb-0" podUID="33286960-496c-4f14-a334-c6dc765ca956" Oct 02 02:02:21 crc kubenswrapper[4885]: I1002 02:02:21.755114 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"079e125d-c938-4962-8f41-09b3109fb0ae","Type":"ContainerStarted","Data":"7ea7ea2b5c311a3809dc71aee9b791da234e49df852b4cfb4e2de180bf5b0de7"} Oct 02 02:02:21 crc kubenswrapper[4885]: I1002 02:02:21.759035 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"33286960-496c-4f14-a334-c6dc765ca956","Type":"ContainerStarted","Data":"d42d975a475126f863952501c9cd2af2a350f12e786c45b4d7e0a1672541c54e"} Oct 02 02:02:21 crc kubenswrapper[4885]: I1002 02:02:21.762599 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-q92rb" event={"ID":"b65df3c4-cc33-43da-9bf1-8f46a480f7eb","Type":"ContainerStarted","Data":"722e6ec997f48af4199f37c9bbbba51670426cb20cac01e900e5323cc11bc1df"} Oct 02 02:02:21 crc kubenswrapper[4885]: E1002 02:02:21.763303 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovsdbserver-sb\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-ovn-sb-db-server:current-podified\\\"\"" pod="openstack/ovsdbserver-sb-0" podUID="33286960-496c-4f14-a334-c6dc765ca956" Oct 02 02:02:21 crc kubenswrapper[4885]: I1002 02:02:21.780161 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=5.6619146879999995 podStartE2EDuration="29.780142468s" podCreationTimestamp="2025-10-02 02:01:52 +0000 UTC" firstStartedPulling="2025-10-02 02:01:57.417349887 +0000 UTC m=+906.229097286" lastFinishedPulling="2025-10-02 02:02:21.535577657 +0000 UTC m=+930.347325066" observedRunningTime="2025-10-02 02:02:21.777863279 +0000 UTC m=+930.589610678" watchObservedRunningTime="2025-10-02 02:02:21.780142468 +0000 UTC m=+930.591889867" Oct 02 02:02:22 crc kubenswrapper[4885]: I1002 02:02:22.774914 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-q92rb" event={"ID":"b65df3c4-cc33-43da-9bf1-8f46a480f7eb","Type":"ContainerStarted","Data":"f64e8e156cdd3e72f949352d40737e334d54a58474d23d13884c7ca64ce7684e"} Oct 02 02:02:22 crc kubenswrapper[4885]: I1002 02:02:22.775530 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-q92rb" Oct 02 02:02:22 crc kubenswrapper[4885]: I1002 02:02:22.775558 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-q92rb" Oct 02 02:02:22 crc kubenswrapper[4885]: E1002 02:02:22.777052 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovsdbserver-sb\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-ovn-sb-db-server:current-podified\\\"\"" pod="openstack/ovsdbserver-sb-0" podUID="33286960-496c-4f14-a334-c6dc765ca956" Oct 02 02:02:22 crc kubenswrapper[4885]: I1002 02:02:22.819596 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Oct 02 02:02:22 crc kubenswrapper[4885]: I1002 02:02:22.866032 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-q92rb" podStartSLOduration=12.178526668 podStartE2EDuration="30.866006937s" podCreationTimestamp="2025-10-02 02:01:52 +0000 UTC" firstStartedPulling="2025-10-02 02:01:57.547841497 +0000 UTC m=+906.359588906" lastFinishedPulling="2025-10-02 02:02:16.235321776 +0000 UTC m=+925.047069175" observedRunningTime="2025-10-02 02:02:22.859033475 +0000 UTC m=+931.670780904" watchObservedRunningTime="2025-10-02 02:02:22.866006937 +0000 UTC m=+931.677754376" Oct 02 02:02:23 crc kubenswrapper[4885]: I1002 02:02:23.854837 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Oct 02 02:02:23 crc kubenswrapper[4885]: I1002 02:02:23.854896 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Oct 02 02:02:23 crc kubenswrapper[4885]: I1002 02:02:23.918887 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Oct 02 02:02:24 crc kubenswrapper[4885]: I1002 02:02:24.798255 4885 generic.go:334] "Generic (PLEG): container finished" podID="87196135-8d40-47cb-b918-98aea04c7c17" containerID="9d57810a37a1cca799c557440d8bd72c570370f072692434ad16f328610f95b0" exitCode=0 Oct 02 02:02:24 crc kubenswrapper[4885]: I1002 02:02:24.798325 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"87196135-8d40-47cb-b918-98aea04c7c17","Type":"ContainerDied","Data":"9d57810a37a1cca799c557440d8bd72c570370f072692434ad16f328610f95b0"} Oct 02 02:02:24 crc kubenswrapper[4885]: I1002 02:02:24.800779 4885 generic.go:334] "Generic (PLEG): container finished" podID="8427e298-965b-48f0-821f-bd078fa2e96f" containerID="f068f8c5a1d031159cc0d6e4c94d1b871a16e9160385c1912db8f0e8cb668289" exitCode=0 Oct 02 02:02:24 crc kubenswrapper[4885]: I1002 02:02:24.800886 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"8427e298-965b-48f0-821f-bd078fa2e96f","Type":"ContainerDied","Data":"f068f8c5a1d031159cc0d6e4c94d1b871a16e9160385c1912db8f0e8cb668289"} Oct 02 02:02:24 crc kubenswrapper[4885]: I1002 02:02:24.920168 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Oct 02 02:02:25 crc kubenswrapper[4885]: I1002 02:02:25.183427 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-556m4"] Oct 02 02:02:25 crc kubenswrapper[4885]: E1002 02:02:25.184295 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32a76623-c272-48c1-85e6-2170cbb017cc" containerName="init" Oct 02 02:02:25 crc kubenswrapper[4885]: I1002 02:02:25.184318 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="32a76623-c272-48c1-85e6-2170cbb017cc" containerName="init" Oct 02 02:02:25 crc kubenswrapper[4885]: E1002 02:02:25.184348 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32a76623-c272-48c1-85e6-2170cbb017cc" containerName="dnsmasq-dns" Oct 02 02:02:25 crc kubenswrapper[4885]: I1002 02:02:25.184357 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="32a76623-c272-48c1-85e6-2170cbb017cc" containerName="dnsmasq-dns" Oct 02 02:02:25 crc kubenswrapper[4885]: I1002 02:02:25.184562 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="32a76623-c272-48c1-85e6-2170cbb017cc" containerName="dnsmasq-dns" Oct 02 02:02:25 crc kubenswrapper[4885]: I1002 02:02:25.189541 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-556m4" Oct 02 02:02:25 crc kubenswrapper[4885]: I1002 02:02:25.195890 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Oct 02 02:02:25 crc kubenswrapper[4885]: I1002 02:02:25.199442 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-556m4"] Oct 02 02:02:25 crc kubenswrapper[4885]: I1002 02:02:25.255649 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-sphmn"] Oct 02 02:02:25 crc kubenswrapper[4885]: I1002 02:02:25.256785 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-sphmn" Oct 02 02:02:25 crc kubenswrapper[4885]: I1002 02:02:25.262150 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-sphmn"] Oct 02 02:02:25 crc kubenswrapper[4885]: I1002 02:02:25.290454 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Oct 02 02:02:25 crc kubenswrapper[4885]: I1002 02:02:25.297439 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/949636bd-3d8d-46f6-bfff-c201cea96ef6-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-556m4\" (UID: \"949636bd-3d8d-46f6-bfff-c201cea96ef6\") " pod="openstack/dnsmasq-dns-7fd796d7df-556m4" Oct 02 02:02:25 crc kubenswrapper[4885]: I1002 02:02:25.297505 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/949636bd-3d8d-46f6-bfff-c201cea96ef6-config\") pod \"dnsmasq-dns-7fd796d7df-556m4\" (UID: \"949636bd-3d8d-46f6-bfff-c201cea96ef6\") " pod="openstack/dnsmasq-dns-7fd796d7df-556m4" Oct 02 02:02:25 crc kubenswrapper[4885]: I1002 02:02:25.297552 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/949636bd-3d8d-46f6-bfff-c201cea96ef6-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-556m4\" (UID: \"949636bd-3d8d-46f6-bfff-c201cea96ef6\") " pod="openstack/dnsmasq-dns-7fd796d7df-556m4" Oct 02 02:02:25 crc kubenswrapper[4885]: I1002 02:02:25.297585 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dzr69\" (UniqueName: \"kubernetes.io/projected/949636bd-3d8d-46f6-bfff-c201cea96ef6-kube-api-access-dzr69\") pod \"dnsmasq-dns-7fd796d7df-556m4\" (UID: \"949636bd-3d8d-46f6-bfff-c201cea96ef6\") " pod="openstack/dnsmasq-dns-7fd796d7df-556m4" Oct 02 02:02:25 crc kubenswrapper[4885]: I1002 02:02:25.396830 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-556m4"] Oct 02 02:02:25 crc kubenswrapper[4885]: E1002 02:02:25.397562 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[config dns-svc kube-api-access-dzr69 ovsdbserver-nb], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/dnsmasq-dns-7fd796d7df-556m4" podUID="949636bd-3d8d-46f6-bfff-c201cea96ef6" Oct 02 02:02:25 crc kubenswrapper[4885]: I1002 02:02:25.398010 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/eb166198-03b5-458f-b86b-5706afa88d7b-ovn-rundir\") pod \"ovn-controller-metrics-sphmn\" (UID: \"eb166198-03b5-458f-b86b-5706afa88d7b\") " pod="openstack/ovn-controller-metrics-sphmn" Oct 02 02:02:25 crc kubenswrapper[4885]: I1002 02:02:25.398118 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/949636bd-3d8d-46f6-bfff-c201cea96ef6-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-556m4\" (UID: \"949636bd-3d8d-46f6-bfff-c201cea96ef6\") " pod="openstack/dnsmasq-dns-7fd796d7df-556m4" Oct 02 02:02:25 crc kubenswrapper[4885]: I1002 02:02:25.398196 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dzr69\" (UniqueName: \"kubernetes.io/projected/949636bd-3d8d-46f6-bfff-c201cea96ef6-kube-api-access-dzr69\") pod \"dnsmasq-dns-7fd796d7df-556m4\" (UID: \"949636bd-3d8d-46f6-bfff-c201cea96ef6\") " pod="openstack/dnsmasq-dns-7fd796d7df-556m4" Oct 02 02:02:25 crc kubenswrapper[4885]: I1002 02:02:25.398303 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fvqw2\" (UniqueName: \"kubernetes.io/projected/eb166198-03b5-458f-b86b-5706afa88d7b-kube-api-access-fvqw2\") pod \"ovn-controller-metrics-sphmn\" (UID: \"eb166198-03b5-458f-b86b-5706afa88d7b\") " pod="openstack/ovn-controller-metrics-sphmn" Oct 02 02:02:25 crc kubenswrapper[4885]: I1002 02:02:25.398395 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/eb166198-03b5-458f-b86b-5706afa88d7b-ovs-rundir\") pod \"ovn-controller-metrics-sphmn\" (UID: \"eb166198-03b5-458f-b86b-5706afa88d7b\") " pod="openstack/ovn-controller-metrics-sphmn" Oct 02 02:02:25 crc kubenswrapper[4885]: I1002 02:02:25.398484 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb166198-03b5-458f-b86b-5706afa88d7b-config\") pod \"ovn-controller-metrics-sphmn\" (UID: \"eb166198-03b5-458f-b86b-5706afa88d7b\") " pod="openstack/ovn-controller-metrics-sphmn" Oct 02 02:02:25 crc kubenswrapper[4885]: I1002 02:02:25.398576 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/eb166198-03b5-458f-b86b-5706afa88d7b-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-sphmn\" (UID: \"eb166198-03b5-458f-b86b-5706afa88d7b\") " pod="openstack/ovn-controller-metrics-sphmn" Oct 02 02:02:25 crc kubenswrapper[4885]: I1002 02:02:25.398689 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/949636bd-3d8d-46f6-bfff-c201cea96ef6-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-556m4\" (UID: \"949636bd-3d8d-46f6-bfff-c201cea96ef6\") " pod="openstack/dnsmasq-dns-7fd796d7df-556m4" Oct 02 02:02:25 crc kubenswrapper[4885]: I1002 02:02:25.398776 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb166198-03b5-458f-b86b-5706afa88d7b-combined-ca-bundle\") pod \"ovn-controller-metrics-sphmn\" (UID: \"eb166198-03b5-458f-b86b-5706afa88d7b\") " pod="openstack/ovn-controller-metrics-sphmn" Oct 02 02:02:25 crc kubenswrapper[4885]: I1002 02:02:25.398873 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/949636bd-3d8d-46f6-bfff-c201cea96ef6-config\") pod \"dnsmasq-dns-7fd796d7df-556m4\" (UID: \"949636bd-3d8d-46f6-bfff-c201cea96ef6\") " pod="openstack/dnsmasq-dns-7fd796d7df-556m4" Oct 02 02:02:25 crc kubenswrapper[4885]: I1002 02:02:25.398993 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/949636bd-3d8d-46f6-bfff-c201cea96ef6-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-556m4\" (UID: \"949636bd-3d8d-46f6-bfff-c201cea96ef6\") " pod="openstack/dnsmasq-dns-7fd796d7df-556m4" Oct 02 02:02:25 crc kubenswrapper[4885]: I1002 02:02:25.399413 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/949636bd-3d8d-46f6-bfff-c201cea96ef6-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-556m4\" (UID: \"949636bd-3d8d-46f6-bfff-c201cea96ef6\") " pod="openstack/dnsmasq-dns-7fd796d7df-556m4" Oct 02 02:02:25 crc kubenswrapper[4885]: I1002 02:02:25.400180 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/949636bd-3d8d-46f6-bfff-c201cea96ef6-config\") pod \"dnsmasq-dns-7fd796d7df-556m4\" (UID: \"949636bd-3d8d-46f6-bfff-c201cea96ef6\") " pod="openstack/dnsmasq-dns-7fd796d7df-556m4" Oct 02 02:02:25 crc kubenswrapper[4885]: I1002 02:02:25.414315 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-z5rln"] Oct 02 02:02:25 crc kubenswrapper[4885]: I1002 02:02:25.415509 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-z5rln" Oct 02 02:02:25 crc kubenswrapper[4885]: I1002 02:02:25.418960 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Oct 02 02:02:25 crc kubenswrapper[4885]: I1002 02:02:25.421433 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dzr69\" (UniqueName: \"kubernetes.io/projected/949636bd-3d8d-46f6-bfff-c201cea96ef6-kube-api-access-dzr69\") pod \"dnsmasq-dns-7fd796d7df-556m4\" (UID: \"949636bd-3d8d-46f6-bfff-c201cea96ef6\") " pod="openstack/dnsmasq-dns-7fd796d7df-556m4" Oct 02 02:02:25 crc kubenswrapper[4885]: I1002 02:02:25.433876 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-z5rln"] Oct 02 02:02:25 crc kubenswrapper[4885]: I1002 02:02:25.500788 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/eb166198-03b5-458f-b86b-5706afa88d7b-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-sphmn\" (UID: \"eb166198-03b5-458f-b86b-5706afa88d7b\") " pod="openstack/ovn-controller-metrics-sphmn" Oct 02 02:02:25 crc kubenswrapper[4885]: I1002 02:02:25.500873 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fd73a7ac-f025-4cf2-b09c-26e3eb7096c6-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-z5rln\" (UID: \"fd73a7ac-f025-4cf2-b09c-26e3eb7096c6\") " pod="openstack/dnsmasq-dns-86db49b7ff-z5rln" Oct 02 02:02:25 crc kubenswrapper[4885]: I1002 02:02:25.500904 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb166198-03b5-458f-b86b-5706afa88d7b-combined-ca-bundle\") pod \"ovn-controller-metrics-sphmn\" (UID: \"eb166198-03b5-458f-b86b-5706afa88d7b\") " pod="openstack/ovn-controller-metrics-sphmn" Oct 02 02:02:25 crc kubenswrapper[4885]: I1002 02:02:25.500934 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t8hrq\" (UniqueName: \"kubernetes.io/projected/fd73a7ac-f025-4cf2-b09c-26e3eb7096c6-kube-api-access-t8hrq\") pod \"dnsmasq-dns-86db49b7ff-z5rln\" (UID: \"fd73a7ac-f025-4cf2-b09c-26e3eb7096c6\") " pod="openstack/dnsmasq-dns-86db49b7ff-z5rln" Oct 02 02:02:25 crc kubenswrapper[4885]: I1002 02:02:25.500950 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fd73a7ac-f025-4cf2-b09c-26e3eb7096c6-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-z5rln\" (UID: \"fd73a7ac-f025-4cf2-b09c-26e3eb7096c6\") " pod="openstack/dnsmasq-dns-86db49b7ff-z5rln" Oct 02 02:02:25 crc kubenswrapper[4885]: I1002 02:02:25.500970 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd73a7ac-f025-4cf2-b09c-26e3eb7096c6-config\") pod \"dnsmasq-dns-86db49b7ff-z5rln\" (UID: \"fd73a7ac-f025-4cf2-b09c-26e3eb7096c6\") " pod="openstack/dnsmasq-dns-86db49b7ff-z5rln" Oct 02 02:02:25 crc kubenswrapper[4885]: I1002 02:02:25.500997 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/eb166198-03b5-458f-b86b-5706afa88d7b-ovn-rundir\") pod \"ovn-controller-metrics-sphmn\" (UID: \"eb166198-03b5-458f-b86b-5706afa88d7b\") " pod="openstack/ovn-controller-metrics-sphmn" Oct 02 02:02:25 crc kubenswrapper[4885]: I1002 02:02:25.501037 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fvqw2\" (UniqueName: \"kubernetes.io/projected/eb166198-03b5-458f-b86b-5706afa88d7b-kube-api-access-fvqw2\") pod \"ovn-controller-metrics-sphmn\" (UID: \"eb166198-03b5-458f-b86b-5706afa88d7b\") " pod="openstack/ovn-controller-metrics-sphmn" Oct 02 02:02:25 crc kubenswrapper[4885]: I1002 02:02:25.501057 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/eb166198-03b5-458f-b86b-5706afa88d7b-ovs-rundir\") pod \"ovn-controller-metrics-sphmn\" (UID: \"eb166198-03b5-458f-b86b-5706afa88d7b\") " pod="openstack/ovn-controller-metrics-sphmn" Oct 02 02:02:25 crc kubenswrapper[4885]: I1002 02:02:25.501077 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fd73a7ac-f025-4cf2-b09c-26e3eb7096c6-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-z5rln\" (UID: \"fd73a7ac-f025-4cf2-b09c-26e3eb7096c6\") " pod="openstack/dnsmasq-dns-86db49b7ff-z5rln" Oct 02 02:02:25 crc kubenswrapper[4885]: I1002 02:02:25.501099 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb166198-03b5-458f-b86b-5706afa88d7b-config\") pod \"ovn-controller-metrics-sphmn\" (UID: \"eb166198-03b5-458f-b86b-5706afa88d7b\") " pod="openstack/ovn-controller-metrics-sphmn" Oct 02 02:02:25 crc kubenswrapper[4885]: I1002 02:02:25.501781 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb166198-03b5-458f-b86b-5706afa88d7b-config\") pod \"ovn-controller-metrics-sphmn\" (UID: \"eb166198-03b5-458f-b86b-5706afa88d7b\") " pod="openstack/ovn-controller-metrics-sphmn" Oct 02 02:02:25 crc kubenswrapper[4885]: I1002 02:02:25.502006 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/eb166198-03b5-458f-b86b-5706afa88d7b-ovn-rundir\") pod \"ovn-controller-metrics-sphmn\" (UID: \"eb166198-03b5-458f-b86b-5706afa88d7b\") " pod="openstack/ovn-controller-metrics-sphmn" Oct 02 02:02:25 crc kubenswrapper[4885]: I1002 02:02:25.502291 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/eb166198-03b5-458f-b86b-5706afa88d7b-ovs-rundir\") pod \"ovn-controller-metrics-sphmn\" (UID: \"eb166198-03b5-458f-b86b-5706afa88d7b\") " pod="openstack/ovn-controller-metrics-sphmn" Oct 02 02:02:25 crc kubenswrapper[4885]: I1002 02:02:25.503765 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/eb166198-03b5-458f-b86b-5706afa88d7b-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-sphmn\" (UID: \"eb166198-03b5-458f-b86b-5706afa88d7b\") " pod="openstack/ovn-controller-metrics-sphmn" Oct 02 02:02:25 crc kubenswrapper[4885]: I1002 02:02:25.504756 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb166198-03b5-458f-b86b-5706afa88d7b-combined-ca-bundle\") pod \"ovn-controller-metrics-sphmn\" (UID: \"eb166198-03b5-458f-b86b-5706afa88d7b\") " pod="openstack/ovn-controller-metrics-sphmn" Oct 02 02:02:25 crc kubenswrapper[4885]: I1002 02:02:25.514910 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fvqw2\" (UniqueName: \"kubernetes.io/projected/eb166198-03b5-458f-b86b-5706afa88d7b-kube-api-access-fvqw2\") pod \"ovn-controller-metrics-sphmn\" (UID: \"eb166198-03b5-458f-b86b-5706afa88d7b\") " pod="openstack/ovn-controller-metrics-sphmn" Oct 02 02:02:25 crc kubenswrapper[4885]: I1002 02:02:25.601330 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-sphmn" Oct 02 02:02:25 crc kubenswrapper[4885]: I1002 02:02:25.602517 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fd73a7ac-f025-4cf2-b09c-26e3eb7096c6-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-z5rln\" (UID: \"fd73a7ac-f025-4cf2-b09c-26e3eb7096c6\") " pod="openstack/dnsmasq-dns-86db49b7ff-z5rln" Oct 02 02:02:25 crc kubenswrapper[4885]: I1002 02:02:25.602566 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t8hrq\" (UniqueName: \"kubernetes.io/projected/fd73a7ac-f025-4cf2-b09c-26e3eb7096c6-kube-api-access-t8hrq\") pod \"dnsmasq-dns-86db49b7ff-z5rln\" (UID: \"fd73a7ac-f025-4cf2-b09c-26e3eb7096c6\") " pod="openstack/dnsmasq-dns-86db49b7ff-z5rln" Oct 02 02:02:25 crc kubenswrapper[4885]: I1002 02:02:25.602584 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fd73a7ac-f025-4cf2-b09c-26e3eb7096c6-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-z5rln\" (UID: \"fd73a7ac-f025-4cf2-b09c-26e3eb7096c6\") " pod="openstack/dnsmasq-dns-86db49b7ff-z5rln" Oct 02 02:02:25 crc kubenswrapper[4885]: I1002 02:02:25.602600 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd73a7ac-f025-4cf2-b09c-26e3eb7096c6-config\") pod \"dnsmasq-dns-86db49b7ff-z5rln\" (UID: \"fd73a7ac-f025-4cf2-b09c-26e3eb7096c6\") " pod="openstack/dnsmasq-dns-86db49b7ff-z5rln" Oct 02 02:02:25 crc kubenswrapper[4885]: I1002 02:02:25.602656 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fd73a7ac-f025-4cf2-b09c-26e3eb7096c6-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-z5rln\" (UID: \"fd73a7ac-f025-4cf2-b09c-26e3eb7096c6\") " pod="openstack/dnsmasq-dns-86db49b7ff-z5rln" Oct 02 02:02:25 crc kubenswrapper[4885]: I1002 02:02:25.603386 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fd73a7ac-f025-4cf2-b09c-26e3eb7096c6-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-z5rln\" (UID: \"fd73a7ac-f025-4cf2-b09c-26e3eb7096c6\") " pod="openstack/dnsmasq-dns-86db49b7ff-z5rln" Oct 02 02:02:25 crc kubenswrapper[4885]: I1002 02:02:25.603445 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd73a7ac-f025-4cf2-b09c-26e3eb7096c6-config\") pod \"dnsmasq-dns-86db49b7ff-z5rln\" (UID: \"fd73a7ac-f025-4cf2-b09c-26e3eb7096c6\") " pod="openstack/dnsmasq-dns-86db49b7ff-z5rln" Oct 02 02:02:25 crc kubenswrapper[4885]: I1002 02:02:25.603791 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fd73a7ac-f025-4cf2-b09c-26e3eb7096c6-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-z5rln\" (UID: \"fd73a7ac-f025-4cf2-b09c-26e3eb7096c6\") " pod="openstack/dnsmasq-dns-86db49b7ff-z5rln" Oct 02 02:02:25 crc kubenswrapper[4885]: I1002 02:02:25.603876 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fd73a7ac-f025-4cf2-b09c-26e3eb7096c6-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-z5rln\" (UID: \"fd73a7ac-f025-4cf2-b09c-26e3eb7096c6\") " pod="openstack/dnsmasq-dns-86db49b7ff-z5rln" Oct 02 02:02:25 crc kubenswrapper[4885]: I1002 02:02:25.619077 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t8hrq\" (UniqueName: \"kubernetes.io/projected/fd73a7ac-f025-4cf2-b09c-26e3eb7096c6-kube-api-access-t8hrq\") pod \"dnsmasq-dns-86db49b7ff-z5rln\" (UID: \"fd73a7ac-f025-4cf2-b09c-26e3eb7096c6\") " pod="openstack/dnsmasq-dns-86db49b7ff-z5rln" Oct 02 02:02:25 crc kubenswrapper[4885]: I1002 02:02:25.768116 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-z5rln" Oct 02 02:02:25 crc kubenswrapper[4885]: I1002 02:02:25.855584 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"8427e298-965b-48f0-821f-bd078fa2e96f","Type":"ContainerStarted","Data":"2e6a4c155cc94b76eb037a97b0e2f889715dcd77b1465dc354aace3b4c727a76"} Oct 02 02:02:25 crc kubenswrapper[4885]: I1002 02:02:25.866108 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-556m4" Oct 02 02:02:25 crc kubenswrapper[4885]: I1002 02:02:25.866768 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"87196135-8d40-47cb-b918-98aea04c7c17","Type":"ContainerStarted","Data":"52fcc6f63f08bc45110f7b4da6dfa608aa366daf9cf749aa05d565783007ad4f"} Oct 02 02:02:25 crc kubenswrapper[4885]: I1002 02:02:25.881761 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-556m4" Oct 02 02:02:25 crc kubenswrapper[4885]: I1002 02:02:25.927022 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=21.85946324 podStartE2EDuration="40.927008824s" podCreationTimestamp="2025-10-02 02:01:45 +0000 UTC" firstStartedPulling="2025-10-02 02:01:57.323027366 +0000 UTC m=+906.134774765" lastFinishedPulling="2025-10-02 02:02:16.39057294 +0000 UTC m=+925.202320349" observedRunningTime="2025-10-02 02:02:25.884560782 +0000 UTC m=+934.696308181" watchObservedRunningTime="2025-10-02 02:02:25.927008824 +0000 UTC m=+934.738756223" Oct 02 02:02:26 crc kubenswrapper[4885]: I1002 02:02:26.020728 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dzr69\" (UniqueName: \"kubernetes.io/projected/949636bd-3d8d-46f6-bfff-c201cea96ef6-kube-api-access-dzr69\") pod \"949636bd-3d8d-46f6-bfff-c201cea96ef6\" (UID: \"949636bd-3d8d-46f6-bfff-c201cea96ef6\") " Oct 02 02:02:26 crc kubenswrapper[4885]: I1002 02:02:26.020777 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/949636bd-3d8d-46f6-bfff-c201cea96ef6-dns-svc\") pod \"949636bd-3d8d-46f6-bfff-c201cea96ef6\" (UID: \"949636bd-3d8d-46f6-bfff-c201cea96ef6\") " Oct 02 02:02:26 crc kubenswrapper[4885]: I1002 02:02:26.020858 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/949636bd-3d8d-46f6-bfff-c201cea96ef6-ovsdbserver-nb\") pod \"949636bd-3d8d-46f6-bfff-c201cea96ef6\" (UID: \"949636bd-3d8d-46f6-bfff-c201cea96ef6\") " Oct 02 02:02:26 crc kubenswrapper[4885]: I1002 02:02:26.020982 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/949636bd-3d8d-46f6-bfff-c201cea96ef6-config\") pod \"949636bd-3d8d-46f6-bfff-c201cea96ef6\" (UID: \"949636bd-3d8d-46f6-bfff-c201cea96ef6\") " Oct 02 02:02:26 crc kubenswrapper[4885]: I1002 02:02:26.021773 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/949636bd-3d8d-46f6-bfff-c201cea96ef6-config" (OuterVolumeSpecName: "config") pod "949636bd-3d8d-46f6-bfff-c201cea96ef6" (UID: "949636bd-3d8d-46f6-bfff-c201cea96ef6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:02:26 crc kubenswrapper[4885]: I1002 02:02:26.024505 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/949636bd-3d8d-46f6-bfff-c201cea96ef6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "949636bd-3d8d-46f6-bfff-c201cea96ef6" (UID: "949636bd-3d8d-46f6-bfff-c201cea96ef6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:02:26 crc kubenswrapper[4885]: I1002 02:02:26.025696 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/949636bd-3d8d-46f6-bfff-c201cea96ef6-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "949636bd-3d8d-46f6-bfff-c201cea96ef6" (UID: "949636bd-3d8d-46f6-bfff-c201cea96ef6"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:02:26 crc kubenswrapper[4885]: I1002 02:02:26.037959 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/949636bd-3d8d-46f6-bfff-c201cea96ef6-kube-api-access-dzr69" (OuterVolumeSpecName: "kube-api-access-dzr69") pod "949636bd-3d8d-46f6-bfff-c201cea96ef6" (UID: "949636bd-3d8d-46f6-bfff-c201cea96ef6"). InnerVolumeSpecName "kube-api-access-dzr69". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:02:26 crc kubenswrapper[4885]: I1002 02:02:26.042800 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=21.236589187 podStartE2EDuration="41.042780206s" podCreationTimestamp="2025-10-02 02:01:45 +0000 UTC" firstStartedPulling="2025-10-02 02:01:56.90971624 +0000 UTC m=+905.721463639" lastFinishedPulling="2025-10-02 02:02:16.715907219 +0000 UTC m=+925.527654658" observedRunningTime="2025-10-02 02:02:25.925986252 +0000 UTC m=+934.737733651" watchObservedRunningTime="2025-10-02 02:02:26.042780206 +0000 UTC m=+934.854527605" Oct 02 02:02:26 crc kubenswrapper[4885]: I1002 02:02:26.057095 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-sphmn"] Oct 02 02:02:26 crc kubenswrapper[4885]: W1002 02:02:26.064736 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeb166198_03b5_458f_b86b_5706afa88d7b.slice/crio-d44206102368a77b84121a5bee42210a7993345c0f316f3d8827320c95ebfcae WatchSource:0}: Error finding container d44206102368a77b84121a5bee42210a7993345c0f316f3d8827320c95ebfcae: Status 404 returned error can't find the container with id d44206102368a77b84121a5bee42210a7993345c0f316f3d8827320c95ebfcae Oct 02 02:02:26 crc kubenswrapper[4885]: I1002 02:02:26.117502 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-z5rln"] Oct 02 02:02:26 crc kubenswrapper[4885]: I1002 02:02:26.122715 4885 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/949636bd-3d8d-46f6-bfff-c201cea96ef6-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 02:02:26 crc kubenswrapper[4885]: I1002 02:02:26.123503 4885 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/949636bd-3d8d-46f6-bfff-c201cea96ef6-config\") on node \"crc\" DevicePath \"\"" Oct 02 02:02:26 crc kubenswrapper[4885]: I1002 02:02:26.123554 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dzr69\" (UniqueName: \"kubernetes.io/projected/949636bd-3d8d-46f6-bfff-c201cea96ef6-kube-api-access-dzr69\") on node \"crc\" DevicePath \"\"" Oct 02 02:02:26 crc kubenswrapper[4885]: I1002 02:02:26.123568 4885 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/949636bd-3d8d-46f6-bfff-c201cea96ef6-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 02:02:26 crc kubenswrapper[4885]: I1002 02:02:26.876080 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-sphmn" event={"ID":"eb166198-03b5-458f-b86b-5706afa88d7b","Type":"ContainerStarted","Data":"381d207490d1b72d727e5545b2c811aa3516fa10a9e6d8019effc08d07313e35"} Oct 02 02:02:26 crc kubenswrapper[4885]: I1002 02:02:26.876492 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-sphmn" event={"ID":"eb166198-03b5-458f-b86b-5706afa88d7b","Type":"ContainerStarted","Data":"d44206102368a77b84121a5bee42210a7993345c0f316f3d8827320c95ebfcae"} Oct 02 02:02:26 crc kubenswrapper[4885]: I1002 02:02:26.878113 4885 generic.go:334] "Generic (PLEG): container finished" podID="fd73a7ac-f025-4cf2-b09c-26e3eb7096c6" containerID="c7f5dd99660e2d1123057f565c6bea67fc73d1d20b56c586e37566f67b0e79e1" exitCode=0 Oct 02 02:02:26 crc kubenswrapper[4885]: I1002 02:02:26.878204 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-556m4" Oct 02 02:02:26 crc kubenswrapper[4885]: I1002 02:02:26.878559 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-z5rln" event={"ID":"fd73a7ac-f025-4cf2-b09c-26e3eb7096c6","Type":"ContainerDied","Data":"c7f5dd99660e2d1123057f565c6bea67fc73d1d20b56c586e37566f67b0e79e1"} Oct 02 02:02:26 crc kubenswrapper[4885]: I1002 02:02:26.878621 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-z5rln" event={"ID":"fd73a7ac-f025-4cf2-b09c-26e3eb7096c6","Type":"ContainerStarted","Data":"72d8e15931c507955b4affeb845000fbd62a3e0d694f29cac6bd8e8e31ec414c"} Oct 02 02:02:26 crc kubenswrapper[4885]: I1002 02:02:26.919939 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-sphmn" podStartSLOduration=1.9199205849999998 podStartE2EDuration="1.919920585s" podCreationTimestamp="2025-10-02 02:02:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:02:26.905195707 +0000 UTC m=+935.716943146" watchObservedRunningTime="2025-10-02 02:02:26.919920585 +0000 UTC m=+935.731667994" Oct 02 02:02:26 crc kubenswrapper[4885]: I1002 02:02:26.977982 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-556m4"] Oct 02 02:02:26 crc kubenswrapper[4885]: I1002 02:02:26.988205 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-556m4"] Oct 02 02:02:27 crc kubenswrapper[4885]: I1002 02:02:27.212837 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Oct 02 02:02:27 crc kubenswrapper[4885]: I1002 02:02:27.212889 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Oct 02 02:02:27 crc kubenswrapper[4885]: I1002 02:02:27.326501 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Oct 02 02:02:27 crc kubenswrapper[4885]: I1002 02:02:27.326824 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Oct 02 02:02:27 crc kubenswrapper[4885]: I1002 02:02:27.890059 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-z5rln" event={"ID":"fd73a7ac-f025-4cf2-b09c-26e3eb7096c6","Type":"ContainerStarted","Data":"2c77e4bde7e6a6a3d5019ef6001f7556d9a38620f43ae8597f351f41e996df8c"} Oct 02 02:02:27 crc kubenswrapper[4885]: I1002 02:02:27.911044 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-86db49b7ff-z5rln" podStartSLOduration=2.911025281 podStartE2EDuration="2.911025281s" podCreationTimestamp="2025-10-02 02:02:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:02:27.907897966 +0000 UTC m=+936.719645395" watchObservedRunningTime="2025-10-02 02:02:27.911025281 +0000 UTC m=+936.722772710" Oct 02 02:02:28 crc kubenswrapper[4885]: I1002 02:02:28.063391 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="949636bd-3d8d-46f6-bfff-c201cea96ef6" path="/var/lib/kubelet/pods/949636bd-3d8d-46f6-bfff-c201cea96ef6/volumes" Oct 02 02:02:28 crc kubenswrapper[4885]: I1002 02:02:28.900407 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-86db49b7ff-z5rln" Oct 02 02:02:29 crc kubenswrapper[4885]: I1002 02:02:29.586382 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-z5rln"] Oct 02 02:02:29 crc kubenswrapper[4885]: I1002 02:02:29.620380 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-698758b865-qkx2j"] Oct 02 02:02:29 crc kubenswrapper[4885]: I1002 02:02:29.621628 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-qkx2j" Oct 02 02:02:29 crc kubenswrapper[4885]: I1002 02:02:29.637199 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-qkx2j"] Oct 02 02:02:29 crc kubenswrapper[4885]: I1002 02:02:29.700527 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/52b80336-4c74-4236-9600-7ea4606e5d6f-dns-svc\") pod \"dnsmasq-dns-698758b865-qkx2j\" (UID: \"52b80336-4c74-4236-9600-7ea4606e5d6f\") " pod="openstack/dnsmasq-dns-698758b865-qkx2j" Oct 02 02:02:29 crc kubenswrapper[4885]: I1002 02:02:29.700577 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/52b80336-4c74-4236-9600-7ea4606e5d6f-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-qkx2j\" (UID: \"52b80336-4c74-4236-9600-7ea4606e5d6f\") " pod="openstack/dnsmasq-dns-698758b865-qkx2j" Oct 02 02:02:29 crc kubenswrapper[4885]: I1002 02:02:29.700648 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/52b80336-4c74-4236-9600-7ea4606e5d6f-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-qkx2j\" (UID: \"52b80336-4c74-4236-9600-7ea4606e5d6f\") " pod="openstack/dnsmasq-dns-698758b865-qkx2j" Oct 02 02:02:29 crc kubenswrapper[4885]: I1002 02:02:29.700694 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/52b80336-4c74-4236-9600-7ea4606e5d6f-config\") pod \"dnsmasq-dns-698758b865-qkx2j\" (UID: \"52b80336-4c74-4236-9600-7ea4606e5d6f\") " pod="openstack/dnsmasq-dns-698758b865-qkx2j" Oct 02 02:02:29 crc kubenswrapper[4885]: I1002 02:02:29.700737 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-br595\" (UniqueName: \"kubernetes.io/projected/52b80336-4c74-4236-9600-7ea4606e5d6f-kube-api-access-br595\") pod \"dnsmasq-dns-698758b865-qkx2j\" (UID: \"52b80336-4c74-4236-9600-7ea4606e5d6f\") " pod="openstack/dnsmasq-dns-698758b865-qkx2j" Oct 02 02:02:29 crc kubenswrapper[4885]: I1002 02:02:29.802575 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/52b80336-4c74-4236-9600-7ea4606e5d6f-dns-svc\") pod \"dnsmasq-dns-698758b865-qkx2j\" (UID: \"52b80336-4c74-4236-9600-7ea4606e5d6f\") " pod="openstack/dnsmasq-dns-698758b865-qkx2j" Oct 02 02:02:29 crc kubenswrapper[4885]: I1002 02:02:29.802669 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/52b80336-4c74-4236-9600-7ea4606e5d6f-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-qkx2j\" (UID: \"52b80336-4c74-4236-9600-7ea4606e5d6f\") " pod="openstack/dnsmasq-dns-698758b865-qkx2j" Oct 02 02:02:29 crc kubenswrapper[4885]: I1002 02:02:29.802747 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/52b80336-4c74-4236-9600-7ea4606e5d6f-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-qkx2j\" (UID: \"52b80336-4c74-4236-9600-7ea4606e5d6f\") " pod="openstack/dnsmasq-dns-698758b865-qkx2j" Oct 02 02:02:29 crc kubenswrapper[4885]: I1002 02:02:29.802814 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/52b80336-4c74-4236-9600-7ea4606e5d6f-config\") pod \"dnsmasq-dns-698758b865-qkx2j\" (UID: \"52b80336-4c74-4236-9600-7ea4606e5d6f\") " pod="openstack/dnsmasq-dns-698758b865-qkx2j" Oct 02 02:02:29 crc kubenswrapper[4885]: I1002 02:02:29.802865 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-br595\" (UniqueName: \"kubernetes.io/projected/52b80336-4c74-4236-9600-7ea4606e5d6f-kube-api-access-br595\") pod \"dnsmasq-dns-698758b865-qkx2j\" (UID: \"52b80336-4c74-4236-9600-7ea4606e5d6f\") " pod="openstack/dnsmasq-dns-698758b865-qkx2j" Oct 02 02:02:29 crc kubenswrapper[4885]: I1002 02:02:29.803561 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/52b80336-4c74-4236-9600-7ea4606e5d6f-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-qkx2j\" (UID: \"52b80336-4c74-4236-9600-7ea4606e5d6f\") " pod="openstack/dnsmasq-dns-698758b865-qkx2j" Oct 02 02:02:29 crc kubenswrapper[4885]: I1002 02:02:29.803656 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/52b80336-4c74-4236-9600-7ea4606e5d6f-dns-svc\") pod \"dnsmasq-dns-698758b865-qkx2j\" (UID: \"52b80336-4c74-4236-9600-7ea4606e5d6f\") " pod="openstack/dnsmasq-dns-698758b865-qkx2j" Oct 02 02:02:29 crc kubenswrapper[4885]: I1002 02:02:29.803663 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/52b80336-4c74-4236-9600-7ea4606e5d6f-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-qkx2j\" (UID: \"52b80336-4c74-4236-9600-7ea4606e5d6f\") " pod="openstack/dnsmasq-dns-698758b865-qkx2j" Oct 02 02:02:29 crc kubenswrapper[4885]: I1002 02:02:29.803947 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/52b80336-4c74-4236-9600-7ea4606e5d6f-config\") pod \"dnsmasq-dns-698758b865-qkx2j\" (UID: \"52b80336-4c74-4236-9600-7ea4606e5d6f\") " pod="openstack/dnsmasq-dns-698758b865-qkx2j" Oct 02 02:02:29 crc kubenswrapper[4885]: I1002 02:02:29.821643 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-br595\" (UniqueName: \"kubernetes.io/projected/52b80336-4c74-4236-9600-7ea4606e5d6f-kube-api-access-br595\") pod \"dnsmasq-dns-698758b865-qkx2j\" (UID: \"52b80336-4c74-4236-9600-7ea4606e5d6f\") " pod="openstack/dnsmasq-dns-698758b865-qkx2j" Oct 02 02:02:29 crc kubenswrapper[4885]: I1002 02:02:29.948046 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-qkx2j" Oct 02 02:02:30 crc kubenswrapper[4885]: I1002 02:02:30.474441 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-qkx2j"] Oct 02 02:02:30 crc kubenswrapper[4885]: W1002 02:02:30.481764 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod52b80336_4c74_4236_9600_7ea4606e5d6f.slice/crio-e4d5fb452419cdbec9b0adfdcfef9e875939e09cde369b3ae1b2665922ca85d6 WatchSource:0}: Error finding container e4d5fb452419cdbec9b0adfdcfef9e875939e09cde369b3ae1b2665922ca85d6: Status 404 returned error can't find the container with id e4d5fb452419cdbec9b0adfdcfef9e875939e09cde369b3ae1b2665922ca85d6 Oct 02 02:02:30 crc kubenswrapper[4885]: I1002 02:02:30.685075 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Oct 02 02:02:30 crc kubenswrapper[4885]: I1002 02:02:30.691908 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 02 02:02:30 crc kubenswrapper[4885]: I1002 02:02:30.693878 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-hs7t6" Oct 02 02:02:30 crc kubenswrapper[4885]: I1002 02:02:30.694524 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Oct 02 02:02:30 crc kubenswrapper[4885]: I1002 02:02:30.694655 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Oct 02 02:02:30 crc kubenswrapper[4885]: I1002 02:02:30.695153 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Oct 02 02:02:30 crc kubenswrapper[4885]: I1002 02:02:30.705118 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Oct 02 02:02:30 crc kubenswrapper[4885]: I1002 02:02:30.816843 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j989w\" (UniqueName: \"kubernetes.io/projected/0c8d183a-f48b-4a69-8685-90c834e678ce-kube-api-access-j989w\") pod \"swift-storage-0\" (UID: \"0c8d183a-f48b-4a69-8685-90c834e678ce\") " pod="openstack/swift-storage-0" Oct 02 02:02:30 crc kubenswrapper[4885]: I1002 02:02:30.816925 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"swift-storage-0\" (UID: \"0c8d183a-f48b-4a69-8685-90c834e678ce\") " pod="openstack/swift-storage-0" Oct 02 02:02:30 crc kubenswrapper[4885]: I1002 02:02:30.817006 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/0c8d183a-f48b-4a69-8685-90c834e678ce-cache\") pod \"swift-storage-0\" (UID: \"0c8d183a-f48b-4a69-8685-90c834e678ce\") " pod="openstack/swift-storage-0" Oct 02 02:02:30 crc kubenswrapper[4885]: I1002 02:02:30.817096 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/0c8d183a-f48b-4a69-8685-90c834e678ce-lock\") pod \"swift-storage-0\" (UID: \"0c8d183a-f48b-4a69-8685-90c834e678ce\") " pod="openstack/swift-storage-0" Oct 02 02:02:30 crc kubenswrapper[4885]: I1002 02:02:30.817165 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/0c8d183a-f48b-4a69-8685-90c834e678ce-etc-swift\") pod \"swift-storage-0\" (UID: \"0c8d183a-f48b-4a69-8685-90c834e678ce\") " pod="openstack/swift-storage-0" Oct 02 02:02:30 crc kubenswrapper[4885]: I1002 02:02:30.918145 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/0c8d183a-f48b-4a69-8685-90c834e678ce-lock\") pod \"swift-storage-0\" (UID: \"0c8d183a-f48b-4a69-8685-90c834e678ce\") " pod="openstack/swift-storage-0" Oct 02 02:02:30 crc kubenswrapper[4885]: I1002 02:02:30.918199 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/0c8d183a-f48b-4a69-8685-90c834e678ce-etc-swift\") pod \"swift-storage-0\" (UID: \"0c8d183a-f48b-4a69-8685-90c834e678ce\") " pod="openstack/swift-storage-0" Oct 02 02:02:30 crc kubenswrapper[4885]: I1002 02:02:30.918244 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j989w\" (UniqueName: \"kubernetes.io/projected/0c8d183a-f48b-4a69-8685-90c834e678ce-kube-api-access-j989w\") pod \"swift-storage-0\" (UID: \"0c8d183a-f48b-4a69-8685-90c834e678ce\") " pod="openstack/swift-storage-0" Oct 02 02:02:30 crc kubenswrapper[4885]: I1002 02:02:30.918284 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"swift-storage-0\" (UID: \"0c8d183a-f48b-4a69-8685-90c834e678ce\") " pod="openstack/swift-storage-0" Oct 02 02:02:30 crc kubenswrapper[4885]: I1002 02:02:30.918321 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/0c8d183a-f48b-4a69-8685-90c834e678ce-cache\") pod \"swift-storage-0\" (UID: \"0c8d183a-f48b-4a69-8685-90c834e678ce\") " pod="openstack/swift-storage-0" Oct 02 02:02:30 crc kubenswrapper[4885]: I1002 02:02:30.918768 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/0c8d183a-f48b-4a69-8685-90c834e678ce-cache\") pod \"swift-storage-0\" (UID: \"0c8d183a-f48b-4a69-8685-90c834e678ce\") " pod="openstack/swift-storage-0" Oct 02 02:02:30 crc kubenswrapper[4885]: I1002 02:02:30.920563 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/0c8d183a-f48b-4a69-8685-90c834e678ce-lock\") pod \"swift-storage-0\" (UID: \"0c8d183a-f48b-4a69-8685-90c834e678ce\") " pod="openstack/swift-storage-0" Oct 02 02:02:30 crc kubenswrapper[4885]: E1002 02:02:30.920689 4885 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 02 02:02:30 crc kubenswrapper[4885]: E1002 02:02:30.920708 4885 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 02 02:02:30 crc kubenswrapper[4885]: E1002 02:02:30.920748 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/0c8d183a-f48b-4a69-8685-90c834e678ce-etc-swift podName:0c8d183a-f48b-4a69-8685-90c834e678ce nodeName:}" failed. No retries permitted until 2025-10-02 02:02:31.420732017 +0000 UTC m=+940.232479406 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/0c8d183a-f48b-4a69-8685-90c834e678ce-etc-swift") pod "swift-storage-0" (UID: "0c8d183a-f48b-4a69-8685-90c834e678ce") : configmap "swift-ring-files" not found Oct 02 02:02:30 crc kubenswrapper[4885]: I1002 02:02:30.921308 4885 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"swift-storage-0\" (UID: \"0c8d183a-f48b-4a69-8685-90c834e678ce\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/swift-storage-0" Oct 02 02:02:30 crc kubenswrapper[4885]: I1002 02:02:30.924169 4885 generic.go:334] "Generic (PLEG): container finished" podID="52b80336-4c74-4236-9600-7ea4606e5d6f" containerID="6d45bac9a24e0df35ab7b6a7861c1ef92d45e7ae755cd19856e778a659e47200" exitCode=0 Oct 02 02:02:30 crc kubenswrapper[4885]: I1002 02:02:30.924491 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-86db49b7ff-z5rln" podUID="fd73a7ac-f025-4cf2-b09c-26e3eb7096c6" containerName="dnsmasq-dns" containerID="cri-o://2c77e4bde7e6a6a3d5019ef6001f7556d9a38620f43ae8597f351f41e996df8c" gracePeriod=10 Oct 02 02:02:30 crc kubenswrapper[4885]: I1002 02:02:30.925464 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-qkx2j" event={"ID":"52b80336-4c74-4236-9600-7ea4606e5d6f","Type":"ContainerDied","Data":"6d45bac9a24e0df35ab7b6a7861c1ef92d45e7ae755cd19856e778a659e47200"} Oct 02 02:02:30 crc kubenswrapper[4885]: I1002 02:02:30.925518 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-qkx2j" event={"ID":"52b80336-4c74-4236-9600-7ea4606e5d6f","Type":"ContainerStarted","Data":"e4d5fb452419cdbec9b0adfdcfef9e875939e09cde369b3ae1b2665922ca85d6"} Oct 02 02:02:30 crc kubenswrapper[4885]: I1002 02:02:30.939907 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j989w\" (UniqueName: \"kubernetes.io/projected/0c8d183a-f48b-4a69-8685-90c834e678ce-kube-api-access-j989w\") pod \"swift-storage-0\" (UID: \"0c8d183a-f48b-4a69-8685-90c834e678ce\") " pod="openstack/swift-storage-0" Oct 02 02:02:30 crc kubenswrapper[4885]: I1002 02:02:30.949494 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"swift-storage-0\" (UID: \"0c8d183a-f48b-4a69-8685-90c834e678ce\") " pod="openstack/swift-storage-0" Oct 02 02:02:31 crc kubenswrapper[4885]: I1002 02:02:31.237203 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-9wd49"] Oct 02 02:02:31 crc kubenswrapper[4885]: I1002 02:02:31.238694 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-9wd49" Oct 02 02:02:31 crc kubenswrapper[4885]: I1002 02:02:31.241399 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Oct 02 02:02:31 crc kubenswrapper[4885]: I1002 02:02:31.241535 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Oct 02 02:02:31 crc kubenswrapper[4885]: I1002 02:02:31.241652 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Oct 02 02:02:31 crc kubenswrapper[4885]: I1002 02:02:31.256903 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-9wd49"] Oct 02 02:02:31 crc kubenswrapper[4885]: I1002 02:02:31.330322 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-z5rln" Oct 02 02:02:31 crc kubenswrapper[4885]: I1002 02:02:31.429166 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/2d471d2b-912c-4f49-a7ec-c07b85c02bd9-ring-data-devices\") pod \"swift-ring-rebalance-9wd49\" (UID: \"2d471d2b-912c-4f49-a7ec-c07b85c02bd9\") " pod="openstack/swift-ring-rebalance-9wd49" Oct 02 02:02:31 crc kubenswrapper[4885]: I1002 02:02:31.429224 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2d471d2b-912c-4f49-a7ec-c07b85c02bd9-scripts\") pod \"swift-ring-rebalance-9wd49\" (UID: \"2d471d2b-912c-4f49-a7ec-c07b85c02bd9\") " pod="openstack/swift-ring-rebalance-9wd49" Oct 02 02:02:31 crc kubenswrapper[4885]: I1002 02:02:31.429254 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/2d471d2b-912c-4f49-a7ec-c07b85c02bd9-dispersionconf\") pod \"swift-ring-rebalance-9wd49\" (UID: \"2d471d2b-912c-4f49-a7ec-c07b85c02bd9\") " pod="openstack/swift-ring-rebalance-9wd49" Oct 02 02:02:31 crc kubenswrapper[4885]: I1002 02:02:31.429331 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/2d471d2b-912c-4f49-a7ec-c07b85c02bd9-etc-swift\") pod \"swift-ring-rebalance-9wd49\" (UID: \"2d471d2b-912c-4f49-a7ec-c07b85c02bd9\") " pod="openstack/swift-ring-rebalance-9wd49" Oct 02 02:02:31 crc kubenswrapper[4885]: I1002 02:02:31.429353 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qc2q7\" (UniqueName: \"kubernetes.io/projected/2d471d2b-912c-4f49-a7ec-c07b85c02bd9-kube-api-access-qc2q7\") pod \"swift-ring-rebalance-9wd49\" (UID: \"2d471d2b-912c-4f49-a7ec-c07b85c02bd9\") " pod="openstack/swift-ring-rebalance-9wd49" Oct 02 02:02:31 crc kubenswrapper[4885]: I1002 02:02:31.430191 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/0c8d183a-f48b-4a69-8685-90c834e678ce-etc-swift\") pod \"swift-storage-0\" (UID: \"0c8d183a-f48b-4a69-8685-90c834e678ce\") " pod="openstack/swift-storage-0" Oct 02 02:02:31 crc kubenswrapper[4885]: I1002 02:02:31.430241 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d471d2b-912c-4f49-a7ec-c07b85c02bd9-combined-ca-bundle\") pod \"swift-ring-rebalance-9wd49\" (UID: \"2d471d2b-912c-4f49-a7ec-c07b85c02bd9\") " pod="openstack/swift-ring-rebalance-9wd49" Oct 02 02:02:31 crc kubenswrapper[4885]: I1002 02:02:31.430316 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/2d471d2b-912c-4f49-a7ec-c07b85c02bd9-swiftconf\") pod \"swift-ring-rebalance-9wd49\" (UID: \"2d471d2b-912c-4f49-a7ec-c07b85c02bd9\") " pod="openstack/swift-ring-rebalance-9wd49" Oct 02 02:02:31 crc kubenswrapper[4885]: E1002 02:02:31.430474 4885 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 02 02:02:31 crc kubenswrapper[4885]: E1002 02:02:31.430491 4885 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 02 02:02:31 crc kubenswrapper[4885]: E1002 02:02:31.430545 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/0c8d183a-f48b-4a69-8685-90c834e678ce-etc-swift podName:0c8d183a-f48b-4a69-8685-90c834e678ce nodeName:}" failed. No retries permitted until 2025-10-02 02:02:32.430514498 +0000 UTC m=+941.242261897 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/0c8d183a-f48b-4a69-8685-90c834e678ce-etc-swift") pod "swift-storage-0" (UID: "0c8d183a-f48b-4a69-8685-90c834e678ce") : configmap "swift-ring-files" not found Oct 02 02:02:31 crc kubenswrapper[4885]: I1002 02:02:31.531146 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd73a7ac-f025-4cf2-b09c-26e3eb7096c6-config\") pod \"fd73a7ac-f025-4cf2-b09c-26e3eb7096c6\" (UID: \"fd73a7ac-f025-4cf2-b09c-26e3eb7096c6\") " Oct 02 02:02:31 crc kubenswrapper[4885]: I1002 02:02:31.531187 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fd73a7ac-f025-4cf2-b09c-26e3eb7096c6-dns-svc\") pod \"fd73a7ac-f025-4cf2-b09c-26e3eb7096c6\" (UID: \"fd73a7ac-f025-4cf2-b09c-26e3eb7096c6\") " Oct 02 02:02:31 crc kubenswrapper[4885]: I1002 02:02:31.531239 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fd73a7ac-f025-4cf2-b09c-26e3eb7096c6-ovsdbserver-sb\") pod \"fd73a7ac-f025-4cf2-b09c-26e3eb7096c6\" (UID: \"fd73a7ac-f025-4cf2-b09c-26e3eb7096c6\") " Oct 02 02:02:31 crc kubenswrapper[4885]: I1002 02:02:31.531377 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t8hrq\" (UniqueName: \"kubernetes.io/projected/fd73a7ac-f025-4cf2-b09c-26e3eb7096c6-kube-api-access-t8hrq\") pod \"fd73a7ac-f025-4cf2-b09c-26e3eb7096c6\" (UID: \"fd73a7ac-f025-4cf2-b09c-26e3eb7096c6\") " Oct 02 02:02:31 crc kubenswrapper[4885]: I1002 02:02:31.531419 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fd73a7ac-f025-4cf2-b09c-26e3eb7096c6-ovsdbserver-nb\") pod \"fd73a7ac-f025-4cf2-b09c-26e3eb7096c6\" (UID: \"fd73a7ac-f025-4cf2-b09c-26e3eb7096c6\") " Oct 02 02:02:31 crc kubenswrapper[4885]: I1002 02:02:31.531603 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/2d471d2b-912c-4f49-a7ec-c07b85c02bd9-swiftconf\") pod \"swift-ring-rebalance-9wd49\" (UID: \"2d471d2b-912c-4f49-a7ec-c07b85c02bd9\") " pod="openstack/swift-ring-rebalance-9wd49" Oct 02 02:02:31 crc kubenswrapper[4885]: I1002 02:02:31.531671 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/2d471d2b-912c-4f49-a7ec-c07b85c02bd9-ring-data-devices\") pod \"swift-ring-rebalance-9wd49\" (UID: \"2d471d2b-912c-4f49-a7ec-c07b85c02bd9\") " pod="openstack/swift-ring-rebalance-9wd49" Oct 02 02:02:31 crc kubenswrapper[4885]: I1002 02:02:31.531710 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2d471d2b-912c-4f49-a7ec-c07b85c02bd9-scripts\") pod \"swift-ring-rebalance-9wd49\" (UID: \"2d471d2b-912c-4f49-a7ec-c07b85c02bd9\") " pod="openstack/swift-ring-rebalance-9wd49" Oct 02 02:02:31 crc kubenswrapper[4885]: I1002 02:02:31.531731 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/2d471d2b-912c-4f49-a7ec-c07b85c02bd9-dispersionconf\") pod \"swift-ring-rebalance-9wd49\" (UID: \"2d471d2b-912c-4f49-a7ec-c07b85c02bd9\") " pod="openstack/swift-ring-rebalance-9wd49" Oct 02 02:02:31 crc kubenswrapper[4885]: I1002 02:02:31.531749 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/2d471d2b-912c-4f49-a7ec-c07b85c02bd9-etc-swift\") pod \"swift-ring-rebalance-9wd49\" (UID: \"2d471d2b-912c-4f49-a7ec-c07b85c02bd9\") " pod="openstack/swift-ring-rebalance-9wd49" Oct 02 02:02:31 crc kubenswrapper[4885]: I1002 02:02:31.531770 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qc2q7\" (UniqueName: \"kubernetes.io/projected/2d471d2b-912c-4f49-a7ec-c07b85c02bd9-kube-api-access-qc2q7\") pod \"swift-ring-rebalance-9wd49\" (UID: \"2d471d2b-912c-4f49-a7ec-c07b85c02bd9\") " pod="openstack/swift-ring-rebalance-9wd49" Oct 02 02:02:31 crc kubenswrapper[4885]: I1002 02:02:31.531807 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d471d2b-912c-4f49-a7ec-c07b85c02bd9-combined-ca-bundle\") pod \"swift-ring-rebalance-9wd49\" (UID: \"2d471d2b-912c-4f49-a7ec-c07b85c02bd9\") " pod="openstack/swift-ring-rebalance-9wd49" Oct 02 02:02:31 crc kubenswrapper[4885]: I1002 02:02:31.532545 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/2d471d2b-912c-4f49-a7ec-c07b85c02bd9-etc-swift\") pod \"swift-ring-rebalance-9wd49\" (UID: \"2d471d2b-912c-4f49-a7ec-c07b85c02bd9\") " pod="openstack/swift-ring-rebalance-9wd49" Oct 02 02:02:31 crc kubenswrapper[4885]: I1002 02:02:31.532853 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2d471d2b-912c-4f49-a7ec-c07b85c02bd9-scripts\") pod \"swift-ring-rebalance-9wd49\" (UID: \"2d471d2b-912c-4f49-a7ec-c07b85c02bd9\") " pod="openstack/swift-ring-rebalance-9wd49" Oct 02 02:02:31 crc kubenswrapper[4885]: I1002 02:02:31.533337 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/2d471d2b-912c-4f49-a7ec-c07b85c02bd9-ring-data-devices\") pod \"swift-ring-rebalance-9wd49\" (UID: \"2d471d2b-912c-4f49-a7ec-c07b85c02bd9\") " pod="openstack/swift-ring-rebalance-9wd49" Oct 02 02:02:31 crc kubenswrapper[4885]: I1002 02:02:31.538808 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fd73a7ac-f025-4cf2-b09c-26e3eb7096c6-kube-api-access-t8hrq" (OuterVolumeSpecName: "kube-api-access-t8hrq") pod "fd73a7ac-f025-4cf2-b09c-26e3eb7096c6" (UID: "fd73a7ac-f025-4cf2-b09c-26e3eb7096c6"). InnerVolumeSpecName "kube-api-access-t8hrq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:02:31 crc kubenswrapper[4885]: I1002 02:02:31.539186 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d471d2b-912c-4f49-a7ec-c07b85c02bd9-combined-ca-bundle\") pod \"swift-ring-rebalance-9wd49\" (UID: \"2d471d2b-912c-4f49-a7ec-c07b85c02bd9\") " pod="openstack/swift-ring-rebalance-9wd49" Oct 02 02:02:31 crc kubenswrapper[4885]: I1002 02:02:31.539201 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/2d471d2b-912c-4f49-a7ec-c07b85c02bd9-swiftconf\") pod \"swift-ring-rebalance-9wd49\" (UID: \"2d471d2b-912c-4f49-a7ec-c07b85c02bd9\") " pod="openstack/swift-ring-rebalance-9wd49" Oct 02 02:02:31 crc kubenswrapper[4885]: I1002 02:02:31.539684 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/2d471d2b-912c-4f49-a7ec-c07b85c02bd9-dispersionconf\") pod \"swift-ring-rebalance-9wd49\" (UID: \"2d471d2b-912c-4f49-a7ec-c07b85c02bd9\") " pod="openstack/swift-ring-rebalance-9wd49" Oct 02 02:02:31 crc kubenswrapper[4885]: I1002 02:02:31.558673 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qc2q7\" (UniqueName: \"kubernetes.io/projected/2d471d2b-912c-4f49-a7ec-c07b85c02bd9-kube-api-access-qc2q7\") pod \"swift-ring-rebalance-9wd49\" (UID: \"2d471d2b-912c-4f49-a7ec-c07b85c02bd9\") " pod="openstack/swift-ring-rebalance-9wd49" Oct 02 02:02:31 crc kubenswrapper[4885]: I1002 02:02:31.567967 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-9wd49" Oct 02 02:02:31 crc kubenswrapper[4885]: I1002 02:02:31.579940 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fd73a7ac-f025-4cf2-b09c-26e3eb7096c6-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "fd73a7ac-f025-4cf2-b09c-26e3eb7096c6" (UID: "fd73a7ac-f025-4cf2-b09c-26e3eb7096c6"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:02:31 crc kubenswrapper[4885]: I1002 02:02:31.580736 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fd73a7ac-f025-4cf2-b09c-26e3eb7096c6-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "fd73a7ac-f025-4cf2-b09c-26e3eb7096c6" (UID: "fd73a7ac-f025-4cf2-b09c-26e3eb7096c6"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:02:31 crc kubenswrapper[4885]: I1002 02:02:31.599355 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fd73a7ac-f025-4cf2-b09c-26e3eb7096c6-config" (OuterVolumeSpecName: "config") pod "fd73a7ac-f025-4cf2-b09c-26e3eb7096c6" (UID: "fd73a7ac-f025-4cf2-b09c-26e3eb7096c6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:02:31 crc kubenswrapper[4885]: I1002 02:02:31.600414 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fd73a7ac-f025-4cf2-b09c-26e3eb7096c6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "fd73a7ac-f025-4cf2-b09c-26e3eb7096c6" (UID: "fd73a7ac-f025-4cf2-b09c-26e3eb7096c6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:02:31 crc kubenswrapper[4885]: I1002 02:02:31.633428 4885 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd73a7ac-f025-4cf2-b09c-26e3eb7096c6-config\") on node \"crc\" DevicePath \"\"" Oct 02 02:02:31 crc kubenswrapper[4885]: I1002 02:02:31.633725 4885 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fd73a7ac-f025-4cf2-b09c-26e3eb7096c6-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 02:02:31 crc kubenswrapper[4885]: I1002 02:02:31.633737 4885 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fd73a7ac-f025-4cf2-b09c-26e3eb7096c6-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 02:02:31 crc kubenswrapper[4885]: I1002 02:02:31.633753 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t8hrq\" (UniqueName: \"kubernetes.io/projected/fd73a7ac-f025-4cf2-b09c-26e3eb7096c6-kube-api-access-t8hrq\") on node \"crc\" DevicePath \"\"" Oct 02 02:02:31 crc kubenswrapper[4885]: I1002 02:02:31.633766 4885 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fd73a7ac-f025-4cf2-b09c-26e3eb7096c6-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 02:02:31 crc kubenswrapper[4885]: I1002 02:02:31.932340 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-qkx2j" event={"ID":"52b80336-4c74-4236-9600-7ea4606e5d6f","Type":"ContainerStarted","Data":"13d632d093ec39287feac2e2cd715160da8bd3673c8eb579a58d27664223a705"} Oct 02 02:02:31 crc kubenswrapper[4885]: I1002 02:02:31.932483 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-698758b865-qkx2j" Oct 02 02:02:31 crc kubenswrapper[4885]: I1002 02:02:31.933807 4885 generic.go:334] "Generic (PLEG): container finished" podID="fd73a7ac-f025-4cf2-b09c-26e3eb7096c6" containerID="2c77e4bde7e6a6a3d5019ef6001f7556d9a38620f43ae8597f351f41e996df8c" exitCode=0 Oct 02 02:02:31 crc kubenswrapper[4885]: I1002 02:02:31.933842 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-z5rln" event={"ID":"fd73a7ac-f025-4cf2-b09c-26e3eb7096c6","Type":"ContainerDied","Data":"2c77e4bde7e6a6a3d5019ef6001f7556d9a38620f43ae8597f351f41e996df8c"} Oct 02 02:02:31 crc kubenswrapper[4885]: I1002 02:02:31.933863 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-z5rln" event={"ID":"fd73a7ac-f025-4cf2-b09c-26e3eb7096c6","Type":"ContainerDied","Data":"72d8e15931c507955b4affeb845000fbd62a3e0d694f29cac6bd8e8e31ec414c"} Oct 02 02:02:31 crc kubenswrapper[4885]: I1002 02:02:31.933881 4885 scope.go:117] "RemoveContainer" containerID="2c77e4bde7e6a6a3d5019ef6001f7556d9a38620f43ae8597f351f41e996df8c" Oct 02 02:02:31 crc kubenswrapper[4885]: I1002 02:02:31.933938 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-z5rln" Oct 02 02:02:31 crc kubenswrapper[4885]: I1002 02:02:31.957984 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-698758b865-qkx2j" podStartSLOduration=2.957962366 podStartE2EDuration="2.957962366s" podCreationTimestamp="2025-10-02 02:02:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:02:31.948169058 +0000 UTC m=+940.759916467" watchObservedRunningTime="2025-10-02 02:02:31.957962366 +0000 UTC m=+940.769709785" Oct 02 02:02:31 crc kubenswrapper[4885]: I1002 02:02:31.960172 4885 scope.go:117] "RemoveContainer" containerID="c7f5dd99660e2d1123057f565c6bea67fc73d1d20b56c586e37566f67b0e79e1" Oct 02 02:02:31 crc kubenswrapper[4885]: I1002 02:02:31.981694 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-z5rln"] Oct 02 02:02:31 crc kubenswrapper[4885]: I1002 02:02:31.989369 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-z5rln"] Oct 02 02:02:32 crc kubenswrapper[4885]: I1002 02:02:32.002464 4885 scope.go:117] "RemoveContainer" containerID="2c77e4bde7e6a6a3d5019ef6001f7556d9a38620f43ae8597f351f41e996df8c" Oct 02 02:02:32 crc kubenswrapper[4885]: E1002 02:02:32.003551 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2c77e4bde7e6a6a3d5019ef6001f7556d9a38620f43ae8597f351f41e996df8c\": container with ID starting with 2c77e4bde7e6a6a3d5019ef6001f7556d9a38620f43ae8597f351f41e996df8c not found: ID does not exist" containerID="2c77e4bde7e6a6a3d5019ef6001f7556d9a38620f43ae8597f351f41e996df8c" Oct 02 02:02:32 crc kubenswrapper[4885]: I1002 02:02:32.003608 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2c77e4bde7e6a6a3d5019ef6001f7556d9a38620f43ae8597f351f41e996df8c"} err="failed to get container status \"2c77e4bde7e6a6a3d5019ef6001f7556d9a38620f43ae8597f351f41e996df8c\": rpc error: code = NotFound desc = could not find container \"2c77e4bde7e6a6a3d5019ef6001f7556d9a38620f43ae8597f351f41e996df8c\": container with ID starting with 2c77e4bde7e6a6a3d5019ef6001f7556d9a38620f43ae8597f351f41e996df8c not found: ID does not exist" Oct 02 02:02:32 crc kubenswrapper[4885]: I1002 02:02:32.003637 4885 scope.go:117] "RemoveContainer" containerID="c7f5dd99660e2d1123057f565c6bea67fc73d1d20b56c586e37566f67b0e79e1" Oct 02 02:02:32 crc kubenswrapper[4885]: E1002 02:02:32.006498 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c7f5dd99660e2d1123057f565c6bea67fc73d1d20b56c586e37566f67b0e79e1\": container with ID starting with c7f5dd99660e2d1123057f565c6bea67fc73d1d20b56c586e37566f67b0e79e1 not found: ID does not exist" containerID="c7f5dd99660e2d1123057f565c6bea67fc73d1d20b56c586e37566f67b0e79e1" Oct 02 02:02:32 crc kubenswrapper[4885]: I1002 02:02:32.006721 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c7f5dd99660e2d1123057f565c6bea67fc73d1d20b56c586e37566f67b0e79e1"} err="failed to get container status \"c7f5dd99660e2d1123057f565c6bea67fc73d1d20b56c586e37566f67b0e79e1\": rpc error: code = NotFound desc = could not find container \"c7f5dd99660e2d1123057f565c6bea67fc73d1d20b56c586e37566f67b0e79e1\": container with ID starting with c7f5dd99660e2d1123057f565c6bea67fc73d1d20b56c586e37566f67b0e79e1 not found: ID does not exist" Oct 02 02:02:32 crc kubenswrapper[4885]: I1002 02:02:32.024859 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-9wd49"] Oct 02 02:02:32 crc kubenswrapper[4885]: I1002 02:02:32.068986 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fd73a7ac-f025-4cf2-b09c-26e3eb7096c6" path="/var/lib/kubelet/pods/fd73a7ac-f025-4cf2-b09c-26e3eb7096c6/volumes" Oct 02 02:02:32 crc kubenswrapper[4885]: I1002 02:02:32.083240 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Oct 02 02:02:32 crc kubenswrapper[4885]: I1002 02:02:32.139840 4885 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-cell1-galera-0" podUID="87196135-8d40-47cb-b918-98aea04c7c17" containerName="galera" probeResult="failure" output=< Oct 02 02:02:32 crc kubenswrapper[4885]: wsrep_local_state_comment (Joined) differs from Synced Oct 02 02:02:32 crc kubenswrapper[4885]: > Oct 02 02:02:32 crc kubenswrapper[4885]: I1002 02:02:32.444795 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/0c8d183a-f48b-4a69-8685-90c834e678ce-etc-swift\") pod \"swift-storage-0\" (UID: \"0c8d183a-f48b-4a69-8685-90c834e678ce\") " pod="openstack/swift-storage-0" Oct 02 02:02:32 crc kubenswrapper[4885]: E1002 02:02:32.444956 4885 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 02 02:02:32 crc kubenswrapper[4885]: E1002 02:02:32.444973 4885 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 02 02:02:32 crc kubenswrapper[4885]: E1002 02:02:32.445028 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/0c8d183a-f48b-4a69-8685-90c834e678ce-etc-swift podName:0c8d183a-f48b-4a69-8685-90c834e678ce nodeName:}" failed. No retries permitted until 2025-10-02 02:02:34.445009845 +0000 UTC m=+943.256757244 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/0c8d183a-f48b-4a69-8685-90c834e678ce-etc-swift") pod "swift-storage-0" (UID: "0c8d183a-f48b-4a69-8685-90c834e678ce") : configmap "swift-ring-files" not found Oct 02 02:02:32 crc kubenswrapper[4885]: E1002 02:02:32.626118 4885 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.39:53134->38.102.83.39:35229: write tcp 38.102.83.39:53134->38.102.83.39:35229: write: broken pipe Oct 02 02:02:32 crc kubenswrapper[4885]: I1002 02:02:32.943003 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"faff5405-00c6-4d75-bbca-77f9230c2811","Type":"ContainerStarted","Data":"80e36c78fd1078a8f7272a7a784beeca583573d3867f310d6a61e997fce2d26a"} Oct 02 02:02:32 crc kubenswrapper[4885]: I1002 02:02:32.943795 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 02 02:02:32 crc kubenswrapper[4885]: I1002 02:02:32.944310 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-9wd49" event={"ID":"2d471d2b-912c-4f49-a7ec-c07b85c02bd9","Type":"ContainerStarted","Data":"ec92831404be01280133ce230b8010328797f0a590781887b3c1ff15695ee3db"} Oct 02 02:02:32 crc kubenswrapper[4885]: I1002 02:02:32.966832 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=8.724046809 podStartE2EDuration="43.966815583s" podCreationTimestamp="2025-10-02 02:01:49 +0000 UTC" firstStartedPulling="2025-10-02 02:01:57.293368604 +0000 UTC m=+906.105116003" lastFinishedPulling="2025-10-02 02:02:32.536137368 +0000 UTC m=+941.347884777" observedRunningTime="2025-10-02 02:02:32.959125119 +0000 UTC m=+941.770872518" watchObservedRunningTime="2025-10-02 02:02:32.966815583 +0000 UTC m=+941.778562982" Oct 02 02:02:34 crc kubenswrapper[4885]: I1002 02:02:34.484715 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/0c8d183a-f48b-4a69-8685-90c834e678ce-etc-swift\") pod \"swift-storage-0\" (UID: \"0c8d183a-f48b-4a69-8685-90c834e678ce\") " pod="openstack/swift-storage-0" Oct 02 02:02:34 crc kubenswrapper[4885]: E1002 02:02:34.485034 4885 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 02 02:02:34 crc kubenswrapper[4885]: E1002 02:02:34.485593 4885 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 02 02:02:34 crc kubenswrapper[4885]: E1002 02:02:34.485722 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/0c8d183a-f48b-4a69-8685-90c834e678ce-etc-swift podName:0c8d183a-f48b-4a69-8685-90c834e678ce nodeName:}" failed. No retries permitted until 2025-10-02 02:02:38.485681787 +0000 UTC m=+947.297429196 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/0c8d183a-f48b-4a69-8685-90c834e678ce-etc-swift") pod "swift-storage-0" (UID: "0c8d183a-f48b-4a69-8685-90c834e678ce") : configmap "swift-ring-files" not found Oct 02 02:02:36 crc kubenswrapper[4885]: I1002 02:02:36.422145 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Oct 02 02:02:36 crc kubenswrapper[4885]: I1002 02:02:36.491982 4885 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-galera-0" podUID="8427e298-965b-48f0-821f-bd078fa2e96f" containerName="galera" probeResult="failure" output=< Oct 02 02:02:36 crc kubenswrapper[4885]: wsrep_local_state_comment (Joined) differs from Synced Oct 02 02:02:36 crc kubenswrapper[4885]: > Oct 02 02:02:37 crc kubenswrapper[4885]: I1002 02:02:37.273542 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Oct 02 02:02:37 crc kubenswrapper[4885]: I1002 02:02:37.408750 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Oct 02 02:02:37 crc kubenswrapper[4885]: I1002 02:02:37.748861 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-7dl92"] Oct 02 02:02:37 crc kubenswrapper[4885]: E1002 02:02:37.749150 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd73a7ac-f025-4cf2-b09c-26e3eb7096c6" containerName="dnsmasq-dns" Oct 02 02:02:37 crc kubenswrapper[4885]: I1002 02:02:37.749162 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd73a7ac-f025-4cf2-b09c-26e3eb7096c6" containerName="dnsmasq-dns" Oct 02 02:02:37 crc kubenswrapper[4885]: E1002 02:02:37.749184 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd73a7ac-f025-4cf2-b09c-26e3eb7096c6" containerName="init" Oct 02 02:02:37 crc kubenswrapper[4885]: I1002 02:02:37.749192 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd73a7ac-f025-4cf2-b09c-26e3eb7096c6" containerName="init" Oct 02 02:02:37 crc kubenswrapper[4885]: I1002 02:02:37.749349 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd73a7ac-f025-4cf2-b09c-26e3eb7096c6" containerName="dnsmasq-dns" Oct 02 02:02:37 crc kubenswrapper[4885]: I1002 02:02:37.749820 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-7dl92" Oct 02 02:02:37 crc kubenswrapper[4885]: I1002 02:02:37.759519 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-7dl92"] Oct 02 02:02:37 crc kubenswrapper[4885]: I1002 02:02:37.849985 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vsjvk\" (UniqueName: \"kubernetes.io/projected/da80ee88-0786-4a7d-9c62-446412d63021-kube-api-access-vsjvk\") pod \"placement-db-create-7dl92\" (UID: \"da80ee88-0786-4a7d-9c62-446412d63021\") " pod="openstack/placement-db-create-7dl92" Oct 02 02:02:37 crc kubenswrapper[4885]: I1002 02:02:37.951925 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vsjvk\" (UniqueName: \"kubernetes.io/projected/da80ee88-0786-4a7d-9c62-446412d63021-kube-api-access-vsjvk\") pod \"placement-db-create-7dl92\" (UID: \"da80ee88-0786-4a7d-9c62-446412d63021\") " pod="openstack/placement-db-create-7dl92" Oct 02 02:02:37 crc kubenswrapper[4885]: I1002 02:02:37.973613 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vsjvk\" (UniqueName: \"kubernetes.io/projected/da80ee88-0786-4a7d-9c62-446412d63021-kube-api-access-vsjvk\") pod \"placement-db-create-7dl92\" (UID: \"da80ee88-0786-4a7d-9c62-446412d63021\") " pod="openstack/placement-db-create-7dl92" Oct 02 02:02:38 crc kubenswrapper[4885]: I1002 02:02:38.098636 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-7dl92" Oct 02 02:02:38 crc kubenswrapper[4885]: I1002 02:02:38.564560 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/0c8d183a-f48b-4a69-8685-90c834e678ce-etc-swift\") pod \"swift-storage-0\" (UID: \"0c8d183a-f48b-4a69-8685-90c834e678ce\") " pod="openstack/swift-storage-0" Oct 02 02:02:38 crc kubenswrapper[4885]: E1002 02:02:38.564848 4885 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 02 02:02:38 crc kubenswrapper[4885]: E1002 02:02:38.564893 4885 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 02 02:02:38 crc kubenswrapper[4885]: E1002 02:02:38.564977 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/0c8d183a-f48b-4a69-8685-90c834e678ce-etc-swift podName:0c8d183a-f48b-4a69-8685-90c834e678ce nodeName:}" failed. No retries permitted until 2025-10-02 02:02:46.564948986 +0000 UTC m=+955.376696415 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/0c8d183a-f48b-4a69-8685-90c834e678ce-etc-swift") pod "swift-storage-0" (UID: "0c8d183a-f48b-4a69-8685-90c834e678ce") : configmap "swift-ring-files" not found Oct 02 02:02:39 crc kubenswrapper[4885]: I1002 02:02:39.621921 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 02 02:02:39 crc kubenswrapper[4885]: I1002 02:02:39.949484 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-698758b865-qkx2j" Oct 02 02:02:40 crc kubenswrapper[4885]: I1002 02:02:40.065003 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-p9cqp"] Oct 02 02:02:40 crc kubenswrapper[4885]: I1002 02:02:40.065390 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57d769cc4f-p9cqp" podUID="90626475-ee00-431f-8992-d5fc20728c6c" containerName="dnsmasq-dns" containerID="cri-o://3b8c253cf16ce810fd1834840e2d1116620e42562f6f50eea87012550966a7b3" gracePeriod=10 Oct 02 02:02:41 crc kubenswrapper[4885]: I1002 02:02:41.034884 4885 generic.go:334] "Generic (PLEG): container finished" podID="90626475-ee00-431f-8992-d5fc20728c6c" containerID="3b8c253cf16ce810fd1834840e2d1116620e42562f6f50eea87012550966a7b3" exitCode=0 Oct 02 02:02:41 crc kubenswrapper[4885]: I1002 02:02:41.035007 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-p9cqp" event={"ID":"90626475-ee00-431f-8992-d5fc20728c6c","Type":"ContainerDied","Data":"3b8c253cf16ce810fd1834840e2d1116620e42562f6f50eea87012550966a7b3"} Oct 02 02:02:42 crc kubenswrapper[4885]: I1002 02:02:42.617871 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-7dl92"] Oct 02 02:02:42 crc kubenswrapper[4885]: W1002 02:02:42.753697 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podda80ee88_0786_4a7d_9c62_446412d63021.slice/crio-b0daa8796672251663d94c655c6e07f91b5df6b746d70b4c4feff916e010f274 WatchSource:0}: Error finding container b0daa8796672251663d94c655c6e07f91b5df6b746d70b4c4feff916e010f274: Status 404 returned error can't find the container with id b0daa8796672251663d94c655c6e07f91b5df6b746d70b4c4feff916e010f274 Oct 02 02:02:42 crc kubenswrapper[4885]: I1002 02:02:42.822631 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-p9cqp" Oct 02 02:02:42 crc kubenswrapper[4885]: I1002 02:02:42.958098 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/90626475-ee00-431f-8992-d5fc20728c6c-config\") pod \"90626475-ee00-431f-8992-d5fc20728c6c\" (UID: \"90626475-ee00-431f-8992-d5fc20728c6c\") " Oct 02 02:02:42 crc kubenswrapper[4885]: I1002 02:02:42.958427 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2tgx5\" (UniqueName: \"kubernetes.io/projected/90626475-ee00-431f-8992-d5fc20728c6c-kube-api-access-2tgx5\") pod \"90626475-ee00-431f-8992-d5fc20728c6c\" (UID: \"90626475-ee00-431f-8992-d5fc20728c6c\") " Oct 02 02:02:42 crc kubenswrapper[4885]: I1002 02:02:42.958665 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/90626475-ee00-431f-8992-d5fc20728c6c-dns-svc\") pod \"90626475-ee00-431f-8992-d5fc20728c6c\" (UID: \"90626475-ee00-431f-8992-d5fc20728c6c\") " Oct 02 02:02:42 crc kubenswrapper[4885]: I1002 02:02:42.961402 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/90626475-ee00-431f-8992-d5fc20728c6c-kube-api-access-2tgx5" (OuterVolumeSpecName: "kube-api-access-2tgx5") pod "90626475-ee00-431f-8992-d5fc20728c6c" (UID: "90626475-ee00-431f-8992-d5fc20728c6c"). InnerVolumeSpecName "kube-api-access-2tgx5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:02:42 crc kubenswrapper[4885]: I1002 02:02:42.997819 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/90626475-ee00-431f-8992-d5fc20728c6c-config" (OuterVolumeSpecName: "config") pod "90626475-ee00-431f-8992-d5fc20728c6c" (UID: "90626475-ee00-431f-8992-d5fc20728c6c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:02:43 crc kubenswrapper[4885]: I1002 02:02:43.011609 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/90626475-ee00-431f-8992-d5fc20728c6c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "90626475-ee00-431f-8992-d5fc20728c6c" (UID: "90626475-ee00-431f-8992-d5fc20728c6c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:02:43 crc kubenswrapper[4885]: I1002 02:02:43.029119 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-dnlx6"] Oct 02 02:02:43 crc kubenswrapper[4885]: E1002 02:02:43.029518 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90626475-ee00-431f-8992-d5fc20728c6c" containerName="dnsmasq-dns" Oct 02 02:02:43 crc kubenswrapper[4885]: I1002 02:02:43.029536 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="90626475-ee00-431f-8992-d5fc20728c6c" containerName="dnsmasq-dns" Oct 02 02:02:43 crc kubenswrapper[4885]: E1002 02:02:43.029558 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90626475-ee00-431f-8992-d5fc20728c6c" containerName="init" Oct 02 02:02:43 crc kubenswrapper[4885]: I1002 02:02:43.029565 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="90626475-ee00-431f-8992-d5fc20728c6c" containerName="init" Oct 02 02:02:43 crc kubenswrapper[4885]: I1002 02:02:43.029719 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="90626475-ee00-431f-8992-d5fc20728c6c" containerName="dnsmasq-dns" Oct 02 02:02:43 crc kubenswrapper[4885]: I1002 02:02:43.030232 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-dnlx6" Oct 02 02:02:43 crc kubenswrapper[4885]: I1002 02:02:43.037342 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-dnlx6"] Oct 02 02:02:43 crc kubenswrapper[4885]: I1002 02:02:43.058788 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-7dl92" event={"ID":"da80ee88-0786-4a7d-9c62-446412d63021","Type":"ContainerStarted","Data":"49363e2fbd3e9324689ef02f8d32d917ce193a0b4ab530b68afaac7b0433606a"} Oct 02 02:02:43 crc kubenswrapper[4885]: I1002 02:02:43.058856 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-7dl92" event={"ID":"da80ee88-0786-4a7d-9c62-446412d63021","Type":"ContainerStarted","Data":"b0daa8796672251663d94c655c6e07f91b5df6b746d70b4c4feff916e010f274"} Oct 02 02:02:43 crc kubenswrapper[4885]: I1002 02:02:43.060232 4885 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/90626475-ee00-431f-8992-d5fc20728c6c-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 02:02:43 crc kubenswrapper[4885]: I1002 02:02:43.060278 4885 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/90626475-ee00-431f-8992-d5fc20728c6c-config\") on node \"crc\" DevicePath \"\"" Oct 02 02:02:43 crc kubenswrapper[4885]: I1002 02:02:43.060292 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2tgx5\" (UniqueName: \"kubernetes.io/projected/90626475-ee00-431f-8992-d5fc20728c6c-kube-api-access-2tgx5\") on node \"crc\" DevicePath \"\"" Oct 02 02:02:43 crc kubenswrapper[4885]: I1002 02:02:43.060756 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"33286960-496c-4f14-a334-c6dc765ca956","Type":"ContainerStarted","Data":"be4517d5fbf06f7fb90d626d231e26fc62f9c7d31b04f2fb13fa66339046b168"} Oct 02 02:02:43 crc kubenswrapper[4885]: I1002 02:02:43.063424 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-p9cqp" event={"ID":"90626475-ee00-431f-8992-d5fc20728c6c","Type":"ContainerDied","Data":"b9202f38c3e6b86dff94c2d73561c7e9d89082b2a9efa38a64f9bfd18a8c4c84"} Oct 02 02:02:43 crc kubenswrapper[4885]: I1002 02:02:43.064167 4885 scope.go:117] "RemoveContainer" containerID="3b8c253cf16ce810fd1834840e2d1116620e42562f6f50eea87012550966a7b3" Oct 02 02:02:43 crc kubenswrapper[4885]: I1002 02:02:43.064333 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-p9cqp" Oct 02 02:02:43 crc kubenswrapper[4885]: I1002 02:02:43.085431 4885 scope.go:117] "RemoveContainer" containerID="35fff824d306ebab1e80c6e4cb5ff69baa9c27248d26050fa3683c38aa8b261c" Oct 02 02:02:43 crc kubenswrapper[4885]: I1002 02:02:43.087966 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=4.223531412 podStartE2EDuration="48.087946737s" podCreationTimestamp="2025-10-02 02:01:55 +0000 UTC" firstStartedPulling="2025-10-02 02:01:58.249243088 +0000 UTC m=+907.060990487" lastFinishedPulling="2025-10-02 02:02:42.113658373 +0000 UTC m=+950.925405812" observedRunningTime="2025-10-02 02:02:43.081660215 +0000 UTC m=+951.893407624" watchObservedRunningTime="2025-10-02 02:02:43.087946737 +0000 UTC m=+951.899694146" Oct 02 02:02:43 crc kubenswrapper[4885]: I1002 02:02:43.117218 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-p9cqp"] Oct 02 02:02:43 crc kubenswrapper[4885]: I1002 02:02:43.123058 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-p9cqp"] Oct 02 02:02:43 crc kubenswrapper[4885]: I1002 02:02:43.162248 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bs42l\" (UniqueName: \"kubernetes.io/projected/7b8634a8-6199-42bf-b18d-592a8ba3d164-kube-api-access-bs42l\") pod \"glance-db-create-dnlx6\" (UID: \"7b8634a8-6199-42bf-b18d-592a8ba3d164\") " pod="openstack/glance-db-create-dnlx6" Oct 02 02:02:43 crc kubenswrapper[4885]: I1002 02:02:43.264628 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bs42l\" (UniqueName: \"kubernetes.io/projected/7b8634a8-6199-42bf-b18d-592a8ba3d164-kube-api-access-bs42l\") pod \"glance-db-create-dnlx6\" (UID: \"7b8634a8-6199-42bf-b18d-592a8ba3d164\") " pod="openstack/glance-db-create-dnlx6" Oct 02 02:02:43 crc kubenswrapper[4885]: I1002 02:02:43.265565 4885 patch_prober.go:28] interesting pod/machine-config-daemon-rttx8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 02:02:43 crc kubenswrapper[4885]: I1002 02:02:43.265644 4885 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 02:02:43 crc kubenswrapper[4885]: I1002 02:02:43.287737 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bs42l\" (UniqueName: \"kubernetes.io/projected/7b8634a8-6199-42bf-b18d-592a8ba3d164-kube-api-access-bs42l\") pod \"glance-db-create-dnlx6\" (UID: \"7b8634a8-6199-42bf-b18d-592a8ba3d164\") " pod="openstack/glance-db-create-dnlx6" Oct 02 02:02:43 crc kubenswrapper[4885]: I1002 02:02:43.354766 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-dnlx6" Oct 02 02:02:43 crc kubenswrapper[4885]: W1002 02:02:43.824591 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7b8634a8_6199_42bf_b18d_592a8ba3d164.slice/crio-d1aaedb0f9b1b5cee5f93363ae4b77247f6d4d3473489c1ec2c926956bd5951d WatchSource:0}: Error finding container d1aaedb0f9b1b5cee5f93363ae4b77247f6d4d3473489c1ec2c926956bd5951d: Status 404 returned error can't find the container with id d1aaedb0f9b1b5cee5f93363ae4b77247f6d4d3473489c1ec2c926956bd5951d Oct 02 02:02:43 crc kubenswrapper[4885]: I1002 02:02:43.826849 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-dnlx6"] Oct 02 02:02:44 crc kubenswrapper[4885]: I1002 02:02:44.063171 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="90626475-ee00-431f-8992-d5fc20728c6c" path="/var/lib/kubelet/pods/90626475-ee00-431f-8992-d5fc20728c6c/volumes" Oct 02 02:02:44 crc kubenswrapper[4885]: I1002 02:02:44.075593 4885 generic.go:334] "Generic (PLEG): container finished" podID="da80ee88-0786-4a7d-9c62-446412d63021" containerID="49363e2fbd3e9324689ef02f8d32d917ce193a0b4ab530b68afaac7b0433606a" exitCode=0 Oct 02 02:02:44 crc kubenswrapper[4885]: I1002 02:02:44.075659 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-7dl92" event={"ID":"da80ee88-0786-4a7d-9c62-446412d63021","Type":"ContainerDied","Data":"49363e2fbd3e9324689ef02f8d32d917ce193a0b4ab530b68afaac7b0433606a"} Oct 02 02:02:44 crc kubenswrapper[4885]: I1002 02:02:44.078113 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-dnlx6" event={"ID":"7b8634a8-6199-42bf-b18d-592a8ba3d164","Type":"ContainerStarted","Data":"e8887f0cf1208a83481402ef4a9f149d8af949103cfd35b53383a946e2763dd8"} Oct 02 02:02:44 crc kubenswrapper[4885]: I1002 02:02:44.078226 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-dnlx6" event={"ID":"7b8634a8-6199-42bf-b18d-592a8ba3d164","Type":"ContainerStarted","Data":"d1aaedb0f9b1b5cee5f93363ae4b77247f6d4d3473489c1ec2c926956bd5951d"} Oct 02 02:02:44 crc kubenswrapper[4885]: I1002 02:02:44.079630 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-9wd49" event={"ID":"2d471d2b-912c-4f49-a7ec-c07b85c02bd9","Type":"ContainerStarted","Data":"a0ec092b50bd992d88812a55d974053d7ddac95be457a288dd049fe5a43b0f2d"} Oct 02 02:02:44 crc kubenswrapper[4885]: I1002 02:02:44.116464 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-9wd49" podStartSLOduration=2.347079713 podStartE2EDuration="13.11643931s" podCreationTimestamp="2025-10-02 02:02:31 +0000 UTC" firstStartedPulling="2025-10-02 02:02:32.034518516 +0000 UTC m=+940.846265935" lastFinishedPulling="2025-10-02 02:02:42.803878093 +0000 UTC m=+951.615625532" observedRunningTime="2025-10-02 02:02:44.107503639 +0000 UTC m=+952.919251048" watchObservedRunningTime="2025-10-02 02:02:44.11643931 +0000 UTC m=+952.928186749" Oct 02 02:02:45 crc kubenswrapper[4885]: I1002 02:02:45.071428 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Oct 02 02:02:45 crc kubenswrapper[4885]: I1002 02:02:45.093517 4885 generic.go:334] "Generic (PLEG): container finished" podID="7b8634a8-6199-42bf-b18d-592a8ba3d164" containerID="e8887f0cf1208a83481402ef4a9f149d8af949103cfd35b53383a946e2763dd8" exitCode=0 Oct 02 02:02:45 crc kubenswrapper[4885]: I1002 02:02:45.093682 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-dnlx6" event={"ID":"7b8634a8-6199-42bf-b18d-592a8ba3d164","Type":"ContainerDied","Data":"e8887f0cf1208a83481402ef4a9f149d8af949103cfd35b53383a946e2763dd8"} Oct 02 02:02:45 crc kubenswrapper[4885]: I1002 02:02:45.152786 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Oct 02 02:02:45 crc kubenswrapper[4885]: I1002 02:02:45.153089 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Oct 02 02:02:45 crc kubenswrapper[4885]: I1002 02:02:45.570648 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-7dl92" Oct 02 02:02:45 crc kubenswrapper[4885]: I1002 02:02:45.606162 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vsjvk\" (UniqueName: \"kubernetes.io/projected/da80ee88-0786-4a7d-9c62-446412d63021-kube-api-access-vsjvk\") pod \"da80ee88-0786-4a7d-9c62-446412d63021\" (UID: \"da80ee88-0786-4a7d-9c62-446412d63021\") " Oct 02 02:02:45 crc kubenswrapper[4885]: I1002 02:02:45.612151 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da80ee88-0786-4a7d-9c62-446412d63021-kube-api-access-vsjvk" (OuterVolumeSpecName: "kube-api-access-vsjvk") pod "da80ee88-0786-4a7d-9c62-446412d63021" (UID: "da80ee88-0786-4a7d-9c62-446412d63021"). InnerVolumeSpecName "kube-api-access-vsjvk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:02:45 crc kubenswrapper[4885]: I1002 02:02:45.709335 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vsjvk\" (UniqueName: \"kubernetes.io/projected/da80ee88-0786-4a7d-9c62-446412d63021-kube-api-access-vsjvk\") on node \"crc\" DevicePath \"\"" Oct 02 02:02:46 crc kubenswrapper[4885]: I1002 02:02:46.107702 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-7dl92" event={"ID":"da80ee88-0786-4a7d-9c62-446412d63021","Type":"ContainerDied","Data":"b0daa8796672251663d94c655c6e07f91b5df6b746d70b4c4feff916e010f274"} Oct 02 02:02:46 crc kubenswrapper[4885]: I1002 02:02:46.107760 4885 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b0daa8796672251663d94c655c6e07f91b5df6b746d70b4c4feff916e010f274" Oct 02 02:02:46 crc kubenswrapper[4885]: I1002 02:02:46.107940 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-7dl92" Oct 02 02:02:46 crc kubenswrapper[4885]: I1002 02:02:46.509972 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-dnlx6" Oct 02 02:02:46 crc kubenswrapper[4885]: I1002 02:02:46.627504 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bs42l\" (UniqueName: \"kubernetes.io/projected/7b8634a8-6199-42bf-b18d-592a8ba3d164-kube-api-access-bs42l\") pod \"7b8634a8-6199-42bf-b18d-592a8ba3d164\" (UID: \"7b8634a8-6199-42bf-b18d-592a8ba3d164\") " Oct 02 02:02:46 crc kubenswrapper[4885]: I1002 02:02:46.627988 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/0c8d183a-f48b-4a69-8685-90c834e678ce-etc-swift\") pod \"swift-storage-0\" (UID: \"0c8d183a-f48b-4a69-8685-90c834e678ce\") " pod="openstack/swift-storage-0" Oct 02 02:02:46 crc kubenswrapper[4885]: E1002 02:02:46.628157 4885 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 02 02:02:46 crc kubenswrapper[4885]: E1002 02:02:46.628177 4885 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 02 02:02:46 crc kubenswrapper[4885]: E1002 02:02:46.628225 4885 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/0c8d183a-f48b-4a69-8685-90c834e678ce-etc-swift podName:0c8d183a-f48b-4a69-8685-90c834e678ce nodeName:}" failed. No retries permitted until 2025-10-02 02:03:02.628210096 +0000 UTC m=+971.439957495 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/0c8d183a-f48b-4a69-8685-90c834e678ce-etc-swift") pod "swift-storage-0" (UID: "0c8d183a-f48b-4a69-8685-90c834e678ce") : configmap "swift-ring-files" not found Oct 02 02:02:46 crc kubenswrapper[4885]: I1002 02:02:46.642126 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b8634a8-6199-42bf-b18d-592a8ba3d164-kube-api-access-bs42l" (OuterVolumeSpecName: "kube-api-access-bs42l") pod "7b8634a8-6199-42bf-b18d-592a8ba3d164" (UID: "7b8634a8-6199-42bf-b18d-592a8ba3d164"). InnerVolumeSpecName "kube-api-access-bs42l". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:02:46 crc kubenswrapper[4885]: I1002 02:02:46.729380 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bs42l\" (UniqueName: \"kubernetes.io/projected/7b8634a8-6199-42bf-b18d-592a8ba3d164-kube-api-access-bs42l\") on node \"crc\" DevicePath \"\"" Oct 02 02:02:47 crc kubenswrapper[4885]: I1002 02:02:47.125732 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-dnlx6" event={"ID":"7b8634a8-6199-42bf-b18d-592a8ba3d164","Type":"ContainerDied","Data":"d1aaedb0f9b1b5cee5f93363ae4b77247f6d4d3473489c1ec2c926956bd5951d"} Oct 02 02:02:47 crc kubenswrapper[4885]: I1002 02:02:47.125795 4885 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d1aaedb0f9b1b5cee5f93363ae4b77247f6d4d3473489c1ec2c926956bd5951d" Oct 02 02:02:47 crc kubenswrapper[4885]: I1002 02:02:47.125811 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-dnlx6" Oct 02 02:02:47 crc kubenswrapper[4885]: I1002 02:02:47.156006 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Oct 02 02:02:47 crc kubenswrapper[4885]: I1002 02:02:47.315892 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Oct 02 02:02:47 crc kubenswrapper[4885]: E1002 02:02:47.316176 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da80ee88-0786-4a7d-9c62-446412d63021" containerName="mariadb-database-create" Oct 02 02:02:47 crc kubenswrapper[4885]: I1002 02:02:47.316193 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="da80ee88-0786-4a7d-9c62-446412d63021" containerName="mariadb-database-create" Oct 02 02:02:47 crc kubenswrapper[4885]: E1002 02:02:47.316211 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b8634a8-6199-42bf-b18d-592a8ba3d164" containerName="mariadb-database-create" Oct 02 02:02:47 crc kubenswrapper[4885]: I1002 02:02:47.316218 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b8634a8-6199-42bf-b18d-592a8ba3d164" containerName="mariadb-database-create" Oct 02 02:02:47 crc kubenswrapper[4885]: I1002 02:02:47.316432 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b8634a8-6199-42bf-b18d-592a8ba3d164" containerName="mariadb-database-create" Oct 02 02:02:47 crc kubenswrapper[4885]: I1002 02:02:47.316445 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="da80ee88-0786-4a7d-9c62-446412d63021" containerName="mariadb-database-create" Oct 02 02:02:47 crc kubenswrapper[4885]: I1002 02:02:47.317174 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 02 02:02:47 crc kubenswrapper[4885]: I1002 02:02:47.319027 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-6mfqm" Oct 02 02:02:47 crc kubenswrapper[4885]: I1002 02:02:47.319407 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Oct 02 02:02:47 crc kubenswrapper[4885]: I1002 02:02:47.319417 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Oct 02 02:02:47 crc kubenswrapper[4885]: I1002 02:02:47.319541 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Oct 02 02:02:47 crc kubenswrapper[4885]: I1002 02:02:47.339567 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n5ckr\" (UniqueName: \"kubernetes.io/projected/d02dcd41-585a-411f-ace9-97c1a662fdc8-kube-api-access-n5ckr\") pod \"ovn-northd-0\" (UID: \"d02dcd41-585a-411f-ace9-97c1a662fdc8\") " pod="openstack/ovn-northd-0" Oct 02 02:02:47 crc kubenswrapper[4885]: I1002 02:02:47.339616 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/d02dcd41-585a-411f-ace9-97c1a662fdc8-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"d02dcd41-585a-411f-ace9-97c1a662fdc8\") " pod="openstack/ovn-northd-0" Oct 02 02:02:47 crc kubenswrapper[4885]: I1002 02:02:47.339669 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d02dcd41-585a-411f-ace9-97c1a662fdc8-scripts\") pod \"ovn-northd-0\" (UID: \"d02dcd41-585a-411f-ace9-97c1a662fdc8\") " pod="openstack/ovn-northd-0" Oct 02 02:02:47 crc kubenswrapper[4885]: I1002 02:02:47.339732 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d02dcd41-585a-411f-ace9-97c1a662fdc8-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"d02dcd41-585a-411f-ace9-97c1a662fdc8\") " pod="openstack/ovn-northd-0" Oct 02 02:02:47 crc kubenswrapper[4885]: I1002 02:02:47.339750 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d02dcd41-585a-411f-ace9-97c1a662fdc8-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"d02dcd41-585a-411f-ace9-97c1a662fdc8\") " pod="openstack/ovn-northd-0" Oct 02 02:02:47 crc kubenswrapper[4885]: I1002 02:02:47.339771 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d02dcd41-585a-411f-ace9-97c1a662fdc8-config\") pod \"ovn-northd-0\" (UID: \"d02dcd41-585a-411f-ace9-97c1a662fdc8\") " pod="openstack/ovn-northd-0" Oct 02 02:02:47 crc kubenswrapper[4885]: I1002 02:02:47.339786 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d02dcd41-585a-411f-ace9-97c1a662fdc8-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"d02dcd41-585a-411f-ace9-97c1a662fdc8\") " pod="openstack/ovn-northd-0" Oct 02 02:02:47 crc kubenswrapper[4885]: I1002 02:02:47.374333 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-jt45g"] Oct 02 02:02:47 crc kubenswrapper[4885]: I1002 02:02:47.375892 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-jt45g" Oct 02 02:02:47 crc kubenswrapper[4885]: I1002 02:02:47.382381 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 02 02:02:47 crc kubenswrapper[4885]: I1002 02:02:47.388064 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-jt45g"] Oct 02 02:02:47 crc kubenswrapper[4885]: I1002 02:02:47.441304 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d02dcd41-585a-411f-ace9-97c1a662fdc8-scripts\") pod \"ovn-northd-0\" (UID: \"d02dcd41-585a-411f-ace9-97c1a662fdc8\") " pod="openstack/ovn-northd-0" Oct 02 02:02:47 crc kubenswrapper[4885]: I1002 02:02:47.441423 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4rfdh\" (UniqueName: \"kubernetes.io/projected/19989c97-8f7e-4141-bbb3-b3ab9aa68d5d-kube-api-access-4rfdh\") pod \"keystone-db-create-jt45g\" (UID: \"19989c97-8f7e-4141-bbb3-b3ab9aa68d5d\") " pod="openstack/keystone-db-create-jt45g" Oct 02 02:02:47 crc kubenswrapper[4885]: I1002 02:02:47.441572 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d02dcd41-585a-411f-ace9-97c1a662fdc8-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"d02dcd41-585a-411f-ace9-97c1a662fdc8\") " pod="openstack/ovn-northd-0" Oct 02 02:02:47 crc kubenswrapper[4885]: I1002 02:02:47.441624 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d02dcd41-585a-411f-ace9-97c1a662fdc8-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"d02dcd41-585a-411f-ace9-97c1a662fdc8\") " pod="openstack/ovn-northd-0" Oct 02 02:02:47 crc kubenswrapper[4885]: I1002 02:02:47.441659 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d02dcd41-585a-411f-ace9-97c1a662fdc8-config\") pod \"ovn-northd-0\" (UID: \"d02dcd41-585a-411f-ace9-97c1a662fdc8\") " pod="openstack/ovn-northd-0" Oct 02 02:02:47 crc kubenswrapper[4885]: I1002 02:02:47.441702 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d02dcd41-585a-411f-ace9-97c1a662fdc8-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"d02dcd41-585a-411f-ace9-97c1a662fdc8\") " pod="openstack/ovn-northd-0" Oct 02 02:02:47 crc kubenswrapper[4885]: I1002 02:02:47.442135 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d02dcd41-585a-411f-ace9-97c1a662fdc8-scripts\") pod \"ovn-northd-0\" (UID: \"d02dcd41-585a-411f-ace9-97c1a662fdc8\") " pod="openstack/ovn-northd-0" Oct 02 02:02:47 crc kubenswrapper[4885]: I1002 02:02:47.442318 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d02dcd41-585a-411f-ace9-97c1a662fdc8-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"d02dcd41-585a-411f-ace9-97c1a662fdc8\") " pod="openstack/ovn-northd-0" Oct 02 02:02:47 crc kubenswrapper[4885]: I1002 02:02:47.442416 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n5ckr\" (UniqueName: \"kubernetes.io/projected/d02dcd41-585a-411f-ace9-97c1a662fdc8-kube-api-access-n5ckr\") pod \"ovn-northd-0\" (UID: \"d02dcd41-585a-411f-ace9-97c1a662fdc8\") " pod="openstack/ovn-northd-0" Oct 02 02:02:47 crc kubenswrapper[4885]: I1002 02:02:47.442782 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/d02dcd41-585a-411f-ace9-97c1a662fdc8-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"d02dcd41-585a-411f-ace9-97c1a662fdc8\") " pod="openstack/ovn-northd-0" Oct 02 02:02:47 crc kubenswrapper[4885]: I1002 02:02:47.442486 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d02dcd41-585a-411f-ace9-97c1a662fdc8-config\") pod \"ovn-northd-0\" (UID: \"d02dcd41-585a-411f-ace9-97c1a662fdc8\") " pod="openstack/ovn-northd-0" Oct 02 02:02:47 crc kubenswrapper[4885]: I1002 02:02:47.448005 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d02dcd41-585a-411f-ace9-97c1a662fdc8-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"d02dcd41-585a-411f-ace9-97c1a662fdc8\") " pod="openstack/ovn-northd-0" Oct 02 02:02:47 crc kubenswrapper[4885]: I1002 02:02:47.448859 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/d02dcd41-585a-411f-ace9-97c1a662fdc8-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"d02dcd41-585a-411f-ace9-97c1a662fdc8\") " pod="openstack/ovn-northd-0" Oct 02 02:02:47 crc kubenswrapper[4885]: I1002 02:02:47.449975 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d02dcd41-585a-411f-ace9-97c1a662fdc8-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"d02dcd41-585a-411f-ace9-97c1a662fdc8\") " pod="openstack/ovn-northd-0" Oct 02 02:02:47 crc kubenswrapper[4885]: I1002 02:02:47.459086 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n5ckr\" (UniqueName: \"kubernetes.io/projected/d02dcd41-585a-411f-ace9-97c1a662fdc8-kube-api-access-n5ckr\") pod \"ovn-northd-0\" (UID: \"d02dcd41-585a-411f-ace9-97c1a662fdc8\") " pod="openstack/ovn-northd-0" Oct 02 02:02:47 crc kubenswrapper[4885]: I1002 02:02:47.530323 4885 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-fgvpt" podUID="924fb321-4073-449a-a546-811b046ed26a" containerName="ovn-controller" probeResult="failure" output=< Oct 02 02:02:47 crc kubenswrapper[4885]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Oct 02 02:02:47 crc kubenswrapper[4885]: > Oct 02 02:02:47 crc kubenswrapper[4885]: I1002 02:02:47.545386 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4rfdh\" (UniqueName: \"kubernetes.io/projected/19989c97-8f7e-4141-bbb3-b3ab9aa68d5d-kube-api-access-4rfdh\") pod \"keystone-db-create-jt45g\" (UID: \"19989c97-8f7e-4141-bbb3-b3ab9aa68d5d\") " pod="openstack/keystone-db-create-jt45g" Oct 02 02:02:47 crc kubenswrapper[4885]: I1002 02:02:47.565217 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4rfdh\" (UniqueName: \"kubernetes.io/projected/19989c97-8f7e-4141-bbb3-b3ab9aa68d5d-kube-api-access-4rfdh\") pod \"keystone-db-create-jt45g\" (UID: \"19989c97-8f7e-4141-bbb3-b3ab9aa68d5d\") " pod="openstack/keystone-db-create-jt45g" Oct 02 02:02:47 crc kubenswrapper[4885]: I1002 02:02:47.640952 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 02 02:02:47 crc kubenswrapper[4885]: I1002 02:02:47.711375 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-jt45g" Oct 02 02:02:48 crc kubenswrapper[4885]: I1002 02:02:48.074668 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 02 02:02:48 crc kubenswrapper[4885]: W1002 02:02:48.083565 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd02dcd41_585a_411f_ace9_97c1a662fdc8.slice/crio-30a2acda4d61594d1e318902a614de7d15d8fac56c6798922833d3590da176cc WatchSource:0}: Error finding container 30a2acda4d61594d1e318902a614de7d15d8fac56c6798922833d3590da176cc: Status 404 returned error can't find the container with id 30a2acda4d61594d1e318902a614de7d15d8fac56c6798922833d3590da176cc Oct 02 02:02:48 crc kubenswrapper[4885]: I1002 02:02:48.134840 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"d02dcd41-585a-411f-ace9-97c1a662fdc8","Type":"ContainerStarted","Data":"30a2acda4d61594d1e318902a614de7d15d8fac56c6798922833d3590da176cc"} Oct 02 02:02:48 crc kubenswrapper[4885]: I1002 02:02:48.242233 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-jt45g"] Oct 02 02:02:49 crc kubenswrapper[4885]: I1002 02:02:49.150379 4885 generic.go:334] "Generic (PLEG): container finished" podID="19989c97-8f7e-4141-bbb3-b3ab9aa68d5d" containerID="05489a5bf3fb1c22237f40e797f5260aa4732bffe7ad2335ce7c1139122fa939" exitCode=0 Oct 02 02:02:49 crc kubenswrapper[4885]: I1002 02:02:49.150624 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-jt45g" event={"ID":"19989c97-8f7e-4141-bbb3-b3ab9aa68d5d","Type":"ContainerDied","Data":"05489a5bf3fb1c22237f40e797f5260aa4732bffe7ad2335ce7c1139122fa939"} Oct 02 02:02:49 crc kubenswrapper[4885]: I1002 02:02:49.150735 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-jt45g" event={"ID":"19989c97-8f7e-4141-bbb3-b3ab9aa68d5d","Type":"ContainerStarted","Data":"22a91eab17c19cfd0e9b05b60cec9a037aa8be6b603f3c368e2a832df510b4cb"} Oct 02 02:02:50 crc kubenswrapper[4885]: I1002 02:02:50.464662 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-jt45g" Oct 02 02:02:50 crc kubenswrapper[4885]: I1002 02:02:50.520988 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4rfdh\" (UniqueName: \"kubernetes.io/projected/19989c97-8f7e-4141-bbb3-b3ab9aa68d5d-kube-api-access-4rfdh\") pod \"19989c97-8f7e-4141-bbb3-b3ab9aa68d5d\" (UID: \"19989c97-8f7e-4141-bbb3-b3ab9aa68d5d\") " Oct 02 02:02:50 crc kubenswrapper[4885]: I1002 02:02:50.525805 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19989c97-8f7e-4141-bbb3-b3ab9aa68d5d-kube-api-access-4rfdh" (OuterVolumeSpecName: "kube-api-access-4rfdh") pod "19989c97-8f7e-4141-bbb3-b3ab9aa68d5d" (UID: "19989c97-8f7e-4141-bbb3-b3ab9aa68d5d"). InnerVolumeSpecName "kube-api-access-4rfdh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:02:50 crc kubenswrapper[4885]: I1002 02:02:50.623631 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4rfdh\" (UniqueName: \"kubernetes.io/projected/19989c97-8f7e-4141-bbb3-b3ab9aa68d5d-kube-api-access-4rfdh\") on node \"crc\" DevicePath \"\"" Oct 02 02:02:51 crc kubenswrapper[4885]: I1002 02:02:51.171035 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-jt45g" event={"ID":"19989c97-8f7e-4141-bbb3-b3ab9aa68d5d","Type":"ContainerDied","Data":"22a91eab17c19cfd0e9b05b60cec9a037aa8be6b603f3c368e2a832df510b4cb"} Oct 02 02:02:51 crc kubenswrapper[4885]: I1002 02:02:51.171450 4885 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="22a91eab17c19cfd0e9b05b60cec9a037aa8be6b603f3c368e2a832df510b4cb" Oct 02 02:02:51 crc kubenswrapper[4885]: I1002 02:02:51.171133 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-jt45g" Oct 02 02:02:51 crc kubenswrapper[4885]: I1002 02:02:51.175191 4885 generic.go:334] "Generic (PLEG): container finished" podID="6692f52c-5ee9-4cd6-8674-0fb3f5fd6679" containerID="a52ec36756e9b2deb9a3362cfb43912af369dd3918986e7fd40070a1385f68f0" exitCode=0 Oct 02 02:02:51 crc kubenswrapper[4885]: I1002 02:02:51.175400 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"6692f52c-5ee9-4cd6-8674-0fb3f5fd6679","Type":"ContainerDied","Data":"a52ec36756e9b2deb9a3362cfb43912af369dd3918986e7fd40070a1385f68f0"} Oct 02 02:02:51 crc kubenswrapper[4885]: I1002 02:02:51.183716 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"d02dcd41-585a-411f-ace9-97c1a662fdc8","Type":"ContainerStarted","Data":"4ed1fbfcd9ee0451437e9e4b712156a979cf9abf33dd760ffd514c0b23a2f982"} Oct 02 02:02:51 crc kubenswrapper[4885]: I1002 02:02:51.183811 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"d02dcd41-585a-411f-ace9-97c1a662fdc8","Type":"ContainerStarted","Data":"64040402a2269f32b4e4fd1b32c2070d8fdd0baac87e060c6d981ab6565b337f"} Oct 02 02:02:51 crc kubenswrapper[4885]: I1002 02:02:51.184079 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Oct 02 02:02:51 crc kubenswrapper[4885]: I1002 02:02:51.189027 4885 generic.go:334] "Generic (PLEG): container finished" podID="2d471d2b-912c-4f49-a7ec-c07b85c02bd9" containerID="a0ec092b50bd992d88812a55d974053d7ddac95be457a288dd049fe5a43b0f2d" exitCode=0 Oct 02 02:02:51 crc kubenswrapper[4885]: I1002 02:02:51.189115 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-9wd49" event={"ID":"2d471d2b-912c-4f49-a7ec-c07b85c02bd9","Type":"ContainerDied","Data":"a0ec092b50bd992d88812a55d974053d7ddac95be457a288dd049fe5a43b0f2d"} Oct 02 02:02:51 crc kubenswrapper[4885]: I1002 02:02:51.193383 4885 generic.go:334] "Generic (PLEG): container finished" podID="eea66006-67ba-45e3-bdf9-9d144d772386" containerID="6f3eaeb38ec37be6825c9165f2dca00ced63746de7714d811aaa270ddf582da9" exitCode=0 Oct 02 02:02:51 crc kubenswrapper[4885]: I1002 02:02:51.193464 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"eea66006-67ba-45e3-bdf9-9d144d772386","Type":"ContainerDied","Data":"6f3eaeb38ec37be6825c9165f2dca00ced63746de7714d811aaa270ddf582da9"} Oct 02 02:02:51 crc kubenswrapper[4885]: I1002 02:02:51.257853 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.473640493 podStartE2EDuration="4.25782099s" podCreationTimestamp="2025-10-02 02:02:47 +0000 UTC" firstStartedPulling="2025-10-02 02:02:48.08791987 +0000 UTC m=+956.899667309" lastFinishedPulling="2025-10-02 02:02:49.872100377 +0000 UTC m=+958.683847806" observedRunningTime="2025-10-02 02:02:51.253984344 +0000 UTC m=+960.065731783" watchObservedRunningTime="2025-10-02 02:02:51.25782099 +0000 UTC m=+960.069568419" Oct 02 02:02:52 crc kubenswrapper[4885]: I1002 02:02:52.204378 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"6692f52c-5ee9-4cd6-8674-0fb3f5fd6679","Type":"ContainerStarted","Data":"18584fe73de75abdffba2d36424556d76845cfba20bc3152c3df90230f533362"} Oct 02 02:02:52 crc kubenswrapper[4885]: I1002 02:02:52.205213 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 02 02:02:52 crc kubenswrapper[4885]: I1002 02:02:52.207530 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"eea66006-67ba-45e3-bdf9-9d144d772386","Type":"ContainerStarted","Data":"6d97c5a8d103c2c2b9ce70a5d8c6ef3e51104fe5a791fe6d1e52bdf3ad332dff"} Oct 02 02:02:52 crc kubenswrapper[4885]: I1002 02:02:52.255355 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=60.522883573 podStartE2EDuration="1m9.255240489s" podCreationTimestamp="2025-10-02 02:01:43 +0000 UTC" firstStartedPulling="2025-10-02 02:01:56.904623416 +0000 UTC m=+905.716370815" lastFinishedPulling="2025-10-02 02:02:05.636980332 +0000 UTC m=+914.448727731" observedRunningTime="2025-10-02 02:02:52.243631226 +0000 UTC m=+961.055378665" watchObservedRunningTime="2025-10-02 02:02:52.255240489 +0000 UTC m=+961.066987888" Oct 02 02:02:52 crc kubenswrapper[4885]: I1002 02:02:52.274783 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=51.218827423 podStartE2EDuration="1m10.274757093s" podCreationTimestamp="2025-10-02 02:01:42 +0000 UTC" firstStartedPulling="2025-10-02 02:01:57.334293869 +0000 UTC m=+906.146041268" lastFinishedPulling="2025-10-02 02:02:16.390223499 +0000 UTC m=+925.201970938" observedRunningTime="2025-10-02 02:02:52.271161804 +0000 UTC m=+961.082909223" watchObservedRunningTime="2025-10-02 02:02:52.274757093 +0000 UTC m=+961.086504522" Oct 02 02:02:52 crc kubenswrapper[4885]: I1002 02:02:52.530330 4885 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-fgvpt" podUID="924fb321-4073-449a-a546-811b046ed26a" containerName="ovn-controller" probeResult="failure" output=< Oct 02 02:02:52 crc kubenswrapper[4885]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Oct 02 02:02:52 crc kubenswrapper[4885]: > Oct 02 02:02:52 crc kubenswrapper[4885]: I1002 02:02:52.572036 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-q92rb" Oct 02 02:02:52 crc kubenswrapper[4885]: I1002 02:02:52.580278 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-q92rb" Oct 02 02:02:52 crc kubenswrapper[4885]: I1002 02:02:52.616472 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-9wd49" Oct 02 02:02:52 crc kubenswrapper[4885]: I1002 02:02:52.762905 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/2d471d2b-912c-4f49-a7ec-c07b85c02bd9-etc-swift\") pod \"2d471d2b-912c-4f49-a7ec-c07b85c02bd9\" (UID: \"2d471d2b-912c-4f49-a7ec-c07b85c02bd9\") " Oct 02 02:02:52 crc kubenswrapper[4885]: I1002 02:02:52.763230 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/2d471d2b-912c-4f49-a7ec-c07b85c02bd9-dispersionconf\") pod \"2d471d2b-912c-4f49-a7ec-c07b85c02bd9\" (UID: \"2d471d2b-912c-4f49-a7ec-c07b85c02bd9\") " Oct 02 02:02:52 crc kubenswrapper[4885]: I1002 02:02:52.763310 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/2d471d2b-912c-4f49-a7ec-c07b85c02bd9-ring-data-devices\") pod \"2d471d2b-912c-4f49-a7ec-c07b85c02bd9\" (UID: \"2d471d2b-912c-4f49-a7ec-c07b85c02bd9\") " Oct 02 02:02:52 crc kubenswrapper[4885]: I1002 02:02:52.763355 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/2d471d2b-912c-4f49-a7ec-c07b85c02bd9-swiftconf\") pod \"2d471d2b-912c-4f49-a7ec-c07b85c02bd9\" (UID: \"2d471d2b-912c-4f49-a7ec-c07b85c02bd9\") " Oct 02 02:02:52 crc kubenswrapper[4885]: I1002 02:02:52.763406 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2d471d2b-912c-4f49-a7ec-c07b85c02bd9-scripts\") pod \"2d471d2b-912c-4f49-a7ec-c07b85c02bd9\" (UID: \"2d471d2b-912c-4f49-a7ec-c07b85c02bd9\") " Oct 02 02:02:52 crc kubenswrapper[4885]: I1002 02:02:52.763431 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qc2q7\" (UniqueName: \"kubernetes.io/projected/2d471d2b-912c-4f49-a7ec-c07b85c02bd9-kube-api-access-qc2q7\") pod \"2d471d2b-912c-4f49-a7ec-c07b85c02bd9\" (UID: \"2d471d2b-912c-4f49-a7ec-c07b85c02bd9\") " Oct 02 02:02:52 crc kubenswrapper[4885]: I1002 02:02:52.763468 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d471d2b-912c-4f49-a7ec-c07b85c02bd9-combined-ca-bundle\") pod \"2d471d2b-912c-4f49-a7ec-c07b85c02bd9\" (UID: \"2d471d2b-912c-4f49-a7ec-c07b85c02bd9\") " Oct 02 02:02:52 crc kubenswrapper[4885]: I1002 02:02:52.763795 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2d471d2b-912c-4f49-a7ec-c07b85c02bd9-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "2d471d2b-912c-4f49-a7ec-c07b85c02bd9" (UID: "2d471d2b-912c-4f49-a7ec-c07b85c02bd9"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:02:52 crc kubenswrapper[4885]: I1002 02:02:52.764174 4885 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/2d471d2b-912c-4f49-a7ec-c07b85c02bd9-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 02 02:02:52 crc kubenswrapper[4885]: I1002 02:02:52.764731 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d471d2b-912c-4f49-a7ec-c07b85c02bd9-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "2d471d2b-912c-4f49-a7ec-c07b85c02bd9" (UID: "2d471d2b-912c-4f49-a7ec-c07b85c02bd9"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:02:52 crc kubenswrapper[4885]: I1002 02:02:52.770364 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d471d2b-912c-4f49-a7ec-c07b85c02bd9-kube-api-access-qc2q7" (OuterVolumeSpecName: "kube-api-access-qc2q7") pod "2d471d2b-912c-4f49-a7ec-c07b85c02bd9" (UID: "2d471d2b-912c-4f49-a7ec-c07b85c02bd9"). InnerVolumeSpecName "kube-api-access-qc2q7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:02:52 crc kubenswrapper[4885]: I1002 02:02:52.773539 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d471d2b-912c-4f49-a7ec-c07b85c02bd9-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "2d471d2b-912c-4f49-a7ec-c07b85c02bd9" (UID: "2d471d2b-912c-4f49-a7ec-c07b85c02bd9"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:02:52 crc kubenswrapper[4885]: I1002 02:02:52.826165 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-fgvpt-config-zhrsw"] Oct 02 02:02:52 crc kubenswrapper[4885]: E1002 02:02:52.826549 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d471d2b-912c-4f49-a7ec-c07b85c02bd9" containerName="swift-ring-rebalance" Oct 02 02:02:52 crc kubenswrapper[4885]: I1002 02:02:52.826568 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d471d2b-912c-4f49-a7ec-c07b85c02bd9" containerName="swift-ring-rebalance" Oct 02 02:02:52 crc kubenswrapper[4885]: E1002 02:02:52.826591 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19989c97-8f7e-4141-bbb3-b3ab9aa68d5d" containerName="mariadb-database-create" Oct 02 02:02:52 crc kubenswrapper[4885]: I1002 02:02:52.826597 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="19989c97-8f7e-4141-bbb3-b3ab9aa68d5d" containerName="mariadb-database-create" Oct 02 02:02:52 crc kubenswrapper[4885]: I1002 02:02:52.826747 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d471d2b-912c-4f49-a7ec-c07b85c02bd9" containerName="swift-ring-rebalance" Oct 02 02:02:52 crc kubenswrapper[4885]: I1002 02:02:52.826756 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="19989c97-8f7e-4141-bbb3-b3ab9aa68d5d" containerName="mariadb-database-create" Oct 02 02:02:52 crc kubenswrapper[4885]: I1002 02:02:52.827228 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-fgvpt-config-zhrsw" Oct 02 02:02:52 crc kubenswrapper[4885]: I1002 02:02:52.827875 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-fgvpt-config-zhrsw"] Oct 02 02:02:52 crc kubenswrapper[4885]: I1002 02:02:52.830050 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Oct 02 02:02:52 crc kubenswrapper[4885]: I1002 02:02:52.863300 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d471d2b-912c-4f49-a7ec-c07b85c02bd9-scripts" (OuterVolumeSpecName: "scripts") pod "2d471d2b-912c-4f49-a7ec-c07b85c02bd9" (UID: "2d471d2b-912c-4f49-a7ec-c07b85c02bd9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:02:52 crc kubenswrapper[4885]: I1002 02:02:52.863562 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d471d2b-912c-4f49-a7ec-c07b85c02bd9-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "2d471d2b-912c-4f49-a7ec-c07b85c02bd9" (UID: "2d471d2b-912c-4f49-a7ec-c07b85c02bd9"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:02:52 crc kubenswrapper[4885]: I1002 02:02:52.865350 4885 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/2d471d2b-912c-4f49-a7ec-c07b85c02bd9-ring-data-devices\") on node \"crc\" DevicePath \"\"" Oct 02 02:02:52 crc kubenswrapper[4885]: I1002 02:02:52.865369 4885 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/2d471d2b-912c-4f49-a7ec-c07b85c02bd9-swiftconf\") on node \"crc\" DevicePath \"\"" Oct 02 02:02:52 crc kubenswrapper[4885]: I1002 02:02:52.865377 4885 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2d471d2b-912c-4f49-a7ec-c07b85c02bd9-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 02:02:52 crc kubenswrapper[4885]: I1002 02:02:52.865387 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qc2q7\" (UniqueName: \"kubernetes.io/projected/2d471d2b-912c-4f49-a7ec-c07b85c02bd9-kube-api-access-qc2q7\") on node \"crc\" DevicePath \"\"" Oct 02 02:02:52 crc kubenswrapper[4885]: I1002 02:02:52.865398 4885 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/2d471d2b-912c-4f49-a7ec-c07b85c02bd9-dispersionconf\") on node \"crc\" DevicePath \"\"" Oct 02 02:02:52 crc kubenswrapper[4885]: I1002 02:02:52.866993 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d471d2b-912c-4f49-a7ec-c07b85c02bd9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2d471d2b-912c-4f49-a7ec-c07b85c02bd9" (UID: "2d471d2b-912c-4f49-a7ec-c07b85c02bd9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:02:52 crc kubenswrapper[4885]: I1002 02:02:52.966376 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-28sfm\" (UniqueName: \"kubernetes.io/projected/895799b2-0358-4bd5-9ead-95fb63e00207-kube-api-access-28sfm\") pod \"ovn-controller-fgvpt-config-zhrsw\" (UID: \"895799b2-0358-4bd5-9ead-95fb63e00207\") " pod="openstack/ovn-controller-fgvpt-config-zhrsw" Oct 02 02:02:52 crc kubenswrapper[4885]: I1002 02:02:52.966478 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/895799b2-0358-4bd5-9ead-95fb63e00207-additional-scripts\") pod \"ovn-controller-fgvpt-config-zhrsw\" (UID: \"895799b2-0358-4bd5-9ead-95fb63e00207\") " pod="openstack/ovn-controller-fgvpt-config-zhrsw" Oct 02 02:02:52 crc kubenswrapper[4885]: I1002 02:02:52.966532 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/895799b2-0358-4bd5-9ead-95fb63e00207-scripts\") pod \"ovn-controller-fgvpt-config-zhrsw\" (UID: \"895799b2-0358-4bd5-9ead-95fb63e00207\") " pod="openstack/ovn-controller-fgvpt-config-zhrsw" Oct 02 02:02:52 crc kubenswrapper[4885]: I1002 02:02:52.966561 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/895799b2-0358-4bd5-9ead-95fb63e00207-var-run-ovn\") pod \"ovn-controller-fgvpt-config-zhrsw\" (UID: \"895799b2-0358-4bd5-9ead-95fb63e00207\") " pod="openstack/ovn-controller-fgvpt-config-zhrsw" Oct 02 02:02:52 crc kubenswrapper[4885]: I1002 02:02:52.966789 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/895799b2-0358-4bd5-9ead-95fb63e00207-var-log-ovn\") pod \"ovn-controller-fgvpt-config-zhrsw\" (UID: \"895799b2-0358-4bd5-9ead-95fb63e00207\") " pod="openstack/ovn-controller-fgvpt-config-zhrsw" Oct 02 02:02:52 crc kubenswrapper[4885]: I1002 02:02:52.967060 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/895799b2-0358-4bd5-9ead-95fb63e00207-var-run\") pod \"ovn-controller-fgvpt-config-zhrsw\" (UID: \"895799b2-0358-4bd5-9ead-95fb63e00207\") " pod="openstack/ovn-controller-fgvpt-config-zhrsw" Oct 02 02:02:52 crc kubenswrapper[4885]: I1002 02:02:52.967160 4885 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d471d2b-912c-4f49-a7ec-c07b85c02bd9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:02:53 crc kubenswrapper[4885]: I1002 02:02:53.068690 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/895799b2-0358-4bd5-9ead-95fb63e00207-var-run\") pod \"ovn-controller-fgvpt-config-zhrsw\" (UID: \"895799b2-0358-4bd5-9ead-95fb63e00207\") " pod="openstack/ovn-controller-fgvpt-config-zhrsw" Oct 02 02:02:53 crc kubenswrapper[4885]: I1002 02:02:53.068727 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/895799b2-0358-4bd5-9ead-95fb63e00207-var-run\") pod \"ovn-controller-fgvpt-config-zhrsw\" (UID: \"895799b2-0358-4bd5-9ead-95fb63e00207\") " pod="openstack/ovn-controller-fgvpt-config-zhrsw" Oct 02 02:02:53 crc kubenswrapper[4885]: I1002 02:02:53.068786 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-28sfm\" (UniqueName: \"kubernetes.io/projected/895799b2-0358-4bd5-9ead-95fb63e00207-kube-api-access-28sfm\") pod \"ovn-controller-fgvpt-config-zhrsw\" (UID: \"895799b2-0358-4bd5-9ead-95fb63e00207\") " pod="openstack/ovn-controller-fgvpt-config-zhrsw" Oct 02 02:02:53 crc kubenswrapper[4885]: I1002 02:02:53.068812 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/895799b2-0358-4bd5-9ead-95fb63e00207-additional-scripts\") pod \"ovn-controller-fgvpt-config-zhrsw\" (UID: \"895799b2-0358-4bd5-9ead-95fb63e00207\") " pod="openstack/ovn-controller-fgvpt-config-zhrsw" Oct 02 02:02:53 crc kubenswrapper[4885]: I1002 02:02:53.068834 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/895799b2-0358-4bd5-9ead-95fb63e00207-scripts\") pod \"ovn-controller-fgvpt-config-zhrsw\" (UID: \"895799b2-0358-4bd5-9ead-95fb63e00207\") " pod="openstack/ovn-controller-fgvpt-config-zhrsw" Oct 02 02:02:53 crc kubenswrapper[4885]: I1002 02:02:53.068855 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/895799b2-0358-4bd5-9ead-95fb63e00207-var-run-ovn\") pod \"ovn-controller-fgvpt-config-zhrsw\" (UID: \"895799b2-0358-4bd5-9ead-95fb63e00207\") " pod="openstack/ovn-controller-fgvpt-config-zhrsw" Oct 02 02:02:53 crc kubenswrapper[4885]: I1002 02:02:53.069195 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/895799b2-0358-4bd5-9ead-95fb63e00207-var-run-ovn\") pod \"ovn-controller-fgvpt-config-zhrsw\" (UID: \"895799b2-0358-4bd5-9ead-95fb63e00207\") " pod="openstack/ovn-controller-fgvpt-config-zhrsw" Oct 02 02:02:53 crc kubenswrapper[4885]: I1002 02:02:53.069716 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/895799b2-0358-4bd5-9ead-95fb63e00207-var-log-ovn\") pod \"ovn-controller-fgvpt-config-zhrsw\" (UID: \"895799b2-0358-4bd5-9ead-95fb63e00207\") " pod="openstack/ovn-controller-fgvpt-config-zhrsw" Oct 02 02:02:53 crc kubenswrapper[4885]: I1002 02:02:53.069818 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/895799b2-0358-4bd5-9ead-95fb63e00207-var-log-ovn\") pod \"ovn-controller-fgvpt-config-zhrsw\" (UID: \"895799b2-0358-4bd5-9ead-95fb63e00207\") " pod="openstack/ovn-controller-fgvpt-config-zhrsw" Oct 02 02:02:53 crc kubenswrapper[4885]: I1002 02:02:53.070042 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/895799b2-0358-4bd5-9ead-95fb63e00207-additional-scripts\") pod \"ovn-controller-fgvpt-config-zhrsw\" (UID: \"895799b2-0358-4bd5-9ead-95fb63e00207\") " pod="openstack/ovn-controller-fgvpt-config-zhrsw" Oct 02 02:02:53 crc kubenswrapper[4885]: I1002 02:02:53.072633 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/895799b2-0358-4bd5-9ead-95fb63e00207-scripts\") pod \"ovn-controller-fgvpt-config-zhrsw\" (UID: \"895799b2-0358-4bd5-9ead-95fb63e00207\") " pod="openstack/ovn-controller-fgvpt-config-zhrsw" Oct 02 02:02:53 crc kubenswrapper[4885]: I1002 02:02:53.107287 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-28sfm\" (UniqueName: \"kubernetes.io/projected/895799b2-0358-4bd5-9ead-95fb63e00207-kube-api-access-28sfm\") pod \"ovn-controller-fgvpt-config-zhrsw\" (UID: \"895799b2-0358-4bd5-9ead-95fb63e00207\") " pod="openstack/ovn-controller-fgvpt-config-zhrsw" Oct 02 02:02:53 crc kubenswrapper[4885]: I1002 02:02:53.142319 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-e8ab-account-create-wrrhx"] Oct 02 02:02:53 crc kubenswrapper[4885]: I1002 02:02:53.143881 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-e8ab-account-create-wrrhx" Oct 02 02:02:53 crc kubenswrapper[4885]: I1002 02:02:53.146684 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Oct 02 02:02:53 crc kubenswrapper[4885]: I1002 02:02:53.151941 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-e8ab-account-create-wrrhx"] Oct 02 02:02:53 crc kubenswrapper[4885]: I1002 02:02:53.170629 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-slwgc\" (UniqueName: \"kubernetes.io/projected/352a9474-7bae-4467-9e56-2d320b55494a-kube-api-access-slwgc\") pod \"glance-e8ab-account-create-wrrhx\" (UID: \"352a9474-7bae-4467-9e56-2d320b55494a\") " pod="openstack/glance-e8ab-account-create-wrrhx" Oct 02 02:02:53 crc kubenswrapper[4885]: I1002 02:02:53.194341 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-fgvpt-config-zhrsw" Oct 02 02:02:53 crc kubenswrapper[4885]: I1002 02:02:53.217571 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-9wd49" event={"ID":"2d471d2b-912c-4f49-a7ec-c07b85c02bd9","Type":"ContainerDied","Data":"ec92831404be01280133ce230b8010328797f0a590781887b3c1ff15695ee3db"} Oct 02 02:02:53 crc kubenswrapper[4885]: I1002 02:02:53.217635 4885 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ec92831404be01280133ce230b8010328797f0a590781887b3c1ff15695ee3db" Oct 02 02:02:53 crc kubenswrapper[4885]: I1002 02:02:53.218312 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-9wd49" Oct 02 02:02:53 crc kubenswrapper[4885]: I1002 02:02:53.271639 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-slwgc\" (UniqueName: \"kubernetes.io/projected/352a9474-7bae-4467-9e56-2d320b55494a-kube-api-access-slwgc\") pod \"glance-e8ab-account-create-wrrhx\" (UID: \"352a9474-7bae-4467-9e56-2d320b55494a\") " pod="openstack/glance-e8ab-account-create-wrrhx" Oct 02 02:02:53 crc kubenswrapper[4885]: I1002 02:02:53.297721 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-slwgc\" (UniqueName: \"kubernetes.io/projected/352a9474-7bae-4467-9e56-2d320b55494a-kube-api-access-slwgc\") pod \"glance-e8ab-account-create-wrrhx\" (UID: \"352a9474-7bae-4467-9e56-2d320b55494a\") " pod="openstack/glance-e8ab-account-create-wrrhx" Oct 02 02:02:53 crc kubenswrapper[4885]: I1002 02:02:53.468176 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-e8ab-account-create-wrrhx" Oct 02 02:02:53 crc kubenswrapper[4885]: I1002 02:02:53.682007 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-fgvpt-config-zhrsw"] Oct 02 02:02:53 crc kubenswrapper[4885]: W1002 02:02:53.690548 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod895799b2_0358_4bd5_9ead_95fb63e00207.slice/crio-b6a5130e233c41e076122c51ee5483f80392c532a630a7ea813bf9cafec5e6e6 WatchSource:0}: Error finding container b6a5130e233c41e076122c51ee5483f80392c532a630a7ea813bf9cafec5e6e6: Status 404 returned error can't find the container with id b6a5130e233c41e076122c51ee5483f80392c532a630a7ea813bf9cafec5e6e6 Oct 02 02:02:54 crc kubenswrapper[4885]: I1002 02:02:54.011955 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-e8ab-account-create-wrrhx"] Oct 02 02:02:54 crc kubenswrapper[4885]: W1002 02:02:54.029807 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod352a9474_7bae_4467_9e56_2d320b55494a.slice/crio-534f42670f36150a104ba7d8634c851c4c2a69019f44859820b993958e197355 WatchSource:0}: Error finding container 534f42670f36150a104ba7d8634c851c4c2a69019f44859820b993958e197355: Status 404 returned error can't find the container with id 534f42670f36150a104ba7d8634c851c4c2a69019f44859820b993958e197355 Oct 02 02:02:54 crc kubenswrapper[4885]: I1002 02:02:54.142010 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:02:54 crc kubenswrapper[4885]: I1002 02:02:54.225359 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-e8ab-account-create-wrrhx" event={"ID":"352a9474-7bae-4467-9e56-2d320b55494a","Type":"ContainerStarted","Data":"534f42670f36150a104ba7d8634c851c4c2a69019f44859820b993958e197355"} Oct 02 02:02:54 crc kubenswrapper[4885]: I1002 02:02:54.227244 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-fgvpt-config-zhrsw" event={"ID":"895799b2-0358-4bd5-9ead-95fb63e00207","Type":"ContainerStarted","Data":"b6a5130e233c41e076122c51ee5483f80392c532a630a7ea813bf9cafec5e6e6"} Oct 02 02:02:56 crc kubenswrapper[4885]: I1002 02:02:56.256055 4885 generic.go:334] "Generic (PLEG): container finished" podID="352a9474-7bae-4467-9e56-2d320b55494a" containerID="b1d356a6538df1c7aaf0a73e390a4014364807d6d4f4a03d7384039af90d4972" exitCode=0 Oct 02 02:02:56 crc kubenswrapper[4885]: I1002 02:02:56.256166 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-e8ab-account-create-wrrhx" event={"ID":"352a9474-7bae-4467-9e56-2d320b55494a","Type":"ContainerDied","Data":"b1d356a6538df1c7aaf0a73e390a4014364807d6d4f4a03d7384039af90d4972"} Oct 02 02:02:56 crc kubenswrapper[4885]: I1002 02:02:56.261722 4885 generic.go:334] "Generic (PLEG): container finished" podID="895799b2-0358-4bd5-9ead-95fb63e00207" containerID="e4469a2fed694e34c2059aa68a82bd35bddc66ce6055feb2c64800dabde10945" exitCode=0 Oct 02 02:02:56 crc kubenswrapper[4885]: I1002 02:02:56.261791 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-fgvpt-config-zhrsw" event={"ID":"895799b2-0358-4bd5-9ead-95fb63e00207","Type":"ContainerDied","Data":"e4469a2fed694e34c2059aa68a82bd35bddc66ce6055feb2c64800dabde10945"} Oct 02 02:02:57 crc kubenswrapper[4885]: I1002 02:02:57.521646 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-45be-account-create-824gn"] Oct 02 02:02:57 crc kubenswrapper[4885]: I1002 02:02:57.523578 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-45be-account-create-824gn" Oct 02 02:02:57 crc kubenswrapper[4885]: I1002 02:02:57.527677 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Oct 02 02:02:57 crc kubenswrapper[4885]: I1002 02:02:57.542505 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-45be-account-create-824gn"] Oct 02 02:02:57 crc kubenswrapper[4885]: I1002 02:02:57.568209 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-fgvpt" Oct 02 02:02:57 crc kubenswrapper[4885]: I1002 02:02:57.658100 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9tf97\" (UniqueName: \"kubernetes.io/projected/b88a1087-86c9-41ad-a0b9-48a2607bfc46-kube-api-access-9tf97\") pod \"keystone-45be-account-create-824gn\" (UID: \"b88a1087-86c9-41ad-a0b9-48a2607bfc46\") " pod="openstack/keystone-45be-account-create-824gn" Oct 02 02:02:57 crc kubenswrapper[4885]: I1002 02:02:57.669707 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-fgvpt-config-zhrsw" Oct 02 02:02:57 crc kubenswrapper[4885]: I1002 02:02:57.759175 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9tf97\" (UniqueName: \"kubernetes.io/projected/b88a1087-86c9-41ad-a0b9-48a2607bfc46-kube-api-access-9tf97\") pod \"keystone-45be-account-create-824gn\" (UID: \"b88a1087-86c9-41ad-a0b9-48a2607bfc46\") " pod="openstack/keystone-45be-account-create-824gn" Oct 02 02:02:57 crc kubenswrapper[4885]: I1002 02:02:57.761412 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-e8ab-account-create-wrrhx" Oct 02 02:02:57 crc kubenswrapper[4885]: I1002 02:02:57.799816 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9tf97\" (UniqueName: \"kubernetes.io/projected/b88a1087-86c9-41ad-a0b9-48a2607bfc46-kube-api-access-9tf97\") pod \"keystone-45be-account-create-824gn\" (UID: \"b88a1087-86c9-41ad-a0b9-48a2607bfc46\") " pod="openstack/keystone-45be-account-create-824gn" Oct 02 02:02:57 crc kubenswrapper[4885]: I1002 02:02:57.846677 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-77fb-account-create-jn758"] Oct 02 02:02:57 crc kubenswrapper[4885]: E1002 02:02:57.846987 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="895799b2-0358-4bd5-9ead-95fb63e00207" containerName="ovn-config" Oct 02 02:02:57 crc kubenswrapper[4885]: I1002 02:02:57.846997 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="895799b2-0358-4bd5-9ead-95fb63e00207" containerName="ovn-config" Oct 02 02:02:57 crc kubenswrapper[4885]: E1002 02:02:57.847009 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="352a9474-7bae-4467-9e56-2d320b55494a" containerName="mariadb-account-create" Oct 02 02:02:57 crc kubenswrapper[4885]: I1002 02:02:57.847015 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="352a9474-7bae-4467-9e56-2d320b55494a" containerName="mariadb-account-create" Oct 02 02:02:57 crc kubenswrapper[4885]: I1002 02:02:57.847165 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="895799b2-0358-4bd5-9ead-95fb63e00207" containerName="ovn-config" Oct 02 02:02:57 crc kubenswrapper[4885]: I1002 02:02:57.847187 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="352a9474-7bae-4467-9e56-2d320b55494a" containerName="mariadb-account-create" Oct 02 02:02:57 crc kubenswrapper[4885]: I1002 02:02:57.847755 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-77fb-account-create-jn758" Oct 02 02:02:57 crc kubenswrapper[4885]: I1002 02:02:57.849781 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Oct 02 02:02:57 crc kubenswrapper[4885]: I1002 02:02:57.859662 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-45be-account-create-824gn" Oct 02 02:02:57 crc kubenswrapper[4885]: I1002 02:02:57.859921 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/895799b2-0358-4bd5-9ead-95fb63e00207-var-run\") pod \"895799b2-0358-4bd5-9ead-95fb63e00207\" (UID: \"895799b2-0358-4bd5-9ead-95fb63e00207\") " Oct 02 02:02:57 crc kubenswrapper[4885]: I1002 02:02:57.860027 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/895799b2-0358-4bd5-9ead-95fb63e00207-scripts\") pod \"895799b2-0358-4bd5-9ead-95fb63e00207\" (UID: \"895799b2-0358-4bd5-9ead-95fb63e00207\") " Oct 02 02:02:57 crc kubenswrapper[4885]: I1002 02:02:57.860036 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/895799b2-0358-4bd5-9ead-95fb63e00207-var-run" (OuterVolumeSpecName: "var-run") pod "895799b2-0358-4bd5-9ead-95fb63e00207" (UID: "895799b2-0358-4bd5-9ead-95fb63e00207"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 02:02:57 crc kubenswrapper[4885]: I1002 02:02:57.860070 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-28sfm\" (UniqueName: \"kubernetes.io/projected/895799b2-0358-4bd5-9ead-95fb63e00207-kube-api-access-28sfm\") pod \"895799b2-0358-4bd5-9ead-95fb63e00207\" (UID: \"895799b2-0358-4bd5-9ead-95fb63e00207\") " Oct 02 02:02:57 crc kubenswrapper[4885]: I1002 02:02:57.860110 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/895799b2-0358-4bd5-9ead-95fb63e00207-additional-scripts\") pod \"895799b2-0358-4bd5-9ead-95fb63e00207\" (UID: \"895799b2-0358-4bd5-9ead-95fb63e00207\") " Oct 02 02:02:57 crc kubenswrapper[4885]: I1002 02:02:57.860176 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/895799b2-0358-4bd5-9ead-95fb63e00207-var-log-ovn\") pod \"895799b2-0358-4bd5-9ead-95fb63e00207\" (UID: \"895799b2-0358-4bd5-9ead-95fb63e00207\") " Oct 02 02:02:57 crc kubenswrapper[4885]: I1002 02:02:57.860289 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/895799b2-0358-4bd5-9ead-95fb63e00207-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "895799b2-0358-4bd5-9ead-95fb63e00207" (UID: "895799b2-0358-4bd5-9ead-95fb63e00207"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 02:02:57 crc kubenswrapper[4885]: I1002 02:02:57.860312 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/895799b2-0358-4bd5-9ead-95fb63e00207-var-run-ovn\") pod \"895799b2-0358-4bd5-9ead-95fb63e00207\" (UID: \"895799b2-0358-4bd5-9ead-95fb63e00207\") " Oct 02 02:02:57 crc kubenswrapper[4885]: I1002 02:02:57.860448 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/895799b2-0358-4bd5-9ead-95fb63e00207-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "895799b2-0358-4bd5-9ead-95fb63e00207" (UID: "895799b2-0358-4bd5-9ead-95fb63e00207"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 02:02:57 crc kubenswrapper[4885]: I1002 02:02:57.860669 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/895799b2-0358-4bd5-9ead-95fb63e00207-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "895799b2-0358-4bd5-9ead-95fb63e00207" (UID: "895799b2-0358-4bd5-9ead-95fb63e00207"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:02:57 crc kubenswrapper[4885]: I1002 02:02:57.860706 4885 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/895799b2-0358-4bd5-9ead-95fb63e00207-var-log-ovn\") on node \"crc\" DevicePath \"\"" Oct 02 02:02:57 crc kubenswrapper[4885]: I1002 02:02:57.860724 4885 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/895799b2-0358-4bd5-9ead-95fb63e00207-var-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 02 02:02:57 crc kubenswrapper[4885]: I1002 02:02:57.860736 4885 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/895799b2-0358-4bd5-9ead-95fb63e00207-var-run\") on node \"crc\" DevicePath \"\"" Oct 02 02:02:57 crc kubenswrapper[4885]: I1002 02:02:57.860853 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/895799b2-0358-4bd5-9ead-95fb63e00207-scripts" (OuterVolumeSpecName: "scripts") pod "895799b2-0358-4bd5-9ead-95fb63e00207" (UID: "895799b2-0358-4bd5-9ead-95fb63e00207"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:02:57 crc kubenswrapper[4885]: I1002 02:02:57.881346 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/895799b2-0358-4bd5-9ead-95fb63e00207-kube-api-access-28sfm" (OuterVolumeSpecName: "kube-api-access-28sfm") pod "895799b2-0358-4bd5-9ead-95fb63e00207" (UID: "895799b2-0358-4bd5-9ead-95fb63e00207"). InnerVolumeSpecName "kube-api-access-28sfm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:02:57 crc kubenswrapper[4885]: I1002 02:02:57.913446 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-77fb-account-create-jn758"] Oct 02 02:02:57 crc kubenswrapper[4885]: I1002 02:02:57.962033 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-slwgc\" (UniqueName: \"kubernetes.io/projected/352a9474-7bae-4467-9e56-2d320b55494a-kube-api-access-slwgc\") pod \"352a9474-7bae-4467-9e56-2d320b55494a\" (UID: \"352a9474-7bae-4467-9e56-2d320b55494a\") " Oct 02 02:02:57 crc kubenswrapper[4885]: I1002 02:02:57.962462 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b69jk\" (UniqueName: \"kubernetes.io/projected/97dba4b7-d361-4408-9b83-a717653acc0f-kube-api-access-b69jk\") pod \"placement-77fb-account-create-jn758\" (UID: \"97dba4b7-d361-4408-9b83-a717653acc0f\") " pod="openstack/placement-77fb-account-create-jn758" Oct 02 02:02:57 crc kubenswrapper[4885]: I1002 02:02:57.962513 4885 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/895799b2-0358-4bd5-9ead-95fb63e00207-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 02:02:57 crc kubenswrapper[4885]: I1002 02:02:57.962524 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-28sfm\" (UniqueName: \"kubernetes.io/projected/895799b2-0358-4bd5-9ead-95fb63e00207-kube-api-access-28sfm\") on node \"crc\" DevicePath \"\"" Oct 02 02:02:57 crc kubenswrapper[4885]: I1002 02:02:57.962534 4885 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/895799b2-0358-4bd5-9ead-95fb63e00207-additional-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 02:02:57 crc kubenswrapper[4885]: I1002 02:02:57.964953 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/352a9474-7bae-4467-9e56-2d320b55494a-kube-api-access-slwgc" (OuterVolumeSpecName: "kube-api-access-slwgc") pod "352a9474-7bae-4467-9e56-2d320b55494a" (UID: "352a9474-7bae-4467-9e56-2d320b55494a"). InnerVolumeSpecName "kube-api-access-slwgc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:02:58 crc kubenswrapper[4885]: I1002 02:02:58.064585 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b69jk\" (UniqueName: \"kubernetes.io/projected/97dba4b7-d361-4408-9b83-a717653acc0f-kube-api-access-b69jk\") pod \"placement-77fb-account-create-jn758\" (UID: \"97dba4b7-d361-4408-9b83-a717653acc0f\") " pod="openstack/placement-77fb-account-create-jn758" Oct 02 02:02:58 crc kubenswrapper[4885]: I1002 02:02:58.071402 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-slwgc\" (UniqueName: \"kubernetes.io/projected/352a9474-7bae-4467-9e56-2d320b55494a-kube-api-access-slwgc\") on node \"crc\" DevicePath \"\"" Oct 02 02:02:58 crc kubenswrapper[4885]: I1002 02:02:58.090455 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b69jk\" (UniqueName: \"kubernetes.io/projected/97dba4b7-d361-4408-9b83-a717653acc0f-kube-api-access-b69jk\") pod \"placement-77fb-account-create-jn758\" (UID: \"97dba4b7-d361-4408-9b83-a717653acc0f\") " pod="openstack/placement-77fb-account-create-jn758" Oct 02 02:02:58 crc kubenswrapper[4885]: I1002 02:02:58.186204 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-77fb-account-create-jn758" Oct 02 02:02:58 crc kubenswrapper[4885]: I1002 02:02:58.290326 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-e8ab-account-create-wrrhx" event={"ID":"352a9474-7bae-4467-9e56-2d320b55494a","Type":"ContainerDied","Data":"534f42670f36150a104ba7d8634c851c4c2a69019f44859820b993958e197355"} Oct 02 02:02:58 crc kubenswrapper[4885]: I1002 02:02:58.290481 4885 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="534f42670f36150a104ba7d8634c851c4c2a69019f44859820b993958e197355" Oct 02 02:02:58 crc kubenswrapper[4885]: I1002 02:02:58.290396 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-e8ab-account-create-wrrhx" Oct 02 02:02:58 crc kubenswrapper[4885]: I1002 02:02:58.295157 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-fgvpt-config-zhrsw" event={"ID":"895799b2-0358-4bd5-9ead-95fb63e00207","Type":"ContainerDied","Data":"b6a5130e233c41e076122c51ee5483f80392c532a630a7ea813bf9cafec5e6e6"} Oct 02 02:02:58 crc kubenswrapper[4885]: I1002 02:02:58.295201 4885 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b6a5130e233c41e076122c51ee5483f80392c532a630a7ea813bf9cafec5e6e6" Oct 02 02:02:58 crc kubenswrapper[4885]: I1002 02:02:58.295348 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-fgvpt-config-zhrsw" Oct 02 02:02:58 crc kubenswrapper[4885]: I1002 02:02:58.351618 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-45be-account-create-824gn"] Oct 02 02:02:58 crc kubenswrapper[4885]: W1002 02:02:58.363312 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb88a1087_86c9_41ad_a0b9_48a2607bfc46.slice/crio-f6649caa79da48df80ee06a637a9c3740ef26d6cd41d020cf756857d980d9735 WatchSource:0}: Error finding container f6649caa79da48df80ee06a637a9c3740ef26d6cd41d020cf756857d980d9735: Status 404 returned error can't find the container with id f6649caa79da48df80ee06a637a9c3740ef26d6cd41d020cf756857d980d9735 Oct 02 02:02:58 crc kubenswrapper[4885]: I1002 02:02:58.662321 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-77fb-account-create-jn758"] Oct 02 02:02:58 crc kubenswrapper[4885]: W1002 02:02:58.709129 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod97dba4b7_d361_4408_9b83_a717653acc0f.slice/crio-6a2a004c5eda2a49cb957c9666a86f1a1a34cc0544f92cedf6a1a5a9cf4d52fe WatchSource:0}: Error finding container 6a2a004c5eda2a49cb957c9666a86f1a1a34cc0544f92cedf6a1a5a9cf4d52fe: Status 404 returned error can't find the container with id 6a2a004c5eda2a49cb957c9666a86f1a1a34cc0544f92cedf6a1a5a9cf4d52fe Oct 02 02:02:58 crc kubenswrapper[4885]: I1002 02:02:58.786077 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-fgvpt-config-zhrsw"] Oct 02 02:02:58 crc kubenswrapper[4885]: I1002 02:02:58.801175 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-fgvpt-config-zhrsw"] Oct 02 02:02:58 crc kubenswrapper[4885]: I1002 02:02:58.892368 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-fgvpt-config-n499l"] Oct 02 02:02:58 crc kubenswrapper[4885]: I1002 02:02:58.893562 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-fgvpt-config-n499l" Oct 02 02:02:58 crc kubenswrapper[4885]: I1002 02:02:58.895319 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Oct 02 02:02:58 crc kubenswrapper[4885]: I1002 02:02:58.905731 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-fgvpt-config-n499l"] Oct 02 02:02:58 crc kubenswrapper[4885]: I1002 02:02:58.991050 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/648c493d-cfa2-4322-a3ca-e03e233cfa44-var-log-ovn\") pod \"ovn-controller-fgvpt-config-n499l\" (UID: \"648c493d-cfa2-4322-a3ca-e03e233cfa44\") " pod="openstack/ovn-controller-fgvpt-config-n499l" Oct 02 02:02:58 crc kubenswrapper[4885]: I1002 02:02:58.991105 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h82pk\" (UniqueName: \"kubernetes.io/projected/648c493d-cfa2-4322-a3ca-e03e233cfa44-kube-api-access-h82pk\") pod \"ovn-controller-fgvpt-config-n499l\" (UID: \"648c493d-cfa2-4322-a3ca-e03e233cfa44\") " pod="openstack/ovn-controller-fgvpt-config-n499l" Oct 02 02:02:58 crc kubenswrapper[4885]: I1002 02:02:58.991217 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/648c493d-cfa2-4322-a3ca-e03e233cfa44-additional-scripts\") pod \"ovn-controller-fgvpt-config-n499l\" (UID: \"648c493d-cfa2-4322-a3ca-e03e233cfa44\") " pod="openstack/ovn-controller-fgvpt-config-n499l" Oct 02 02:02:58 crc kubenswrapper[4885]: I1002 02:02:58.991242 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/648c493d-cfa2-4322-a3ca-e03e233cfa44-var-run-ovn\") pod \"ovn-controller-fgvpt-config-n499l\" (UID: \"648c493d-cfa2-4322-a3ca-e03e233cfa44\") " pod="openstack/ovn-controller-fgvpt-config-n499l" Oct 02 02:02:58 crc kubenswrapper[4885]: I1002 02:02:58.991386 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/648c493d-cfa2-4322-a3ca-e03e233cfa44-scripts\") pod \"ovn-controller-fgvpt-config-n499l\" (UID: \"648c493d-cfa2-4322-a3ca-e03e233cfa44\") " pod="openstack/ovn-controller-fgvpt-config-n499l" Oct 02 02:02:58 crc kubenswrapper[4885]: I1002 02:02:58.991546 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/648c493d-cfa2-4322-a3ca-e03e233cfa44-var-run\") pod \"ovn-controller-fgvpt-config-n499l\" (UID: \"648c493d-cfa2-4322-a3ca-e03e233cfa44\") " pod="openstack/ovn-controller-fgvpt-config-n499l" Oct 02 02:02:59 crc kubenswrapper[4885]: I1002 02:02:59.093380 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/648c493d-cfa2-4322-a3ca-e03e233cfa44-additional-scripts\") pod \"ovn-controller-fgvpt-config-n499l\" (UID: \"648c493d-cfa2-4322-a3ca-e03e233cfa44\") " pod="openstack/ovn-controller-fgvpt-config-n499l" Oct 02 02:02:59 crc kubenswrapper[4885]: I1002 02:02:59.093462 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/648c493d-cfa2-4322-a3ca-e03e233cfa44-var-run-ovn\") pod \"ovn-controller-fgvpt-config-n499l\" (UID: \"648c493d-cfa2-4322-a3ca-e03e233cfa44\") " pod="openstack/ovn-controller-fgvpt-config-n499l" Oct 02 02:02:59 crc kubenswrapper[4885]: I1002 02:02:59.093503 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/648c493d-cfa2-4322-a3ca-e03e233cfa44-scripts\") pod \"ovn-controller-fgvpt-config-n499l\" (UID: \"648c493d-cfa2-4322-a3ca-e03e233cfa44\") " pod="openstack/ovn-controller-fgvpt-config-n499l" Oct 02 02:02:59 crc kubenswrapper[4885]: I1002 02:02:59.093573 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/648c493d-cfa2-4322-a3ca-e03e233cfa44-var-run\") pod \"ovn-controller-fgvpt-config-n499l\" (UID: \"648c493d-cfa2-4322-a3ca-e03e233cfa44\") " pod="openstack/ovn-controller-fgvpt-config-n499l" Oct 02 02:02:59 crc kubenswrapper[4885]: I1002 02:02:59.093708 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/648c493d-cfa2-4322-a3ca-e03e233cfa44-var-log-ovn\") pod \"ovn-controller-fgvpt-config-n499l\" (UID: \"648c493d-cfa2-4322-a3ca-e03e233cfa44\") " pod="openstack/ovn-controller-fgvpt-config-n499l" Oct 02 02:02:59 crc kubenswrapper[4885]: I1002 02:02:59.093766 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h82pk\" (UniqueName: \"kubernetes.io/projected/648c493d-cfa2-4322-a3ca-e03e233cfa44-kube-api-access-h82pk\") pod \"ovn-controller-fgvpt-config-n499l\" (UID: \"648c493d-cfa2-4322-a3ca-e03e233cfa44\") " pod="openstack/ovn-controller-fgvpt-config-n499l" Oct 02 02:02:59 crc kubenswrapper[4885]: I1002 02:02:59.093803 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/648c493d-cfa2-4322-a3ca-e03e233cfa44-var-run-ovn\") pod \"ovn-controller-fgvpt-config-n499l\" (UID: \"648c493d-cfa2-4322-a3ca-e03e233cfa44\") " pod="openstack/ovn-controller-fgvpt-config-n499l" Oct 02 02:02:59 crc kubenswrapper[4885]: I1002 02:02:59.093848 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/648c493d-cfa2-4322-a3ca-e03e233cfa44-var-run\") pod \"ovn-controller-fgvpt-config-n499l\" (UID: \"648c493d-cfa2-4322-a3ca-e03e233cfa44\") " pod="openstack/ovn-controller-fgvpt-config-n499l" Oct 02 02:02:59 crc kubenswrapper[4885]: I1002 02:02:59.093907 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/648c493d-cfa2-4322-a3ca-e03e233cfa44-var-log-ovn\") pod \"ovn-controller-fgvpt-config-n499l\" (UID: \"648c493d-cfa2-4322-a3ca-e03e233cfa44\") " pod="openstack/ovn-controller-fgvpt-config-n499l" Oct 02 02:02:59 crc kubenswrapper[4885]: I1002 02:02:59.095907 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/648c493d-cfa2-4322-a3ca-e03e233cfa44-scripts\") pod \"ovn-controller-fgvpt-config-n499l\" (UID: \"648c493d-cfa2-4322-a3ca-e03e233cfa44\") " pod="openstack/ovn-controller-fgvpt-config-n499l" Oct 02 02:02:59 crc kubenswrapper[4885]: I1002 02:02:59.097762 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/648c493d-cfa2-4322-a3ca-e03e233cfa44-additional-scripts\") pod \"ovn-controller-fgvpt-config-n499l\" (UID: \"648c493d-cfa2-4322-a3ca-e03e233cfa44\") " pod="openstack/ovn-controller-fgvpt-config-n499l" Oct 02 02:02:59 crc kubenswrapper[4885]: I1002 02:02:59.126017 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h82pk\" (UniqueName: \"kubernetes.io/projected/648c493d-cfa2-4322-a3ca-e03e233cfa44-kube-api-access-h82pk\") pod \"ovn-controller-fgvpt-config-n499l\" (UID: \"648c493d-cfa2-4322-a3ca-e03e233cfa44\") " pod="openstack/ovn-controller-fgvpt-config-n499l" Oct 02 02:02:59 crc kubenswrapper[4885]: I1002 02:02:59.213476 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-fgvpt-config-n499l" Oct 02 02:02:59 crc kubenswrapper[4885]: I1002 02:02:59.312165 4885 generic.go:334] "Generic (PLEG): container finished" podID="b88a1087-86c9-41ad-a0b9-48a2607bfc46" containerID="af22f9c7d8e8ba413cc823f2e85402994be30148633b974c8c2c302351d03e4e" exitCode=0 Oct 02 02:02:59 crc kubenswrapper[4885]: I1002 02:02:59.312296 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-45be-account-create-824gn" event={"ID":"b88a1087-86c9-41ad-a0b9-48a2607bfc46","Type":"ContainerDied","Data":"af22f9c7d8e8ba413cc823f2e85402994be30148633b974c8c2c302351d03e4e"} Oct 02 02:02:59 crc kubenswrapper[4885]: I1002 02:02:59.312327 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-45be-account-create-824gn" event={"ID":"b88a1087-86c9-41ad-a0b9-48a2607bfc46","Type":"ContainerStarted","Data":"f6649caa79da48df80ee06a637a9c3740ef26d6cd41d020cf756857d980d9735"} Oct 02 02:02:59 crc kubenswrapper[4885]: I1002 02:02:59.325233 4885 generic.go:334] "Generic (PLEG): container finished" podID="97dba4b7-d361-4408-9b83-a717653acc0f" containerID="c9e734a87f217a4029999cbcf9a65483cb0267cdff71063783527577082855bf" exitCode=0 Oct 02 02:02:59 crc kubenswrapper[4885]: I1002 02:02:59.325389 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-77fb-account-create-jn758" event={"ID":"97dba4b7-d361-4408-9b83-a717653acc0f","Type":"ContainerDied","Data":"c9e734a87f217a4029999cbcf9a65483cb0267cdff71063783527577082855bf"} Oct 02 02:02:59 crc kubenswrapper[4885]: I1002 02:02:59.325431 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-77fb-account-create-jn758" event={"ID":"97dba4b7-d361-4408-9b83-a717653acc0f","Type":"ContainerStarted","Data":"6a2a004c5eda2a49cb957c9666a86f1a1a34cc0544f92cedf6a1a5a9cf4d52fe"} Oct 02 02:02:59 crc kubenswrapper[4885]: I1002 02:02:59.669274 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-fgvpt-config-n499l"] Oct 02 02:03:00 crc kubenswrapper[4885]: I1002 02:03:00.055466 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="895799b2-0358-4bd5-9ead-95fb63e00207" path="/var/lib/kubelet/pods/895799b2-0358-4bd5-9ead-95fb63e00207/volumes" Oct 02 02:03:00 crc kubenswrapper[4885]: I1002 02:03:00.336583 4885 generic.go:334] "Generic (PLEG): container finished" podID="648c493d-cfa2-4322-a3ca-e03e233cfa44" containerID="671ef76134a5c82d3bedd4b9993e1330e7a9f078e6f02c8f3cca851b8d9ec988" exitCode=0 Oct 02 02:03:00 crc kubenswrapper[4885]: I1002 02:03:00.336695 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-fgvpt-config-n499l" event={"ID":"648c493d-cfa2-4322-a3ca-e03e233cfa44","Type":"ContainerDied","Data":"671ef76134a5c82d3bedd4b9993e1330e7a9f078e6f02c8f3cca851b8d9ec988"} Oct 02 02:03:00 crc kubenswrapper[4885]: I1002 02:03:00.336955 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-fgvpt-config-n499l" event={"ID":"648c493d-cfa2-4322-a3ca-e03e233cfa44","Type":"ContainerStarted","Data":"0e6b7dbfcd14a109961d047533a90ca0b13f04b696bb4328985df07bd0229476"} Oct 02 02:03:00 crc kubenswrapper[4885]: I1002 02:03:00.716200 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-45be-account-create-824gn" Oct 02 02:03:00 crc kubenswrapper[4885]: I1002 02:03:00.720976 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-77fb-account-create-jn758" Oct 02 02:03:00 crc kubenswrapper[4885]: I1002 02:03:00.820026 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b69jk\" (UniqueName: \"kubernetes.io/projected/97dba4b7-d361-4408-9b83-a717653acc0f-kube-api-access-b69jk\") pod \"97dba4b7-d361-4408-9b83-a717653acc0f\" (UID: \"97dba4b7-d361-4408-9b83-a717653acc0f\") " Oct 02 02:03:00 crc kubenswrapper[4885]: I1002 02:03:00.820249 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9tf97\" (UniqueName: \"kubernetes.io/projected/b88a1087-86c9-41ad-a0b9-48a2607bfc46-kube-api-access-9tf97\") pod \"b88a1087-86c9-41ad-a0b9-48a2607bfc46\" (UID: \"b88a1087-86c9-41ad-a0b9-48a2607bfc46\") " Oct 02 02:03:00 crc kubenswrapper[4885]: I1002 02:03:00.826004 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/97dba4b7-d361-4408-9b83-a717653acc0f-kube-api-access-b69jk" (OuterVolumeSpecName: "kube-api-access-b69jk") pod "97dba4b7-d361-4408-9b83-a717653acc0f" (UID: "97dba4b7-d361-4408-9b83-a717653acc0f"). InnerVolumeSpecName "kube-api-access-b69jk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:03:00 crc kubenswrapper[4885]: I1002 02:03:00.827683 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b88a1087-86c9-41ad-a0b9-48a2607bfc46-kube-api-access-9tf97" (OuterVolumeSpecName: "kube-api-access-9tf97") pod "b88a1087-86c9-41ad-a0b9-48a2607bfc46" (UID: "b88a1087-86c9-41ad-a0b9-48a2607bfc46"). InnerVolumeSpecName "kube-api-access-9tf97". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:03:00 crc kubenswrapper[4885]: I1002 02:03:00.923884 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9tf97\" (UniqueName: \"kubernetes.io/projected/b88a1087-86c9-41ad-a0b9-48a2607bfc46-kube-api-access-9tf97\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:00 crc kubenswrapper[4885]: I1002 02:03:00.923960 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b69jk\" (UniqueName: \"kubernetes.io/projected/97dba4b7-d361-4408-9b83-a717653acc0f-kube-api-access-b69jk\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:01 crc kubenswrapper[4885]: I1002 02:03:01.348298 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-45be-account-create-824gn" event={"ID":"b88a1087-86c9-41ad-a0b9-48a2607bfc46","Type":"ContainerDied","Data":"f6649caa79da48df80ee06a637a9c3740ef26d6cd41d020cf756857d980d9735"} Oct 02 02:03:01 crc kubenswrapper[4885]: I1002 02:03:01.348595 4885 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f6649caa79da48df80ee06a637a9c3740ef26d6cd41d020cf756857d980d9735" Oct 02 02:03:01 crc kubenswrapper[4885]: I1002 02:03:01.348322 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-45be-account-create-824gn" Oct 02 02:03:01 crc kubenswrapper[4885]: I1002 02:03:01.351188 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-77fb-account-create-jn758" Oct 02 02:03:01 crc kubenswrapper[4885]: I1002 02:03:01.351163 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-77fb-account-create-jn758" event={"ID":"97dba4b7-d361-4408-9b83-a717653acc0f","Type":"ContainerDied","Data":"6a2a004c5eda2a49cb957c9666a86f1a1a34cc0544f92cedf6a1a5a9cf4d52fe"} Oct 02 02:03:01 crc kubenswrapper[4885]: I1002 02:03:01.351408 4885 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6a2a004c5eda2a49cb957c9666a86f1a1a34cc0544f92cedf6a1a5a9cf4d52fe" Oct 02 02:03:01 crc kubenswrapper[4885]: I1002 02:03:01.742458 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-fgvpt-config-n499l" Oct 02 02:03:01 crc kubenswrapper[4885]: I1002 02:03:01.842883 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h82pk\" (UniqueName: \"kubernetes.io/projected/648c493d-cfa2-4322-a3ca-e03e233cfa44-kube-api-access-h82pk\") pod \"648c493d-cfa2-4322-a3ca-e03e233cfa44\" (UID: \"648c493d-cfa2-4322-a3ca-e03e233cfa44\") " Oct 02 02:03:01 crc kubenswrapper[4885]: I1002 02:03:01.842928 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/648c493d-cfa2-4322-a3ca-e03e233cfa44-additional-scripts\") pod \"648c493d-cfa2-4322-a3ca-e03e233cfa44\" (UID: \"648c493d-cfa2-4322-a3ca-e03e233cfa44\") " Oct 02 02:03:01 crc kubenswrapper[4885]: I1002 02:03:01.842996 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/648c493d-cfa2-4322-a3ca-e03e233cfa44-var-run\") pod \"648c493d-cfa2-4322-a3ca-e03e233cfa44\" (UID: \"648c493d-cfa2-4322-a3ca-e03e233cfa44\") " Oct 02 02:03:01 crc kubenswrapper[4885]: I1002 02:03:01.843026 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/648c493d-cfa2-4322-a3ca-e03e233cfa44-scripts\") pod \"648c493d-cfa2-4322-a3ca-e03e233cfa44\" (UID: \"648c493d-cfa2-4322-a3ca-e03e233cfa44\") " Oct 02 02:03:01 crc kubenswrapper[4885]: I1002 02:03:01.843055 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/648c493d-cfa2-4322-a3ca-e03e233cfa44-var-log-ovn\") pod \"648c493d-cfa2-4322-a3ca-e03e233cfa44\" (UID: \"648c493d-cfa2-4322-a3ca-e03e233cfa44\") " Oct 02 02:03:01 crc kubenswrapper[4885]: I1002 02:03:01.843106 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/648c493d-cfa2-4322-a3ca-e03e233cfa44-var-run-ovn\") pod \"648c493d-cfa2-4322-a3ca-e03e233cfa44\" (UID: \"648c493d-cfa2-4322-a3ca-e03e233cfa44\") " Oct 02 02:03:01 crc kubenswrapper[4885]: I1002 02:03:01.843106 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/648c493d-cfa2-4322-a3ca-e03e233cfa44-var-run" (OuterVolumeSpecName: "var-run") pod "648c493d-cfa2-4322-a3ca-e03e233cfa44" (UID: "648c493d-cfa2-4322-a3ca-e03e233cfa44"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 02:03:01 crc kubenswrapper[4885]: I1002 02:03:01.843162 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/648c493d-cfa2-4322-a3ca-e03e233cfa44-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "648c493d-cfa2-4322-a3ca-e03e233cfa44" (UID: "648c493d-cfa2-4322-a3ca-e03e233cfa44"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 02:03:01 crc kubenswrapper[4885]: I1002 02:03:01.843350 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/648c493d-cfa2-4322-a3ca-e03e233cfa44-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "648c493d-cfa2-4322-a3ca-e03e233cfa44" (UID: "648c493d-cfa2-4322-a3ca-e03e233cfa44"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 02:03:01 crc kubenswrapper[4885]: I1002 02:03:01.843767 4885 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/648c493d-cfa2-4322-a3ca-e03e233cfa44-var-run\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:01 crc kubenswrapper[4885]: I1002 02:03:01.843781 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/648c493d-cfa2-4322-a3ca-e03e233cfa44-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "648c493d-cfa2-4322-a3ca-e03e233cfa44" (UID: "648c493d-cfa2-4322-a3ca-e03e233cfa44"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:03:01 crc kubenswrapper[4885]: I1002 02:03:01.843793 4885 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/648c493d-cfa2-4322-a3ca-e03e233cfa44-var-log-ovn\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:01 crc kubenswrapper[4885]: I1002 02:03:01.843816 4885 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/648c493d-cfa2-4322-a3ca-e03e233cfa44-var-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:01 crc kubenswrapper[4885]: I1002 02:03:01.844068 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/648c493d-cfa2-4322-a3ca-e03e233cfa44-scripts" (OuterVolumeSpecName: "scripts") pod "648c493d-cfa2-4322-a3ca-e03e233cfa44" (UID: "648c493d-cfa2-4322-a3ca-e03e233cfa44"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:03:01 crc kubenswrapper[4885]: I1002 02:03:01.857088 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/648c493d-cfa2-4322-a3ca-e03e233cfa44-kube-api-access-h82pk" (OuterVolumeSpecName: "kube-api-access-h82pk") pod "648c493d-cfa2-4322-a3ca-e03e233cfa44" (UID: "648c493d-cfa2-4322-a3ca-e03e233cfa44"). InnerVolumeSpecName "kube-api-access-h82pk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:03:01 crc kubenswrapper[4885]: I1002 02:03:01.945910 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h82pk\" (UniqueName: \"kubernetes.io/projected/648c493d-cfa2-4322-a3ca-e03e233cfa44-kube-api-access-h82pk\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:01 crc kubenswrapper[4885]: I1002 02:03:01.945959 4885 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/648c493d-cfa2-4322-a3ca-e03e233cfa44-additional-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:01 crc kubenswrapper[4885]: I1002 02:03:01.945978 4885 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/648c493d-cfa2-4322-a3ca-e03e233cfa44-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:02 crc kubenswrapper[4885]: I1002 02:03:02.365094 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-fgvpt-config-n499l" event={"ID":"648c493d-cfa2-4322-a3ca-e03e233cfa44","Type":"ContainerDied","Data":"0e6b7dbfcd14a109961d047533a90ca0b13f04b696bb4328985df07bd0229476"} Oct 02 02:03:02 crc kubenswrapper[4885]: I1002 02:03:02.365157 4885 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0e6b7dbfcd14a109961d047533a90ca0b13f04b696bb4328985df07bd0229476" Oct 02 02:03:02 crc kubenswrapper[4885]: I1002 02:03:02.365200 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-fgvpt-config-n499l" Oct 02 02:03:02 crc kubenswrapper[4885]: I1002 02:03:02.658501 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/0c8d183a-f48b-4a69-8685-90c834e678ce-etc-swift\") pod \"swift-storage-0\" (UID: \"0c8d183a-f48b-4a69-8685-90c834e678ce\") " pod="openstack/swift-storage-0" Oct 02 02:03:02 crc kubenswrapper[4885]: I1002 02:03:02.664216 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/0c8d183a-f48b-4a69-8685-90c834e678ce-etc-swift\") pod \"swift-storage-0\" (UID: \"0c8d183a-f48b-4a69-8685-90c834e678ce\") " pod="openstack/swift-storage-0" Oct 02 02:03:02 crc kubenswrapper[4885]: I1002 02:03:02.718214 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Oct 02 02:03:02 crc kubenswrapper[4885]: I1002 02:03:02.809376 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 02 02:03:02 crc kubenswrapper[4885]: I1002 02:03:02.830095 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-fgvpt-config-n499l"] Oct 02 02:03:02 crc kubenswrapper[4885]: I1002 02:03:02.834827 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-fgvpt-config-n499l"] Oct 02 02:03:03 crc kubenswrapper[4885]: I1002 02:03:03.334623 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-spxxt"] Oct 02 02:03:03 crc kubenswrapper[4885]: E1002 02:03:03.335319 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b88a1087-86c9-41ad-a0b9-48a2607bfc46" containerName="mariadb-account-create" Oct 02 02:03:03 crc kubenswrapper[4885]: I1002 02:03:03.335340 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="b88a1087-86c9-41ad-a0b9-48a2607bfc46" containerName="mariadb-account-create" Oct 02 02:03:03 crc kubenswrapper[4885]: E1002 02:03:03.335370 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="648c493d-cfa2-4322-a3ca-e03e233cfa44" containerName="ovn-config" Oct 02 02:03:03 crc kubenswrapper[4885]: I1002 02:03:03.335378 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="648c493d-cfa2-4322-a3ca-e03e233cfa44" containerName="ovn-config" Oct 02 02:03:03 crc kubenswrapper[4885]: E1002 02:03:03.335401 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97dba4b7-d361-4408-9b83-a717653acc0f" containerName="mariadb-account-create" Oct 02 02:03:03 crc kubenswrapper[4885]: I1002 02:03:03.335408 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="97dba4b7-d361-4408-9b83-a717653acc0f" containerName="mariadb-account-create" Oct 02 02:03:03 crc kubenswrapper[4885]: I1002 02:03:03.335593 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="648c493d-cfa2-4322-a3ca-e03e233cfa44" containerName="ovn-config" Oct 02 02:03:03 crc kubenswrapper[4885]: I1002 02:03:03.335605 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="97dba4b7-d361-4408-9b83-a717653acc0f" containerName="mariadb-account-create" Oct 02 02:03:03 crc kubenswrapper[4885]: I1002 02:03:03.335622 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="b88a1087-86c9-41ad-a0b9-48a2607bfc46" containerName="mariadb-account-create" Oct 02 02:03:03 crc kubenswrapper[4885]: I1002 02:03:03.336272 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-spxxt" Oct 02 02:03:03 crc kubenswrapper[4885]: I1002 02:03:03.339150 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-fkkck" Oct 02 02:03:03 crc kubenswrapper[4885]: I1002 02:03:03.339208 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Oct 02 02:03:03 crc kubenswrapper[4885]: I1002 02:03:03.352646 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-spxxt"] Oct 02 02:03:03 crc kubenswrapper[4885]: W1002 02:03:03.399210 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0c8d183a_f48b_4a69_8685_90c834e678ce.slice/crio-6927b151dcdda2cfa5edf5fc0c9d8ac9c2ee13fdca5397cf4f2ec5436dfcc9ae WatchSource:0}: Error finding container 6927b151dcdda2cfa5edf5fc0c9d8ac9c2ee13fdca5397cf4f2ec5436dfcc9ae: Status 404 returned error can't find the container with id 6927b151dcdda2cfa5edf5fc0c9d8ac9c2ee13fdca5397cf4f2ec5436dfcc9ae Oct 02 02:03:03 crc kubenswrapper[4885]: I1002 02:03:03.401992 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Oct 02 02:03:03 crc kubenswrapper[4885]: I1002 02:03:03.472359 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6a9ed972-b8c7-4ff7-af7b-899842ecc19c-db-sync-config-data\") pod \"glance-db-sync-spxxt\" (UID: \"6a9ed972-b8c7-4ff7-af7b-899842ecc19c\") " pod="openstack/glance-db-sync-spxxt" Oct 02 02:03:03 crc kubenswrapper[4885]: I1002 02:03:03.472405 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tk2sb\" (UniqueName: \"kubernetes.io/projected/6a9ed972-b8c7-4ff7-af7b-899842ecc19c-kube-api-access-tk2sb\") pod \"glance-db-sync-spxxt\" (UID: \"6a9ed972-b8c7-4ff7-af7b-899842ecc19c\") " pod="openstack/glance-db-sync-spxxt" Oct 02 02:03:03 crc kubenswrapper[4885]: I1002 02:03:03.472452 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a9ed972-b8c7-4ff7-af7b-899842ecc19c-combined-ca-bundle\") pod \"glance-db-sync-spxxt\" (UID: \"6a9ed972-b8c7-4ff7-af7b-899842ecc19c\") " pod="openstack/glance-db-sync-spxxt" Oct 02 02:03:03 crc kubenswrapper[4885]: I1002 02:03:03.472560 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a9ed972-b8c7-4ff7-af7b-899842ecc19c-config-data\") pod \"glance-db-sync-spxxt\" (UID: \"6a9ed972-b8c7-4ff7-af7b-899842ecc19c\") " pod="openstack/glance-db-sync-spxxt" Oct 02 02:03:03 crc kubenswrapper[4885]: I1002 02:03:03.573992 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a9ed972-b8c7-4ff7-af7b-899842ecc19c-combined-ca-bundle\") pod \"glance-db-sync-spxxt\" (UID: \"6a9ed972-b8c7-4ff7-af7b-899842ecc19c\") " pod="openstack/glance-db-sync-spxxt" Oct 02 02:03:03 crc kubenswrapper[4885]: I1002 02:03:03.574047 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a9ed972-b8c7-4ff7-af7b-899842ecc19c-config-data\") pod \"glance-db-sync-spxxt\" (UID: \"6a9ed972-b8c7-4ff7-af7b-899842ecc19c\") " pod="openstack/glance-db-sync-spxxt" Oct 02 02:03:03 crc kubenswrapper[4885]: I1002 02:03:03.574141 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6a9ed972-b8c7-4ff7-af7b-899842ecc19c-db-sync-config-data\") pod \"glance-db-sync-spxxt\" (UID: \"6a9ed972-b8c7-4ff7-af7b-899842ecc19c\") " pod="openstack/glance-db-sync-spxxt" Oct 02 02:03:03 crc kubenswrapper[4885]: I1002 02:03:03.574165 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tk2sb\" (UniqueName: \"kubernetes.io/projected/6a9ed972-b8c7-4ff7-af7b-899842ecc19c-kube-api-access-tk2sb\") pod \"glance-db-sync-spxxt\" (UID: \"6a9ed972-b8c7-4ff7-af7b-899842ecc19c\") " pod="openstack/glance-db-sync-spxxt" Oct 02 02:03:03 crc kubenswrapper[4885]: I1002 02:03:03.582722 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a9ed972-b8c7-4ff7-af7b-899842ecc19c-combined-ca-bundle\") pod \"glance-db-sync-spxxt\" (UID: \"6a9ed972-b8c7-4ff7-af7b-899842ecc19c\") " pod="openstack/glance-db-sync-spxxt" Oct 02 02:03:03 crc kubenswrapper[4885]: I1002 02:03:03.582997 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6a9ed972-b8c7-4ff7-af7b-899842ecc19c-db-sync-config-data\") pod \"glance-db-sync-spxxt\" (UID: \"6a9ed972-b8c7-4ff7-af7b-899842ecc19c\") " pod="openstack/glance-db-sync-spxxt" Oct 02 02:03:03 crc kubenswrapper[4885]: I1002 02:03:03.584475 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a9ed972-b8c7-4ff7-af7b-899842ecc19c-config-data\") pod \"glance-db-sync-spxxt\" (UID: \"6a9ed972-b8c7-4ff7-af7b-899842ecc19c\") " pod="openstack/glance-db-sync-spxxt" Oct 02 02:03:03 crc kubenswrapper[4885]: I1002 02:03:03.598422 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tk2sb\" (UniqueName: \"kubernetes.io/projected/6a9ed972-b8c7-4ff7-af7b-899842ecc19c-kube-api-access-tk2sb\") pod \"glance-db-sync-spxxt\" (UID: \"6a9ed972-b8c7-4ff7-af7b-899842ecc19c\") " pod="openstack/glance-db-sync-spxxt" Oct 02 02:03:03 crc kubenswrapper[4885]: I1002 02:03:03.655013 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-spxxt" Oct 02 02:03:04 crc kubenswrapper[4885]: I1002 02:03:04.063755 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="648c493d-cfa2-4322-a3ca-e03e233cfa44" path="/var/lib/kubelet/pods/648c493d-cfa2-4322-a3ca-e03e233cfa44/volumes" Oct 02 02:03:04 crc kubenswrapper[4885]: I1002 02:03:04.148306 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:03:04 crc kubenswrapper[4885]: I1002 02:03:04.241909 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-spxxt"] Oct 02 02:03:04 crc kubenswrapper[4885]: I1002 02:03:04.368054 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 02 02:03:04 crc kubenswrapper[4885]: I1002 02:03:04.393419 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-spxxt" event={"ID":"6a9ed972-b8c7-4ff7-af7b-899842ecc19c","Type":"ContainerStarted","Data":"c4c16af334ec7909627f304bcf3404a3f330aae644750d65ad8deafc948d085b"} Oct 02 02:03:04 crc kubenswrapper[4885]: I1002 02:03:04.402495 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0c8d183a-f48b-4a69-8685-90c834e678ce","Type":"ContainerStarted","Data":"6927b151dcdda2cfa5edf5fc0c9d8ac9c2ee13fdca5397cf4f2ec5436dfcc9ae"} Oct 02 02:03:05 crc kubenswrapper[4885]: I1002 02:03:05.411368 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0c8d183a-f48b-4a69-8685-90c834e678ce","Type":"ContainerStarted","Data":"69f0dbee9d483a9385ec646d94b6c7a91c020eb87fc99265e128afc2c529c416"} Oct 02 02:03:05 crc kubenswrapper[4885]: I1002 02:03:05.411667 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0c8d183a-f48b-4a69-8685-90c834e678ce","Type":"ContainerStarted","Data":"55bf226e103041b6f6a77d9d2a85babbd1cd01f55c8b4307413fe6f860f34eb4"} Oct 02 02:03:05 crc kubenswrapper[4885]: I1002 02:03:05.411678 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0c8d183a-f48b-4a69-8685-90c834e678ce","Type":"ContainerStarted","Data":"e07f633de5d3d0231b9821485ac197e179b8287adf6afb38bf52b7c8ffcc4761"} Oct 02 02:03:05 crc kubenswrapper[4885]: I1002 02:03:05.411686 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0c8d183a-f48b-4a69-8685-90c834e678ce","Type":"ContainerStarted","Data":"1cfab235919595d811fa714ee9ef1554ba6f734b5ff804f6ad472e879090503c"} Oct 02 02:03:05 crc kubenswrapper[4885]: I1002 02:03:05.993354 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-sff5w"] Oct 02 02:03:05 crc kubenswrapper[4885]: I1002 02:03:05.994883 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-sff5w" Oct 02 02:03:06 crc kubenswrapper[4885]: I1002 02:03:06.007697 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-sff5w"] Oct 02 02:03:06 crc kubenswrapper[4885]: I1002 02:03:06.088186 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-d2d8l"] Oct 02 02:03:06 crc kubenswrapper[4885]: I1002 02:03:06.089100 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-d2d8l" Oct 02 02:03:06 crc kubenswrapper[4885]: I1002 02:03:06.100288 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-d2d8l"] Oct 02 02:03:06 crc kubenswrapper[4885]: I1002 02:03:06.115129 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cgqrc\" (UniqueName: \"kubernetes.io/projected/ab748dfb-c9b3-412c-8c8d-fe40d3c1d9e7-kube-api-access-cgqrc\") pod \"cinder-db-create-sff5w\" (UID: \"ab748dfb-c9b3-412c-8c8d-fe40d3c1d9e7\") " pod="openstack/cinder-db-create-sff5w" Oct 02 02:03:06 crc kubenswrapper[4885]: I1002 02:03:06.216214 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-76f7l\" (UniqueName: \"kubernetes.io/projected/5ab1b5ad-d6a5-4bfa-96bd-9688c397a87b-kube-api-access-76f7l\") pod \"barbican-db-create-d2d8l\" (UID: \"5ab1b5ad-d6a5-4bfa-96bd-9688c397a87b\") " pod="openstack/barbican-db-create-d2d8l" Oct 02 02:03:06 crc kubenswrapper[4885]: I1002 02:03:06.216321 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cgqrc\" (UniqueName: \"kubernetes.io/projected/ab748dfb-c9b3-412c-8c8d-fe40d3c1d9e7-kube-api-access-cgqrc\") pod \"cinder-db-create-sff5w\" (UID: \"ab748dfb-c9b3-412c-8c8d-fe40d3c1d9e7\") " pod="openstack/cinder-db-create-sff5w" Oct 02 02:03:06 crc kubenswrapper[4885]: I1002 02:03:06.236997 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cgqrc\" (UniqueName: \"kubernetes.io/projected/ab748dfb-c9b3-412c-8c8d-fe40d3c1d9e7-kube-api-access-cgqrc\") pod \"cinder-db-create-sff5w\" (UID: \"ab748dfb-c9b3-412c-8c8d-fe40d3c1d9e7\") " pod="openstack/cinder-db-create-sff5w" Oct 02 02:03:06 crc kubenswrapper[4885]: I1002 02:03:06.308228 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-tvjc2"] Oct 02 02:03:06 crc kubenswrapper[4885]: I1002 02:03:06.309139 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-tvjc2" Oct 02 02:03:06 crc kubenswrapper[4885]: I1002 02:03:06.312097 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 02 02:03:06 crc kubenswrapper[4885]: I1002 02:03:06.312326 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 02 02:03:06 crc kubenswrapper[4885]: I1002 02:03:06.312739 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 02 02:03:06 crc kubenswrapper[4885]: I1002 02:03:06.312878 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-h4g4g" Oct 02 02:03:06 crc kubenswrapper[4885]: I1002 02:03:06.321528 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-76f7l\" (UniqueName: \"kubernetes.io/projected/5ab1b5ad-d6a5-4bfa-96bd-9688c397a87b-kube-api-access-76f7l\") pod \"barbican-db-create-d2d8l\" (UID: \"5ab1b5ad-d6a5-4bfa-96bd-9688c397a87b\") " pod="openstack/barbican-db-create-d2d8l" Oct 02 02:03:06 crc kubenswrapper[4885]: I1002 02:03:06.329787 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-tvjc2"] Oct 02 02:03:06 crc kubenswrapper[4885]: I1002 02:03:06.358102 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-sff5w" Oct 02 02:03:06 crc kubenswrapper[4885]: I1002 02:03:06.361980 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-76f7l\" (UniqueName: \"kubernetes.io/projected/5ab1b5ad-d6a5-4bfa-96bd-9688c397a87b-kube-api-access-76f7l\") pod \"barbican-db-create-d2d8l\" (UID: \"5ab1b5ad-d6a5-4bfa-96bd-9688c397a87b\") " pod="openstack/barbican-db-create-d2d8l" Oct 02 02:03:06 crc kubenswrapper[4885]: I1002 02:03:06.382179 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-t2x9j"] Oct 02 02:03:06 crc kubenswrapper[4885]: I1002 02:03:06.383298 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-t2x9j" Oct 02 02:03:06 crc kubenswrapper[4885]: I1002 02:03:06.396575 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-t2x9j"] Oct 02 02:03:06 crc kubenswrapper[4885]: I1002 02:03:06.406731 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-d2d8l" Oct 02 02:03:06 crc kubenswrapper[4885]: I1002 02:03:06.423413 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lvvjp\" (UniqueName: \"kubernetes.io/projected/d0092499-91ac-4e53-8c91-04d30bd49e68-kube-api-access-lvvjp\") pod \"keystone-db-sync-tvjc2\" (UID: \"d0092499-91ac-4e53-8c91-04d30bd49e68\") " pod="openstack/keystone-db-sync-tvjc2" Oct 02 02:03:06 crc kubenswrapper[4885]: I1002 02:03:06.423450 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0092499-91ac-4e53-8c91-04d30bd49e68-config-data\") pod \"keystone-db-sync-tvjc2\" (UID: \"d0092499-91ac-4e53-8c91-04d30bd49e68\") " pod="openstack/keystone-db-sync-tvjc2" Oct 02 02:03:06 crc kubenswrapper[4885]: I1002 02:03:06.423522 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0092499-91ac-4e53-8c91-04d30bd49e68-combined-ca-bundle\") pod \"keystone-db-sync-tvjc2\" (UID: \"d0092499-91ac-4e53-8c91-04d30bd49e68\") " pod="openstack/keystone-db-sync-tvjc2" Oct 02 02:03:06 crc kubenswrapper[4885]: I1002 02:03:06.525099 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lvvjp\" (UniqueName: \"kubernetes.io/projected/d0092499-91ac-4e53-8c91-04d30bd49e68-kube-api-access-lvvjp\") pod \"keystone-db-sync-tvjc2\" (UID: \"d0092499-91ac-4e53-8c91-04d30bd49e68\") " pod="openstack/keystone-db-sync-tvjc2" Oct 02 02:03:06 crc kubenswrapper[4885]: I1002 02:03:06.525371 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0092499-91ac-4e53-8c91-04d30bd49e68-config-data\") pod \"keystone-db-sync-tvjc2\" (UID: \"d0092499-91ac-4e53-8c91-04d30bd49e68\") " pod="openstack/keystone-db-sync-tvjc2" Oct 02 02:03:06 crc kubenswrapper[4885]: I1002 02:03:06.525421 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0092499-91ac-4e53-8c91-04d30bd49e68-combined-ca-bundle\") pod \"keystone-db-sync-tvjc2\" (UID: \"d0092499-91ac-4e53-8c91-04d30bd49e68\") " pod="openstack/keystone-db-sync-tvjc2" Oct 02 02:03:06 crc kubenswrapper[4885]: I1002 02:03:06.525474 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6g68n\" (UniqueName: \"kubernetes.io/projected/e7eb74c5-7bff-43b5-9bd5-89b05eb43ff6-kube-api-access-6g68n\") pod \"neutron-db-create-t2x9j\" (UID: \"e7eb74c5-7bff-43b5-9bd5-89b05eb43ff6\") " pod="openstack/neutron-db-create-t2x9j" Oct 02 02:03:06 crc kubenswrapper[4885]: I1002 02:03:06.533539 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0092499-91ac-4e53-8c91-04d30bd49e68-config-data\") pod \"keystone-db-sync-tvjc2\" (UID: \"d0092499-91ac-4e53-8c91-04d30bd49e68\") " pod="openstack/keystone-db-sync-tvjc2" Oct 02 02:03:06 crc kubenswrapper[4885]: I1002 02:03:06.534186 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0092499-91ac-4e53-8c91-04d30bd49e68-combined-ca-bundle\") pod \"keystone-db-sync-tvjc2\" (UID: \"d0092499-91ac-4e53-8c91-04d30bd49e68\") " pod="openstack/keystone-db-sync-tvjc2" Oct 02 02:03:06 crc kubenswrapper[4885]: I1002 02:03:06.544196 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lvvjp\" (UniqueName: \"kubernetes.io/projected/d0092499-91ac-4e53-8c91-04d30bd49e68-kube-api-access-lvvjp\") pod \"keystone-db-sync-tvjc2\" (UID: \"d0092499-91ac-4e53-8c91-04d30bd49e68\") " pod="openstack/keystone-db-sync-tvjc2" Oct 02 02:03:06 crc kubenswrapper[4885]: I1002 02:03:06.626927 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6g68n\" (UniqueName: \"kubernetes.io/projected/e7eb74c5-7bff-43b5-9bd5-89b05eb43ff6-kube-api-access-6g68n\") pod \"neutron-db-create-t2x9j\" (UID: \"e7eb74c5-7bff-43b5-9bd5-89b05eb43ff6\") " pod="openstack/neutron-db-create-t2x9j" Oct 02 02:03:06 crc kubenswrapper[4885]: I1002 02:03:06.629379 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-tvjc2" Oct 02 02:03:06 crc kubenswrapper[4885]: I1002 02:03:06.644063 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6g68n\" (UniqueName: \"kubernetes.io/projected/e7eb74c5-7bff-43b5-9bd5-89b05eb43ff6-kube-api-access-6g68n\") pod \"neutron-db-create-t2x9j\" (UID: \"e7eb74c5-7bff-43b5-9bd5-89b05eb43ff6\") " pod="openstack/neutron-db-create-t2x9j" Oct 02 02:03:06 crc kubenswrapper[4885]: I1002 02:03:06.777046 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-t2x9j" Oct 02 02:03:06 crc kubenswrapper[4885]: I1002 02:03:06.851778 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-sff5w"] Oct 02 02:03:06 crc kubenswrapper[4885]: W1002 02:03:06.863063 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podab748dfb_c9b3_412c_8c8d_fe40d3c1d9e7.slice/crio-2b1277a4327b32910e8fd1128d829988d6002d3dfa594a8554585e5f238a6804 WatchSource:0}: Error finding container 2b1277a4327b32910e8fd1128d829988d6002d3dfa594a8554585e5f238a6804: Status 404 returned error can't find the container with id 2b1277a4327b32910e8fd1128d829988d6002d3dfa594a8554585e5f238a6804 Oct 02 02:03:06 crc kubenswrapper[4885]: I1002 02:03:06.921605 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-d2d8l"] Oct 02 02:03:07 crc kubenswrapper[4885]: I1002 02:03:07.399873 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-tvjc2"] Oct 02 02:03:07 crc kubenswrapper[4885]: I1002 02:03:07.447705 4885 generic.go:334] "Generic (PLEG): container finished" podID="5ab1b5ad-d6a5-4bfa-96bd-9688c397a87b" containerID="af98d4d991ccc2571344d27684d911494c041b09417917524d78ad7f42ba6f84" exitCode=0 Oct 02 02:03:07 crc kubenswrapper[4885]: I1002 02:03:07.448921 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-d2d8l" event={"ID":"5ab1b5ad-d6a5-4bfa-96bd-9688c397a87b","Type":"ContainerDied","Data":"af98d4d991ccc2571344d27684d911494c041b09417917524d78ad7f42ba6f84"} Oct 02 02:03:07 crc kubenswrapper[4885]: I1002 02:03:07.448967 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-d2d8l" event={"ID":"5ab1b5ad-d6a5-4bfa-96bd-9688c397a87b","Type":"ContainerStarted","Data":"74f60e607a645c9bf5e881d869c880e9dc13ac6a9c315dc2ba00cf501a15229f"} Oct 02 02:03:07 crc kubenswrapper[4885]: I1002 02:03:07.456953 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0c8d183a-f48b-4a69-8685-90c834e678ce","Type":"ContainerStarted","Data":"8d39cb30ab5e7cde4bc8fe34b35bf66fc371a249acba0c6ced0e975e826263d5"} Oct 02 02:03:07 crc kubenswrapper[4885]: I1002 02:03:07.456987 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0c8d183a-f48b-4a69-8685-90c834e678ce","Type":"ContainerStarted","Data":"6450263e82fc65fd2e1840d84f0ee630ffaa544ccb4288322e72816181fbe6f6"} Oct 02 02:03:07 crc kubenswrapper[4885]: I1002 02:03:07.458025 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-tvjc2" event={"ID":"d0092499-91ac-4e53-8c91-04d30bd49e68","Type":"ContainerStarted","Data":"f55777e4b40637ee51de638450c8feb5694e9d731c777be79bc1158d6f020c70"} Oct 02 02:03:07 crc kubenswrapper[4885]: I1002 02:03:07.461978 4885 generic.go:334] "Generic (PLEG): container finished" podID="ab748dfb-c9b3-412c-8c8d-fe40d3c1d9e7" containerID="f25a01d3dd742ffada042f77bc5438c703eadf29a06cfa21a978326393093670" exitCode=0 Oct 02 02:03:07 crc kubenswrapper[4885]: I1002 02:03:07.462009 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-sff5w" event={"ID":"ab748dfb-c9b3-412c-8c8d-fe40d3c1d9e7","Type":"ContainerDied","Data":"f25a01d3dd742ffada042f77bc5438c703eadf29a06cfa21a978326393093670"} Oct 02 02:03:07 crc kubenswrapper[4885]: I1002 02:03:07.462136 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-sff5w" event={"ID":"ab748dfb-c9b3-412c-8c8d-fe40d3c1d9e7","Type":"ContainerStarted","Data":"2b1277a4327b32910e8fd1128d829988d6002d3dfa594a8554585e5f238a6804"} Oct 02 02:03:07 crc kubenswrapper[4885]: I1002 02:03:07.480981 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-t2x9j"] Oct 02 02:03:07 crc kubenswrapper[4885]: W1002 02:03:07.493359 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode7eb74c5_7bff_43b5_9bd5_89b05eb43ff6.slice/crio-cff7f5ecebfe11bcb7f68bdaedbed4be7bad0d4da999fc74f4bc82f3e7f1ce2c WatchSource:0}: Error finding container cff7f5ecebfe11bcb7f68bdaedbed4be7bad0d4da999fc74f4bc82f3e7f1ce2c: Status 404 returned error can't find the container with id cff7f5ecebfe11bcb7f68bdaedbed4be7bad0d4da999fc74f4bc82f3e7f1ce2c Oct 02 02:03:08 crc kubenswrapper[4885]: I1002 02:03:08.478155 4885 generic.go:334] "Generic (PLEG): container finished" podID="e7eb74c5-7bff-43b5-9bd5-89b05eb43ff6" containerID="417844a7916dd6611a997b92463595cb69243430f38bd4bc93e7d8ec63b1a43b" exitCode=0 Oct 02 02:03:08 crc kubenswrapper[4885]: I1002 02:03:08.478233 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-t2x9j" event={"ID":"e7eb74c5-7bff-43b5-9bd5-89b05eb43ff6","Type":"ContainerDied","Data":"417844a7916dd6611a997b92463595cb69243430f38bd4bc93e7d8ec63b1a43b"} Oct 02 02:03:08 crc kubenswrapper[4885]: I1002 02:03:08.478469 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-t2x9j" event={"ID":"e7eb74c5-7bff-43b5-9bd5-89b05eb43ff6","Type":"ContainerStarted","Data":"cff7f5ecebfe11bcb7f68bdaedbed4be7bad0d4da999fc74f4bc82f3e7f1ce2c"} Oct 02 02:03:08 crc kubenswrapper[4885]: I1002 02:03:08.484077 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0c8d183a-f48b-4a69-8685-90c834e678ce","Type":"ContainerStarted","Data":"3f2220b4c02c29511c6a5d038d10de83ea5c09bec957084d9d0768f832f8ea8a"} Oct 02 02:03:08 crc kubenswrapper[4885]: I1002 02:03:08.484126 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0c8d183a-f48b-4a69-8685-90c834e678ce","Type":"ContainerStarted","Data":"3473e98af0a6ffcc4a3bb2533ef546475a7d172a0aad3ec8f1e1bfe44187cc8e"} Oct 02 02:03:08 crc kubenswrapper[4885]: I1002 02:03:08.842349 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-d2d8l" Oct 02 02:03:08 crc kubenswrapper[4885]: I1002 02:03:08.974920 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-76f7l\" (UniqueName: \"kubernetes.io/projected/5ab1b5ad-d6a5-4bfa-96bd-9688c397a87b-kube-api-access-76f7l\") pod \"5ab1b5ad-d6a5-4bfa-96bd-9688c397a87b\" (UID: \"5ab1b5ad-d6a5-4bfa-96bd-9688c397a87b\") " Oct 02 02:03:08 crc kubenswrapper[4885]: I1002 02:03:08.981250 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ab1b5ad-d6a5-4bfa-96bd-9688c397a87b-kube-api-access-76f7l" (OuterVolumeSpecName: "kube-api-access-76f7l") pod "5ab1b5ad-d6a5-4bfa-96bd-9688c397a87b" (UID: "5ab1b5ad-d6a5-4bfa-96bd-9688c397a87b"). InnerVolumeSpecName "kube-api-access-76f7l". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:03:09 crc kubenswrapper[4885]: I1002 02:03:09.076357 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-76f7l\" (UniqueName: \"kubernetes.io/projected/5ab1b5ad-d6a5-4bfa-96bd-9688c397a87b-kube-api-access-76f7l\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:09 crc kubenswrapper[4885]: I1002 02:03:09.203943 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-sff5w" Oct 02 02:03:09 crc kubenswrapper[4885]: I1002 02:03:09.381438 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cgqrc\" (UniqueName: \"kubernetes.io/projected/ab748dfb-c9b3-412c-8c8d-fe40d3c1d9e7-kube-api-access-cgqrc\") pod \"ab748dfb-c9b3-412c-8c8d-fe40d3c1d9e7\" (UID: \"ab748dfb-c9b3-412c-8c8d-fe40d3c1d9e7\") " Oct 02 02:03:09 crc kubenswrapper[4885]: I1002 02:03:09.393548 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ab748dfb-c9b3-412c-8c8d-fe40d3c1d9e7-kube-api-access-cgqrc" (OuterVolumeSpecName: "kube-api-access-cgqrc") pod "ab748dfb-c9b3-412c-8c8d-fe40d3c1d9e7" (UID: "ab748dfb-c9b3-412c-8c8d-fe40d3c1d9e7"). InnerVolumeSpecName "kube-api-access-cgqrc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:03:09 crc kubenswrapper[4885]: I1002 02:03:09.483463 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cgqrc\" (UniqueName: \"kubernetes.io/projected/ab748dfb-c9b3-412c-8c8d-fe40d3c1d9e7-kube-api-access-cgqrc\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:09 crc kubenswrapper[4885]: I1002 02:03:09.494381 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0c8d183a-f48b-4a69-8685-90c834e678ce","Type":"ContainerStarted","Data":"b069a254e080010c087d2992b11d75245385b86666bdee1a6d61a8cba5046863"} Oct 02 02:03:09 crc kubenswrapper[4885]: I1002 02:03:09.495907 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-sff5w" event={"ID":"ab748dfb-c9b3-412c-8c8d-fe40d3c1d9e7","Type":"ContainerDied","Data":"2b1277a4327b32910e8fd1128d829988d6002d3dfa594a8554585e5f238a6804"} Oct 02 02:03:09 crc kubenswrapper[4885]: I1002 02:03:09.495963 4885 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2b1277a4327b32910e8fd1128d829988d6002d3dfa594a8554585e5f238a6804" Oct 02 02:03:09 crc kubenswrapper[4885]: I1002 02:03:09.495965 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-sff5w" Oct 02 02:03:09 crc kubenswrapper[4885]: I1002 02:03:09.497354 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-d2d8l" Oct 02 02:03:09 crc kubenswrapper[4885]: I1002 02:03:09.497345 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-d2d8l" event={"ID":"5ab1b5ad-d6a5-4bfa-96bd-9688c397a87b","Type":"ContainerDied","Data":"74f60e607a645c9bf5e881d869c880e9dc13ac6a9c315dc2ba00cf501a15229f"} Oct 02 02:03:09 crc kubenswrapper[4885]: I1002 02:03:09.497476 4885 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="74f60e607a645c9bf5e881d869c880e9dc13ac6a9c315dc2ba00cf501a15229f" Oct 02 02:03:09 crc kubenswrapper[4885]: I1002 02:03:09.819311 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-t2x9j" Oct 02 02:03:09 crc kubenswrapper[4885]: I1002 02:03:09.991307 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g68n\" (UniqueName: \"kubernetes.io/projected/e7eb74c5-7bff-43b5-9bd5-89b05eb43ff6-kube-api-access-6g68n\") pod \"e7eb74c5-7bff-43b5-9bd5-89b05eb43ff6\" (UID: \"e7eb74c5-7bff-43b5-9bd5-89b05eb43ff6\") " Oct 02 02:03:09 crc kubenswrapper[4885]: I1002 02:03:09.995345 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7eb74c5-7bff-43b5-9bd5-89b05eb43ff6-kube-api-access-6g68n" (OuterVolumeSpecName: "kube-api-access-6g68n") pod "e7eb74c5-7bff-43b5-9bd5-89b05eb43ff6" (UID: "e7eb74c5-7bff-43b5-9bd5-89b05eb43ff6"). InnerVolumeSpecName "kube-api-access-6g68n". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:03:10 crc kubenswrapper[4885]: I1002 02:03:10.094245 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g68n\" (UniqueName: \"kubernetes.io/projected/e7eb74c5-7bff-43b5-9bd5-89b05eb43ff6-kube-api-access-6g68n\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:10 crc kubenswrapper[4885]: I1002 02:03:10.530387 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0c8d183a-f48b-4a69-8685-90c834e678ce","Type":"ContainerStarted","Data":"de221c004d34ddaf75094bb5274238778603d276dbcfbfb9b07b38366b49cfb1"} Oct 02 02:03:10 crc kubenswrapper[4885]: I1002 02:03:10.532020 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-t2x9j" event={"ID":"e7eb74c5-7bff-43b5-9bd5-89b05eb43ff6","Type":"ContainerDied","Data":"cff7f5ecebfe11bcb7f68bdaedbed4be7bad0d4da999fc74f4bc82f3e7f1ce2c"} Oct 02 02:03:10 crc kubenswrapper[4885]: I1002 02:03:10.532077 4885 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cff7f5ecebfe11bcb7f68bdaedbed4be7bad0d4da999fc74f4bc82f3e7f1ce2c" Oct 02 02:03:10 crc kubenswrapper[4885]: I1002 02:03:10.532087 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-t2x9j" Oct 02 02:03:13 crc kubenswrapper[4885]: I1002 02:03:13.265633 4885 patch_prober.go:28] interesting pod/machine-config-daemon-rttx8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 02:03:13 crc kubenswrapper[4885]: I1002 02:03:13.266041 4885 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 02:03:13 crc kubenswrapper[4885]: I1002 02:03:13.266093 4885 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" Oct 02 02:03:13 crc kubenswrapper[4885]: I1002 02:03:13.266745 4885 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4708a92cba43cfdb8c289fa5ff76e06b7b83a40fb102347f523fcefc325254bd"} pod="openshift-machine-config-operator/machine-config-daemon-rttx8" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 02:03:13 crc kubenswrapper[4885]: I1002 02:03:13.266831 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" containerName="machine-config-daemon" containerID="cri-o://4708a92cba43cfdb8c289fa5ff76e06b7b83a40fb102347f523fcefc325254bd" gracePeriod=600 Oct 02 02:03:13 crc kubenswrapper[4885]: I1002 02:03:13.559861 4885 generic.go:334] "Generic (PLEG): container finished" podID="d169c8ba-000a-4218-8e3b-4ae53035b110" containerID="4708a92cba43cfdb8c289fa5ff76e06b7b83a40fb102347f523fcefc325254bd" exitCode=0 Oct 02 02:03:13 crc kubenswrapper[4885]: I1002 02:03:13.559929 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" event={"ID":"d169c8ba-000a-4218-8e3b-4ae53035b110","Type":"ContainerDied","Data":"4708a92cba43cfdb8c289fa5ff76e06b7b83a40fb102347f523fcefc325254bd"} Oct 02 02:03:13 crc kubenswrapper[4885]: I1002 02:03:13.559980 4885 scope.go:117] "RemoveContainer" containerID="83cae60dd6556fc08f2eddf3fd26fd0abb56b3d0a4d62534298ea5f48fc39e18" Oct 02 02:03:16 crc kubenswrapper[4885]: I1002 02:03:16.143888 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-fe7d-account-create-qtl4l"] Oct 02 02:03:16 crc kubenswrapper[4885]: E1002 02:03:16.144412 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab748dfb-c9b3-412c-8c8d-fe40d3c1d9e7" containerName="mariadb-database-create" Oct 02 02:03:16 crc kubenswrapper[4885]: I1002 02:03:16.144423 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab748dfb-c9b3-412c-8c8d-fe40d3c1d9e7" containerName="mariadb-database-create" Oct 02 02:03:16 crc kubenswrapper[4885]: E1002 02:03:16.144442 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7eb74c5-7bff-43b5-9bd5-89b05eb43ff6" containerName="mariadb-database-create" Oct 02 02:03:16 crc kubenswrapper[4885]: I1002 02:03:16.144448 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7eb74c5-7bff-43b5-9bd5-89b05eb43ff6" containerName="mariadb-database-create" Oct 02 02:03:16 crc kubenswrapper[4885]: E1002 02:03:16.144471 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ab1b5ad-d6a5-4bfa-96bd-9688c397a87b" containerName="mariadb-database-create" Oct 02 02:03:16 crc kubenswrapper[4885]: I1002 02:03:16.144476 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ab1b5ad-d6a5-4bfa-96bd-9688c397a87b" containerName="mariadb-database-create" Oct 02 02:03:16 crc kubenswrapper[4885]: I1002 02:03:16.144626 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ab1b5ad-d6a5-4bfa-96bd-9688c397a87b" containerName="mariadb-database-create" Oct 02 02:03:16 crc kubenswrapper[4885]: I1002 02:03:16.144662 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab748dfb-c9b3-412c-8c8d-fe40d3c1d9e7" containerName="mariadb-database-create" Oct 02 02:03:16 crc kubenswrapper[4885]: I1002 02:03:16.144682 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="e7eb74c5-7bff-43b5-9bd5-89b05eb43ff6" containerName="mariadb-database-create" Oct 02 02:03:16 crc kubenswrapper[4885]: I1002 02:03:16.145324 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-fe7d-account-create-qtl4l" Oct 02 02:03:16 crc kubenswrapper[4885]: I1002 02:03:16.147318 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Oct 02 02:03:16 crc kubenswrapper[4885]: I1002 02:03:16.149774 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-fe7d-account-create-qtl4l"] Oct 02 02:03:16 crc kubenswrapper[4885]: I1002 02:03:16.236221 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-1c48-account-create-6cz92"] Oct 02 02:03:16 crc kubenswrapper[4885]: I1002 02:03:16.237451 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-1c48-account-create-6cz92" Oct 02 02:03:16 crc kubenswrapper[4885]: I1002 02:03:16.239568 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Oct 02 02:03:16 crc kubenswrapper[4885]: I1002 02:03:16.262240 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-1c48-account-create-6cz92"] Oct 02 02:03:16 crc kubenswrapper[4885]: I1002 02:03:16.299578 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5pxm2\" (UniqueName: \"kubernetes.io/projected/32627c09-5c38-4a0b-90a6-d6c665654402-kube-api-access-5pxm2\") pod \"cinder-fe7d-account-create-qtl4l\" (UID: \"32627c09-5c38-4a0b-90a6-d6c665654402\") " pod="openstack/cinder-fe7d-account-create-qtl4l" Oct 02 02:03:16 crc kubenswrapper[4885]: I1002 02:03:16.400946 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5pxm2\" (UniqueName: \"kubernetes.io/projected/32627c09-5c38-4a0b-90a6-d6c665654402-kube-api-access-5pxm2\") pod \"cinder-fe7d-account-create-qtl4l\" (UID: \"32627c09-5c38-4a0b-90a6-d6c665654402\") " pod="openstack/cinder-fe7d-account-create-qtl4l" Oct 02 02:03:16 crc kubenswrapper[4885]: I1002 02:03:16.401118 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sxb68\" (UniqueName: \"kubernetes.io/projected/ef4d5777-35ad-4551-a2df-28af567ebeb5-kube-api-access-sxb68\") pod \"barbican-1c48-account-create-6cz92\" (UID: \"ef4d5777-35ad-4551-a2df-28af567ebeb5\") " pod="openstack/barbican-1c48-account-create-6cz92" Oct 02 02:03:16 crc kubenswrapper[4885]: I1002 02:03:16.430270 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-f842-account-create-fsxd8"] Oct 02 02:03:16 crc kubenswrapper[4885]: I1002 02:03:16.431235 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-f842-account-create-fsxd8" Oct 02 02:03:16 crc kubenswrapper[4885]: I1002 02:03:16.436885 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Oct 02 02:03:16 crc kubenswrapper[4885]: I1002 02:03:16.439097 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5pxm2\" (UniqueName: \"kubernetes.io/projected/32627c09-5c38-4a0b-90a6-d6c665654402-kube-api-access-5pxm2\") pod \"cinder-fe7d-account-create-qtl4l\" (UID: \"32627c09-5c38-4a0b-90a6-d6c665654402\") " pod="openstack/cinder-fe7d-account-create-qtl4l" Oct 02 02:03:16 crc kubenswrapper[4885]: I1002 02:03:16.451321 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-f842-account-create-fsxd8"] Oct 02 02:03:16 crc kubenswrapper[4885]: I1002 02:03:16.467902 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-fe7d-account-create-qtl4l" Oct 02 02:03:16 crc kubenswrapper[4885]: I1002 02:03:16.502362 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sxb68\" (UniqueName: \"kubernetes.io/projected/ef4d5777-35ad-4551-a2df-28af567ebeb5-kube-api-access-sxb68\") pod \"barbican-1c48-account-create-6cz92\" (UID: \"ef4d5777-35ad-4551-a2df-28af567ebeb5\") " pod="openstack/barbican-1c48-account-create-6cz92" Oct 02 02:03:16 crc kubenswrapper[4885]: I1002 02:03:16.521958 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sxb68\" (UniqueName: \"kubernetes.io/projected/ef4d5777-35ad-4551-a2df-28af567ebeb5-kube-api-access-sxb68\") pod \"barbican-1c48-account-create-6cz92\" (UID: \"ef4d5777-35ad-4551-a2df-28af567ebeb5\") " pod="openstack/barbican-1c48-account-create-6cz92" Oct 02 02:03:16 crc kubenswrapper[4885]: I1002 02:03:16.565858 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-1c48-account-create-6cz92" Oct 02 02:03:16 crc kubenswrapper[4885]: I1002 02:03:16.603966 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xwbjs\" (UniqueName: \"kubernetes.io/projected/2147f705-82c1-4cce-98c8-85c7aee85542-kube-api-access-xwbjs\") pod \"neutron-f842-account-create-fsxd8\" (UID: \"2147f705-82c1-4cce-98c8-85c7aee85542\") " pod="openstack/neutron-f842-account-create-fsxd8" Oct 02 02:03:16 crc kubenswrapper[4885]: I1002 02:03:16.706057 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xwbjs\" (UniqueName: \"kubernetes.io/projected/2147f705-82c1-4cce-98c8-85c7aee85542-kube-api-access-xwbjs\") pod \"neutron-f842-account-create-fsxd8\" (UID: \"2147f705-82c1-4cce-98c8-85c7aee85542\") " pod="openstack/neutron-f842-account-create-fsxd8" Oct 02 02:03:16 crc kubenswrapper[4885]: I1002 02:03:16.724079 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xwbjs\" (UniqueName: \"kubernetes.io/projected/2147f705-82c1-4cce-98c8-85c7aee85542-kube-api-access-xwbjs\") pod \"neutron-f842-account-create-fsxd8\" (UID: \"2147f705-82c1-4cce-98c8-85c7aee85542\") " pod="openstack/neutron-f842-account-create-fsxd8" Oct 02 02:03:16 crc kubenswrapper[4885]: I1002 02:03:16.785614 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-f842-account-create-fsxd8" Oct 02 02:03:19 crc kubenswrapper[4885]: I1002 02:03:19.453364 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-f842-account-create-fsxd8"] Oct 02 02:03:19 crc kubenswrapper[4885]: I1002 02:03:19.515724 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-1c48-account-create-6cz92"] Oct 02 02:03:19 crc kubenswrapper[4885]: I1002 02:03:19.530640 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-fe7d-account-create-qtl4l"] Oct 02 02:03:19 crc kubenswrapper[4885]: I1002 02:03:19.619168 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-1c48-account-create-6cz92" event={"ID":"ef4d5777-35ad-4551-a2df-28af567ebeb5","Type":"ContainerStarted","Data":"499a0278f96f0441fb1ef801097f84f04cc5293cadacabf9b83cbb21794a94db"} Oct 02 02:03:19 crc kubenswrapper[4885]: I1002 02:03:19.625993 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" event={"ID":"d169c8ba-000a-4218-8e3b-4ae53035b110","Type":"ContainerStarted","Data":"b3432e8413cbb9773cfd116ef5983ce1a5243df994d7ca0829906833f1dfa903"} Oct 02 02:03:19 crc kubenswrapper[4885]: I1002 02:03:19.635962 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0c8d183a-f48b-4a69-8685-90c834e678ce","Type":"ContainerStarted","Data":"a9459deb98811483b51431a772c0384e03007c3a1241258e1d0ec6a9bcae57cf"} Oct 02 02:03:19 crc kubenswrapper[4885]: I1002 02:03:19.636014 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0c8d183a-f48b-4a69-8685-90c834e678ce","Type":"ContainerStarted","Data":"9e7cdd1b06a286245fc9ffb1414d4c4f8183081f939516dd716586c6568f8867"} Oct 02 02:03:19 crc kubenswrapper[4885]: I1002 02:03:19.636024 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0c8d183a-f48b-4a69-8685-90c834e678ce","Type":"ContainerStarted","Data":"61191e5ce808da69b5b4895df8e17cd77f8e2fd6731d0485238dee6a37b5bc8e"} Oct 02 02:03:19 crc kubenswrapper[4885]: I1002 02:03:19.643796 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-tvjc2" event={"ID":"d0092499-91ac-4e53-8c91-04d30bd49e68","Type":"ContainerStarted","Data":"4f1562bee17ffaa4f7203145677f0398a269b28de57c734ed5c2f625b10afe1c"} Oct 02 02:03:19 crc kubenswrapper[4885]: I1002 02:03:19.651770 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-fe7d-account-create-qtl4l" event={"ID":"32627c09-5c38-4a0b-90a6-d6c665654402","Type":"ContainerStarted","Data":"7c64d02a1a3b8fa513a8d009cf35936ec7ea112a9a02f86645c41f7c9869aeb7"} Oct 02 02:03:19 crc kubenswrapper[4885]: I1002 02:03:19.657865 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-f842-account-create-fsxd8" event={"ID":"2147f705-82c1-4cce-98c8-85c7aee85542","Type":"ContainerStarted","Data":"5d827e2e9b465734e5af7d62b99adc565007a089d3d9076c8bdd1f68e4e5e879"} Oct 02 02:03:19 crc kubenswrapper[4885]: I1002 02:03:19.669873 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-tvjc2" podStartSLOduration=2.149256911 podStartE2EDuration="13.669857097s" podCreationTimestamp="2025-10-02 02:03:06 +0000 UTC" firstStartedPulling="2025-10-02 02:03:07.40759218 +0000 UTC m=+976.219339579" lastFinishedPulling="2025-10-02 02:03:18.928192326 +0000 UTC m=+987.739939765" observedRunningTime="2025-10-02 02:03:19.668820095 +0000 UTC m=+988.480567484" watchObservedRunningTime="2025-10-02 02:03:19.669857097 +0000 UTC m=+988.481604496" Oct 02 02:03:19 crc kubenswrapper[4885]: I1002 02:03:19.685097 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-f842-account-create-fsxd8" podStartSLOduration=3.685086309 podStartE2EDuration="3.685086309s" podCreationTimestamp="2025-10-02 02:03:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:03:19.67886592 +0000 UTC m=+988.490613359" watchObservedRunningTime="2025-10-02 02:03:19.685086309 +0000 UTC m=+988.496833708" Oct 02 02:03:20 crc kubenswrapper[4885]: I1002 02:03:20.680661 4885 generic.go:334] "Generic (PLEG): container finished" podID="32627c09-5c38-4a0b-90a6-d6c665654402" containerID="29d8b0d4bf1525e7f40bd8accd471c2f106dfae1d0fb0eb9c3ccac5d65c6ee07" exitCode=0 Oct 02 02:03:20 crc kubenswrapper[4885]: I1002 02:03:20.680922 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-fe7d-account-create-qtl4l" event={"ID":"32627c09-5c38-4a0b-90a6-d6c665654402","Type":"ContainerDied","Data":"29d8b0d4bf1525e7f40bd8accd471c2f106dfae1d0fb0eb9c3ccac5d65c6ee07"} Oct 02 02:03:20 crc kubenswrapper[4885]: I1002 02:03:20.688064 4885 generic.go:334] "Generic (PLEG): container finished" podID="2147f705-82c1-4cce-98c8-85c7aee85542" containerID="2673effa29be2d50feab73fa410b48c8a801fdbb1e2991b20d9fd88761ccdee5" exitCode=0 Oct 02 02:03:20 crc kubenswrapper[4885]: I1002 02:03:20.688176 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-f842-account-create-fsxd8" event={"ID":"2147f705-82c1-4cce-98c8-85c7aee85542","Type":"ContainerDied","Data":"2673effa29be2d50feab73fa410b48c8a801fdbb1e2991b20d9fd88761ccdee5"} Oct 02 02:03:20 crc kubenswrapper[4885]: I1002 02:03:20.693414 4885 generic.go:334] "Generic (PLEG): container finished" podID="ef4d5777-35ad-4551-a2df-28af567ebeb5" containerID="d1c29d0f0646e6b585de70535fb0505faeed21ff4aac88795e86f26e0758e872" exitCode=0 Oct 02 02:03:20 crc kubenswrapper[4885]: I1002 02:03:20.693487 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-1c48-account-create-6cz92" event={"ID":"ef4d5777-35ad-4551-a2df-28af567ebeb5","Type":"ContainerDied","Data":"d1c29d0f0646e6b585de70535fb0505faeed21ff4aac88795e86f26e0758e872"} Oct 02 02:03:20 crc kubenswrapper[4885]: I1002 02:03:20.698478 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-spxxt" event={"ID":"6a9ed972-b8c7-4ff7-af7b-899842ecc19c","Type":"ContainerStarted","Data":"8d716973aace109fb672bfa98c5dffccc3dd5fa5f26180d6b4abf24b760c199c"} Oct 02 02:03:20 crc kubenswrapper[4885]: I1002 02:03:20.716323 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0c8d183a-f48b-4a69-8685-90c834e678ce","Type":"ContainerStarted","Data":"878a0e733522a9a1a8d25f03558346dcccaea0b68b2489cf1635f1a96544f3ec"} Oct 02 02:03:20 crc kubenswrapper[4885]: I1002 02:03:20.716363 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0c8d183a-f48b-4a69-8685-90c834e678ce","Type":"ContainerStarted","Data":"16509ae4d140e50f5a4e8e86ea820221efdfacafc965ac27ce69a769a9c5ebbe"} Oct 02 02:03:20 crc kubenswrapper[4885]: I1002 02:03:20.722506 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-spxxt" podStartSLOduration=2.9386753949999997 podStartE2EDuration="17.722486578s" podCreationTimestamp="2025-10-02 02:03:03 +0000 UTC" firstStartedPulling="2025-10-02 02:03:04.261305286 +0000 UTC m=+973.073052695" lastFinishedPulling="2025-10-02 02:03:19.045116469 +0000 UTC m=+987.856863878" observedRunningTime="2025-10-02 02:03:20.718674572 +0000 UTC m=+989.530421971" watchObservedRunningTime="2025-10-02 02:03:20.722486578 +0000 UTC m=+989.534233987" Oct 02 02:03:20 crc kubenswrapper[4885]: I1002 02:03:20.798564 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=46.001959922 podStartE2EDuration="51.798542493s" podCreationTimestamp="2025-10-02 02:02:29 +0000 UTC" firstStartedPulling="2025-10-02 02:03:03.40231679 +0000 UTC m=+972.214064189" lastFinishedPulling="2025-10-02 02:03:09.198899351 +0000 UTC m=+978.010646760" observedRunningTime="2025-10-02 02:03:20.792584832 +0000 UTC m=+989.604332311" watchObservedRunningTime="2025-10-02 02:03:20.798542493 +0000 UTC m=+989.610289902" Oct 02 02:03:21 crc kubenswrapper[4885]: I1002 02:03:21.062413 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-x2nx7"] Oct 02 02:03:21 crc kubenswrapper[4885]: I1002 02:03:21.065802 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77585f5f8c-x2nx7" Oct 02 02:03:21 crc kubenswrapper[4885]: I1002 02:03:21.071007 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Oct 02 02:03:21 crc kubenswrapper[4885]: I1002 02:03:21.109145 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9fc15c77-b663-4395-ae92-922b1e4756ba-dns-svc\") pod \"dnsmasq-dns-77585f5f8c-x2nx7\" (UID: \"9fc15c77-b663-4395-ae92-922b1e4756ba\") " pod="openstack/dnsmasq-dns-77585f5f8c-x2nx7" Oct 02 02:03:21 crc kubenswrapper[4885]: I1002 02:03:21.109221 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ht7d9\" (UniqueName: \"kubernetes.io/projected/9fc15c77-b663-4395-ae92-922b1e4756ba-kube-api-access-ht7d9\") pod \"dnsmasq-dns-77585f5f8c-x2nx7\" (UID: \"9fc15c77-b663-4395-ae92-922b1e4756ba\") " pod="openstack/dnsmasq-dns-77585f5f8c-x2nx7" Oct 02 02:03:21 crc kubenswrapper[4885]: I1002 02:03:21.109277 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9fc15c77-b663-4395-ae92-922b1e4756ba-ovsdbserver-sb\") pod \"dnsmasq-dns-77585f5f8c-x2nx7\" (UID: \"9fc15c77-b663-4395-ae92-922b1e4756ba\") " pod="openstack/dnsmasq-dns-77585f5f8c-x2nx7" Oct 02 02:03:21 crc kubenswrapper[4885]: I1002 02:03:21.109313 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9fc15c77-b663-4395-ae92-922b1e4756ba-dns-swift-storage-0\") pod \"dnsmasq-dns-77585f5f8c-x2nx7\" (UID: \"9fc15c77-b663-4395-ae92-922b1e4756ba\") " pod="openstack/dnsmasq-dns-77585f5f8c-x2nx7" Oct 02 02:03:21 crc kubenswrapper[4885]: I1002 02:03:21.110282 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9fc15c77-b663-4395-ae92-922b1e4756ba-ovsdbserver-nb\") pod \"dnsmasq-dns-77585f5f8c-x2nx7\" (UID: \"9fc15c77-b663-4395-ae92-922b1e4756ba\") " pod="openstack/dnsmasq-dns-77585f5f8c-x2nx7" Oct 02 02:03:21 crc kubenswrapper[4885]: I1002 02:03:21.110543 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9fc15c77-b663-4395-ae92-922b1e4756ba-config\") pod \"dnsmasq-dns-77585f5f8c-x2nx7\" (UID: \"9fc15c77-b663-4395-ae92-922b1e4756ba\") " pod="openstack/dnsmasq-dns-77585f5f8c-x2nx7" Oct 02 02:03:21 crc kubenswrapper[4885]: I1002 02:03:21.121587 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-x2nx7"] Oct 02 02:03:21 crc kubenswrapper[4885]: I1002 02:03:21.212368 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9fc15c77-b663-4395-ae92-922b1e4756ba-dns-svc\") pod \"dnsmasq-dns-77585f5f8c-x2nx7\" (UID: \"9fc15c77-b663-4395-ae92-922b1e4756ba\") " pod="openstack/dnsmasq-dns-77585f5f8c-x2nx7" Oct 02 02:03:21 crc kubenswrapper[4885]: I1002 02:03:21.212435 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ht7d9\" (UniqueName: \"kubernetes.io/projected/9fc15c77-b663-4395-ae92-922b1e4756ba-kube-api-access-ht7d9\") pod \"dnsmasq-dns-77585f5f8c-x2nx7\" (UID: \"9fc15c77-b663-4395-ae92-922b1e4756ba\") " pod="openstack/dnsmasq-dns-77585f5f8c-x2nx7" Oct 02 02:03:21 crc kubenswrapper[4885]: I1002 02:03:21.212465 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9fc15c77-b663-4395-ae92-922b1e4756ba-ovsdbserver-sb\") pod \"dnsmasq-dns-77585f5f8c-x2nx7\" (UID: \"9fc15c77-b663-4395-ae92-922b1e4756ba\") " pod="openstack/dnsmasq-dns-77585f5f8c-x2nx7" Oct 02 02:03:21 crc kubenswrapper[4885]: I1002 02:03:21.212487 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9fc15c77-b663-4395-ae92-922b1e4756ba-dns-swift-storage-0\") pod \"dnsmasq-dns-77585f5f8c-x2nx7\" (UID: \"9fc15c77-b663-4395-ae92-922b1e4756ba\") " pod="openstack/dnsmasq-dns-77585f5f8c-x2nx7" Oct 02 02:03:21 crc kubenswrapper[4885]: I1002 02:03:21.212523 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9fc15c77-b663-4395-ae92-922b1e4756ba-ovsdbserver-nb\") pod \"dnsmasq-dns-77585f5f8c-x2nx7\" (UID: \"9fc15c77-b663-4395-ae92-922b1e4756ba\") " pod="openstack/dnsmasq-dns-77585f5f8c-x2nx7" Oct 02 02:03:21 crc kubenswrapper[4885]: I1002 02:03:21.212550 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9fc15c77-b663-4395-ae92-922b1e4756ba-config\") pod \"dnsmasq-dns-77585f5f8c-x2nx7\" (UID: \"9fc15c77-b663-4395-ae92-922b1e4756ba\") " pod="openstack/dnsmasq-dns-77585f5f8c-x2nx7" Oct 02 02:03:21 crc kubenswrapper[4885]: I1002 02:03:21.213194 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9fc15c77-b663-4395-ae92-922b1e4756ba-dns-svc\") pod \"dnsmasq-dns-77585f5f8c-x2nx7\" (UID: \"9fc15c77-b663-4395-ae92-922b1e4756ba\") " pod="openstack/dnsmasq-dns-77585f5f8c-x2nx7" Oct 02 02:03:21 crc kubenswrapper[4885]: I1002 02:03:21.213300 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9fc15c77-b663-4395-ae92-922b1e4756ba-config\") pod \"dnsmasq-dns-77585f5f8c-x2nx7\" (UID: \"9fc15c77-b663-4395-ae92-922b1e4756ba\") " pod="openstack/dnsmasq-dns-77585f5f8c-x2nx7" Oct 02 02:03:21 crc kubenswrapper[4885]: I1002 02:03:21.213755 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9fc15c77-b663-4395-ae92-922b1e4756ba-ovsdbserver-sb\") pod \"dnsmasq-dns-77585f5f8c-x2nx7\" (UID: \"9fc15c77-b663-4395-ae92-922b1e4756ba\") " pod="openstack/dnsmasq-dns-77585f5f8c-x2nx7" Oct 02 02:03:21 crc kubenswrapper[4885]: I1002 02:03:21.217514 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9fc15c77-b663-4395-ae92-922b1e4756ba-dns-swift-storage-0\") pod \"dnsmasq-dns-77585f5f8c-x2nx7\" (UID: \"9fc15c77-b663-4395-ae92-922b1e4756ba\") " pod="openstack/dnsmasq-dns-77585f5f8c-x2nx7" Oct 02 02:03:21 crc kubenswrapper[4885]: I1002 02:03:21.218102 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9fc15c77-b663-4395-ae92-922b1e4756ba-ovsdbserver-nb\") pod \"dnsmasq-dns-77585f5f8c-x2nx7\" (UID: \"9fc15c77-b663-4395-ae92-922b1e4756ba\") " pod="openstack/dnsmasq-dns-77585f5f8c-x2nx7" Oct 02 02:03:21 crc kubenswrapper[4885]: I1002 02:03:21.233348 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ht7d9\" (UniqueName: \"kubernetes.io/projected/9fc15c77-b663-4395-ae92-922b1e4756ba-kube-api-access-ht7d9\") pod \"dnsmasq-dns-77585f5f8c-x2nx7\" (UID: \"9fc15c77-b663-4395-ae92-922b1e4756ba\") " pod="openstack/dnsmasq-dns-77585f5f8c-x2nx7" Oct 02 02:03:21 crc kubenswrapper[4885]: I1002 02:03:21.405377 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77585f5f8c-x2nx7" Oct 02 02:03:32 crc kubenswrapper[4885]: I1002 02:03:21.947612 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-x2nx7"] Oct 02 02:03:32 crc kubenswrapper[4885]: W1002 02:03:21.949598 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9fc15c77_b663_4395_ae92_922b1e4756ba.slice/crio-859f4908b20a9cf03af8c87905ea30ffa64e16851d1dade87d6d05b7f1599118 WatchSource:0}: Error finding container 859f4908b20a9cf03af8c87905ea30ffa64e16851d1dade87d6d05b7f1599118: Status 404 returned error can't find the container with id 859f4908b20a9cf03af8c87905ea30ffa64e16851d1dade87d6d05b7f1599118 Oct 02 02:03:32 crc kubenswrapper[4885]: I1002 02:03:22.740856 4885 generic.go:334] "Generic (PLEG): container finished" podID="9fc15c77-b663-4395-ae92-922b1e4756ba" containerID="fe1dec9f2387d5c95aec1c710183ab2006a76467906b5a9001f650f6227ed39d" exitCode=0 Oct 02 02:03:32 crc kubenswrapper[4885]: I1002 02:03:22.740894 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-x2nx7" event={"ID":"9fc15c77-b663-4395-ae92-922b1e4756ba","Type":"ContainerDied","Data":"fe1dec9f2387d5c95aec1c710183ab2006a76467906b5a9001f650f6227ed39d"} Oct 02 02:03:32 crc kubenswrapper[4885]: I1002 02:03:22.740916 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-x2nx7" event={"ID":"9fc15c77-b663-4395-ae92-922b1e4756ba","Type":"ContainerStarted","Data":"859f4908b20a9cf03af8c87905ea30ffa64e16851d1dade87d6d05b7f1599118"} Oct 02 02:03:32 crc kubenswrapper[4885]: I1002 02:03:32.724401 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-fe7d-account-create-qtl4l" Oct 02 02:03:32 crc kubenswrapper[4885]: I1002 02:03:32.731085 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-1c48-account-create-6cz92" Oct 02 02:03:32 crc kubenswrapper[4885]: I1002 02:03:32.745541 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-f842-account-create-fsxd8" Oct 02 02:03:32 crc kubenswrapper[4885]: I1002 02:03:32.753970 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5pxm2\" (UniqueName: \"kubernetes.io/projected/32627c09-5c38-4a0b-90a6-d6c665654402-kube-api-access-5pxm2\") pod \"32627c09-5c38-4a0b-90a6-d6c665654402\" (UID: \"32627c09-5c38-4a0b-90a6-d6c665654402\") " Oct 02 02:03:32 crc kubenswrapper[4885]: I1002 02:03:32.754190 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xwbjs\" (UniqueName: \"kubernetes.io/projected/2147f705-82c1-4cce-98c8-85c7aee85542-kube-api-access-xwbjs\") pod \"2147f705-82c1-4cce-98c8-85c7aee85542\" (UID: \"2147f705-82c1-4cce-98c8-85c7aee85542\") " Oct 02 02:03:32 crc kubenswrapper[4885]: I1002 02:03:32.754238 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sxb68\" (UniqueName: \"kubernetes.io/projected/ef4d5777-35ad-4551-a2df-28af567ebeb5-kube-api-access-sxb68\") pod \"ef4d5777-35ad-4551-a2df-28af567ebeb5\" (UID: \"ef4d5777-35ad-4551-a2df-28af567ebeb5\") " Oct 02 02:03:32 crc kubenswrapper[4885]: I1002 02:03:32.762737 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef4d5777-35ad-4551-a2df-28af567ebeb5-kube-api-access-sxb68" (OuterVolumeSpecName: "kube-api-access-sxb68") pod "ef4d5777-35ad-4551-a2df-28af567ebeb5" (UID: "ef4d5777-35ad-4551-a2df-28af567ebeb5"). InnerVolumeSpecName "kube-api-access-sxb68". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:03:32 crc kubenswrapper[4885]: I1002 02:03:32.764452 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2147f705-82c1-4cce-98c8-85c7aee85542-kube-api-access-xwbjs" (OuterVolumeSpecName: "kube-api-access-xwbjs") pod "2147f705-82c1-4cce-98c8-85c7aee85542" (UID: "2147f705-82c1-4cce-98c8-85c7aee85542"). InnerVolumeSpecName "kube-api-access-xwbjs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:03:32 crc kubenswrapper[4885]: I1002 02:03:32.777406 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32627c09-5c38-4a0b-90a6-d6c665654402-kube-api-access-5pxm2" (OuterVolumeSpecName: "kube-api-access-5pxm2") pod "32627c09-5c38-4a0b-90a6-d6c665654402" (UID: "32627c09-5c38-4a0b-90a6-d6c665654402"). InnerVolumeSpecName "kube-api-access-5pxm2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:03:32 crc kubenswrapper[4885]: I1002 02:03:32.858746 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xwbjs\" (UniqueName: \"kubernetes.io/projected/2147f705-82c1-4cce-98c8-85c7aee85542-kube-api-access-xwbjs\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:32 crc kubenswrapper[4885]: I1002 02:03:32.858776 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sxb68\" (UniqueName: \"kubernetes.io/projected/ef4d5777-35ad-4551-a2df-28af567ebeb5-kube-api-access-sxb68\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:32 crc kubenswrapper[4885]: I1002 02:03:32.858791 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5pxm2\" (UniqueName: \"kubernetes.io/projected/32627c09-5c38-4a0b-90a6-d6c665654402-kube-api-access-5pxm2\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:32 crc kubenswrapper[4885]: I1002 02:03:32.858876 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-f842-account-create-fsxd8" Oct 02 02:03:32 crc kubenswrapper[4885]: I1002 02:03:32.858920 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-f842-account-create-fsxd8" event={"ID":"2147f705-82c1-4cce-98c8-85c7aee85542","Type":"ContainerDied","Data":"5d827e2e9b465734e5af7d62b99adc565007a089d3d9076c8bdd1f68e4e5e879"} Oct 02 02:03:32 crc kubenswrapper[4885]: I1002 02:03:32.858960 4885 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5d827e2e9b465734e5af7d62b99adc565007a089d3d9076c8bdd1f68e4e5e879" Oct 02 02:03:32 crc kubenswrapper[4885]: I1002 02:03:32.861090 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-1c48-account-create-6cz92" Oct 02 02:03:32 crc kubenswrapper[4885]: I1002 02:03:32.861103 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-1c48-account-create-6cz92" event={"ID":"ef4d5777-35ad-4551-a2df-28af567ebeb5","Type":"ContainerDied","Data":"499a0278f96f0441fb1ef801097f84f04cc5293cadacabf9b83cbb21794a94db"} Oct 02 02:03:32 crc kubenswrapper[4885]: I1002 02:03:32.861307 4885 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="499a0278f96f0441fb1ef801097f84f04cc5293cadacabf9b83cbb21794a94db" Oct 02 02:03:32 crc kubenswrapper[4885]: I1002 02:03:32.867598 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-x2nx7" event={"ID":"9fc15c77-b663-4395-ae92-922b1e4756ba","Type":"ContainerStarted","Data":"4b4ebce37f94f38134bb618a2002d6b8fb4bc1bec4fc68a0d3ab6a914e2c84a0"} Oct 02 02:03:32 crc kubenswrapper[4885]: I1002 02:03:32.873907 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-77585f5f8c-x2nx7" Oct 02 02:03:32 crc kubenswrapper[4885]: I1002 02:03:32.876930 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-fe7d-account-create-qtl4l" event={"ID":"32627c09-5c38-4a0b-90a6-d6c665654402","Type":"ContainerDied","Data":"7c64d02a1a3b8fa513a8d009cf35936ec7ea112a9a02f86645c41f7c9869aeb7"} Oct 02 02:03:32 crc kubenswrapper[4885]: I1002 02:03:32.876995 4885 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7c64d02a1a3b8fa513a8d009cf35936ec7ea112a9a02f86645c41f7c9869aeb7" Oct 02 02:03:32 crc kubenswrapper[4885]: I1002 02:03:32.878633 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-fe7d-account-create-qtl4l" Oct 02 02:03:32 crc kubenswrapper[4885]: I1002 02:03:32.906199 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-77585f5f8c-x2nx7" podStartSLOduration=11.906174215 podStartE2EDuration="11.906174215s" podCreationTimestamp="2025-10-02 02:03:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:03:32.896575953 +0000 UTC m=+1001.708323362" watchObservedRunningTime="2025-10-02 02:03:32.906174215 +0000 UTC m=+1001.717921614" Oct 02 02:03:35 crc kubenswrapper[4885]: I1002 02:03:35.911698 4885 generic.go:334] "Generic (PLEG): container finished" podID="d0092499-91ac-4e53-8c91-04d30bd49e68" containerID="4f1562bee17ffaa4f7203145677f0398a269b28de57c734ed5c2f625b10afe1c" exitCode=0 Oct 02 02:03:35 crc kubenswrapper[4885]: I1002 02:03:35.911851 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-tvjc2" event={"ID":"d0092499-91ac-4e53-8c91-04d30bd49e68","Type":"ContainerDied","Data":"4f1562bee17ffaa4f7203145677f0398a269b28de57c734ed5c2f625b10afe1c"} Oct 02 02:03:37 crc kubenswrapper[4885]: I1002 02:03:37.265658 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-tvjc2" Oct 02 02:03:37 crc kubenswrapper[4885]: I1002 02:03:37.349982 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0092499-91ac-4e53-8c91-04d30bd49e68-combined-ca-bundle\") pod \"d0092499-91ac-4e53-8c91-04d30bd49e68\" (UID: \"d0092499-91ac-4e53-8c91-04d30bd49e68\") " Oct 02 02:03:37 crc kubenswrapper[4885]: I1002 02:03:37.350193 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0092499-91ac-4e53-8c91-04d30bd49e68-config-data\") pod \"d0092499-91ac-4e53-8c91-04d30bd49e68\" (UID: \"d0092499-91ac-4e53-8c91-04d30bd49e68\") " Oct 02 02:03:37 crc kubenswrapper[4885]: I1002 02:03:37.350225 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lvvjp\" (UniqueName: \"kubernetes.io/projected/d0092499-91ac-4e53-8c91-04d30bd49e68-kube-api-access-lvvjp\") pod \"d0092499-91ac-4e53-8c91-04d30bd49e68\" (UID: \"d0092499-91ac-4e53-8c91-04d30bd49e68\") " Oct 02 02:03:37 crc kubenswrapper[4885]: I1002 02:03:37.356212 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d0092499-91ac-4e53-8c91-04d30bd49e68-kube-api-access-lvvjp" (OuterVolumeSpecName: "kube-api-access-lvvjp") pod "d0092499-91ac-4e53-8c91-04d30bd49e68" (UID: "d0092499-91ac-4e53-8c91-04d30bd49e68"). InnerVolumeSpecName "kube-api-access-lvvjp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:03:37 crc kubenswrapper[4885]: I1002 02:03:37.406179 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0092499-91ac-4e53-8c91-04d30bd49e68-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d0092499-91ac-4e53-8c91-04d30bd49e68" (UID: "d0092499-91ac-4e53-8c91-04d30bd49e68"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:37 crc kubenswrapper[4885]: I1002 02:03:37.409590 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0092499-91ac-4e53-8c91-04d30bd49e68-config-data" (OuterVolumeSpecName: "config-data") pod "d0092499-91ac-4e53-8c91-04d30bd49e68" (UID: "d0092499-91ac-4e53-8c91-04d30bd49e68"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:37 crc kubenswrapper[4885]: I1002 02:03:37.451631 4885 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0092499-91ac-4e53-8c91-04d30bd49e68-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:37 crc kubenswrapper[4885]: I1002 02:03:37.451675 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lvvjp\" (UniqueName: \"kubernetes.io/projected/d0092499-91ac-4e53-8c91-04d30bd49e68-kube-api-access-lvvjp\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:37 crc kubenswrapper[4885]: I1002 02:03:37.451690 4885 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0092499-91ac-4e53-8c91-04d30bd49e68-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:37 crc kubenswrapper[4885]: I1002 02:03:37.935901 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-tvjc2" event={"ID":"d0092499-91ac-4e53-8c91-04d30bd49e68","Type":"ContainerDied","Data":"f55777e4b40637ee51de638450c8feb5694e9d731c777be79bc1158d6f020c70"} Oct 02 02:03:37 crc kubenswrapper[4885]: I1002 02:03:37.935959 4885 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f55777e4b40637ee51de638450c8feb5694e9d731c777be79bc1158d6f020c70" Oct 02 02:03:37 crc kubenswrapper[4885]: I1002 02:03:37.936016 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-tvjc2" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.189309 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-x2nx7"] Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.189589 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-77585f5f8c-x2nx7" podUID="9fc15c77-b663-4395-ae92-922b1e4756ba" containerName="dnsmasq-dns" containerID="cri-o://4b4ebce37f94f38134bb618a2002d6b8fb4bc1bec4fc68a0d3ab6a914e2c84a0" gracePeriod=10 Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.195411 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-77585f5f8c-x2nx7" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.233711 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-55fff446b9-84t5x"] Oct 02 02:03:38 crc kubenswrapper[4885]: E1002 02:03:38.234092 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32627c09-5c38-4a0b-90a6-d6c665654402" containerName="mariadb-account-create" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.234106 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="32627c09-5c38-4a0b-90a6-d6c665654402" containerName="mariadb-account-create" Oct 02 02:03:38 crc kubenswrapper[4885]: E1002 02:03:38.234115 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef4d5777-35ad-4551-a2df-28af567ebeb5" containerName="mariadb-account-create" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.234121 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef4d5777-35ad-4551-a2df-28af567ebeb5" containerName="mariadb-account-create" Oct 02 02:03:38 crc kubenswrapper[4885]: E1002 02:03:38.234144 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0092499-91ac-4e53-8c91-04d30bd49e68" containerName="keystone-db-sync" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.234150 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0092499-91ac-4e53-8c91-04d30bd49e68" containerName="keystone-db-sync" Oct 02 02:03:38 crc kubenswrapper[4885]: E1002 02:03:38.234171 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2147f705-82c1-4cce-98c8-85c7aee85542" containerName="mariadb-account-create" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.234179 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="2147f705-82c1-4cce-98c8-85c7aee85542" containerName="mariadb-account-create" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.234343 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef4d5777-35ad-4551-a2df-28af567ebeb5" containerName="mariadb-account-create" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.234362 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0092499-91ac-4e53-8c91-04d30bd49e68" containerName="keystone-db-sync" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.234375 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="2147f705-82c1-4cce-98c8-85c7aee85542" containerName="mariadb-account-create" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.234386 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="32627c09-5c38-4a0b-90a6-d6c665654402" containerName="mariadb-account-create" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.235222 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55fff446b9-84t5x" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.262094 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/afa17394-b736-4acb-bf4f-ca79fc4908ff-dns-svc\") pod \"dnsmasq-dns-55fff446b9-84t5x\" (UID: \"afa17394-b736-4acb-bf4f-ca79fc4908ff\") " pod="openstack/dnsmasq-dns-55fff446b9-84t5x" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.262383 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/afa17394-b736-4acb-bf4f-ca79fc4908ff-ovsdbserver-sb\") pod \"dnsmasq-dns-55fff446b9-84t5x\" (UID: \"afa17394-b736-4acb-bf4f-ca79fc4908ff\") " pod="openstack/dnsmasq-dns-55fff446b9-84t5x" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.262410 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vx58g\" (UniqueName: \"kubernetes.io/projected/afa17394-b736-4acb-bf4f-ca79fc4908ff-kube-api-access-vx58g\") pod \"dnsmasq-dns-55fff446b9-84t5x\" (UID: \"afa17394-b736-4acb-bf4f-ca79fc4908ff\") " pod="openstack/dnsmasq-dns-55fff446b9-84t5x" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.262459 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/afa17394-b736-4acb-bf4f-ca79fc4908ff-dns-swift-storage-0\") pod \"dnsmasq-dns-55fff446b9-84t5x\" (UID: \"afa17394-b736-4acb-bf4f-ca79fc4908ff\") " pod="openstack/dnsmasq-dns-55fff446b9-84t5x" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.262505 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/afa17394-b736-4acb-bf4f-ca79fc4908ff-config\") pod \"dnsmasq-dns-55fff446b9-84t5x\" (UID: \"afa17394-b736-4acb-bf4f-ca79fc4908ff\") " pod="openstack/dnsmasq-dns-55fff446b9-84t5x" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.262530 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/afa17394-b736-4acb-bf4f-ca79fc4908ff-ovsdbserver-nb\") pod \"dnsmasq-dns-55fff446b9-84t5x\" (UID: \"afa17394-b736-4acb-bf4f-ca79fc4908ff\") " pod="openstack/dnsmasq-dns-55fff446b9-84t5x" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.262619 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55fff446b9-84t5x"] Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.280849 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-mj282"] Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.282000 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-mj282" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.285566 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.285875 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.286056 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.286344 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-h4g4g" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.289612 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-mj282"] Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.368171 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9347e7b4-1d53-4a28-9cbd-6a392b8145d8-combined-ca-bundle\") pod \"keystone-bootstrap-mj282\" (UID: \"9347e7b4-1d53-4a28-9cbd-6a392b8145d8\") " pod="openstack/keystone-bootstrap-mj282" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.369080 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/afa17394-b736-4acb-bf4f-ca79fc4908ff-ovsdbserver-nb\") pod \"dnsmasq-dns-55fff446b9-84t5x\" (UID: \"afa17394-b736-4acb-bf4f-ca79fc4908ff\") " pod="openstack/dnsmasq-dns-55fff446b9-84t5x" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.368229 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/afa17394-b736-4acb-bf4f-ca79fc4908ff-ovsdbserver-nb\") pod \"dnsmasq-dns-55fff446b9-84t5x\" (UID: \"afa17394-b736-4acb-bf4f-ca79fc4908ff\") " pod="openstack/dnsmasq-dns-55fff446b9-84t5x" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.369186 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9347e7b4-1d53-4a28-9cbd-6a392b8145d8-fernet-keys\") pod \"keystone-bootstrap-mj282\" (UID: \"9347e7b4-1d53-4a28-9cbd-6a392b8145d8\") " pod="openstack/keystone-bootstrap-mj282" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.369208 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-72dzl\" (UniqueName: \"kubernetes.io/projected/9347e7b4-1d53-4a28-9cbd-6a392b8145d8-kube-api-access-72dzl\") pod \"keystone-bootstrap-mj282\" (UID: \"9347e7b4-1d53-4a28-9cbd-6a392b8145d8\") " pod="openstack/keystone-bootstrap-mj282" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.369308 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/afa17394-b736-4acb-bf4f-ca79fc4908ff-ovsdbserver-sb\") pod \"dnsmasq-dns-55fff446b9-84t5x\" (UID: \"afa17394-b736-4acb-bf4f-ca79fc4908ff\") " pod="openstack/dnsmasq-dns-55fff446b9-84t5x" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.369344 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/afa17394-b736-4acb-bf4f-ca79fc4908ff-dns-svc\") pod \"dnsmasq-dns-55fff446b9-84t5x\" (UID: \"afa17394-b736-4acb-bf4f-ca79fc4908ff\") " pod="openstack/dnsmasq-dns-55fff446b9-84t5x" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.369932 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/afa17394-b736-4acb-bf4f-ca79fc4908ff-ovsdbserver-sb\") pod \"dnsmasq-dns-55fff446b9-84t5x\" (UID: \"afa17394-b736-4acb-bf4f-ca79fc4908ff\") " pod="openstack/dnsmasq-dns-55fff446b9-84t5x" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.369965 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vx58g\" (UniqueName: \"kubernetes.io/projected/afa17394-b736-4acb-bf4f-ca79fc4908ff-kube-api-access-vx58g\") pod \"dnsmasq-dns-55fff446b9-84t5x\" (UID: \"afa17394-b736-4acb-bf4f-ca79fc4908ff\") " pod="openstack/dnsmasq-dns-55fff446b9-84t5x" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.370024 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9347e7b4-1d53-4a28-9cbd-6a392b8145d8-scripts\") pod \"keystone-bootstrap-mj282\" (UID: \"9347e7b4-1d53-4a28-9cbd-6a392b8145d8\") " pod="openstack/keystone-bootstrap-mj282" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.370061 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/afa17394-b736-4acb-bf4f-ca79fc4908ff-dns-swift-storage-0\") pod \"dnsmasq-dns-55fff446b9-84t5x\" (UID: \"afa17394-b736-4acb-bf4f-ca79fc4908ff\") " pod="openstack/dnsmasq-dns-55fff446b9-84t5x" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.370131 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9347e7b4-1d53-4a28-9cbd-6a392b8145d8-config-data\") pod \"keystone-bootstrap-mj282\" (UID: \"9347e7b4-1d53-4a28-9cbd-6a392b8145d8\") " pod="openstack/keystone-bootstrap-mj282" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.370429 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/afa17394-b736-4acb-bf4f-ca79fc4908ff-dns-svc\") pod \"dnsmasq-dns-55fff446b9-84t5x\" (UID: \"afa17394-b736-4acb-bf4f-ca79fc4908ff\") " pod="openstack/dnsmasq-dns-55fff446b9-84t5x" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.370751 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/afa17394-b736-4acb-bf4f-ca79fc4908ff-dns-swift-storage-0\") pod \"dnsmasq-dns-55fff446b9-84t5x\" (UID: \"afa17394-b736-4acb-bf4f-ca79fc4908ff\") " pod="openstack/dnsmasq-dns-55fff446b9-84t5x" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.370818 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9347e7b4-1d53-4a28-9cbd-6a392b8145d8-credential-keys\") pod \"keystone-bootstrap-mj282\" (UID: \"9347e7b4-1d53-4a28-9cbd-6a392b8145d8\") " pod="openstack/keystone-bootstrap-mj282" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.370841 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/afa17394-b736-4acb-bf4f-ca79fc4908ff-config\") pod \"dnsmasq-dns-55fff446b9-84t5x\" (UID: \"afa17394-b736-4acb-bf4f-ca79fc4908ff\") " pod="openstack/dnsmasq-dns-55fff446b9-84t5x" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.371462 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/afa17394-b736-4acb-bf4f-ca79fc4908ff-config\") pod \"dnsmasq-dns-55fff446b9-84t5x\" (UID: \"afa17394-b736-4acb-bf4f-ca79fc4908ff\") " pod="openstack/dnsmasq-dns-55fff446b9-84t5x" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.387461 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-79659d7df7-6dkj8"] Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.388704 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-79659d7df7-6dkj8" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.394497 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vx58g\" (UniqueName: \"kubernetes.io/projected/afa17394-b736-4acb-bf4f-ca79fc4908ff-kube-api-access-vx58g\") pod \"dnsmasq-dns-55fff446b9-84t5x\" (UID: \"afa17394-b736-4acb-bf4f-ca79fc4908ff\") " pod="openstack/dnsmasq-dns-55fff446b9-84t5x" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.396161 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.402117 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.420626 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.420899 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-zwt4w" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.421005 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-79659d7df7-6dkj8"] Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.477525 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9347e7b4-1d53-4a28-9cbd-6a392b8145d8-scripts\") pod \"keystone-bootstrap-mj282\" (UID: \"9347e7b4-1d53-4a28-9cbd-6a392b8145d8\") " pod="openstack/keystone-bootstrap-mj282" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.477586 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/35248aca-4932-408d-a758-766cce4ae517-scripts\") pod \"horizon-79659d7df7-6dkj8\" (UID: \"35248aca-4932-408d-a758-766cce4ae517\") " pod="openstack/horizon-79659d7df7-6dkj8" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.477621 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9347e7b4-1d53-4a28-9cbd-6a392b8145d8-config-data\") pod \"keystone-bootstrap-mj282\" (UID: \"9347e7b4-1d53-4a28-9cbd-6a392b8145d8\") " pod="openstack/keystone-bootstrap-mj282" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.477644 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/35248aca-4932-408d-a758-766cce4ae517-logs\") pod \"horizon-79659d7df7-6dkj8\" (UID: \"35248aca-4932-408d-a758-766cce4ae517\") " pod="openstack/horizon-79659d7df7-6dkj8" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.477662 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tm2rl\" (UniqueName: \"kubernetes.io/projected/35248aca-4932-408d-a758-766cce4ae517-kube-api-access-tm2rl\") pod \"horizon-79659d7df7-6dkj8\" (UID: \"35248aca-4932-408d-a758-766cce4ae517\") " pod="openstack/horizon-79659d7df7-6dkj8" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.477685 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9347e7b4-1d53-4a28-9cbd-6a392b8145d8-credential-keys\") pod \"keystone-bootstrap-mj282\" (UID: \"9347e7b4-1d53-4a28-9cbd-6a392b8145d8\") " pod="openstack/keystone-bootstrap-mj282" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.477707 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/35248aca-4932-408d-a758-766cce4ae517-horizon-secret-key\") pod \"horizon-79659d7df7-6dkj8\" (UID: \"35248aca-4932-408d-a758-766cce4ae517\") " pod="openstack/horizon-79659d7df7-6dkj8" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.477725 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9347e7b4-1d53-4a28-9cbd-6a392b8145d8-combined-ca-bundle\") pod \"keystone-bootstrap-mj282\" (UID: \"9347e7b4-1d53-4a28-9cbd-6a392b8145d8\") " pod="openstack/keystone-bootstrap-mj282" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.477740 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/35248aca-4932-408d-a758-766cce4ae517-config-data\") pod \"horizon-79659d7df7-6dkj8\" (UID: \"35248aca-4932-408d-a758-766cce4ae517\") " pod="openstack/horizon-79659d7df7-6dkj8" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.477775 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9347e7b4-1d53-4a28-9cbd-6a392b8145d8-fernet-keys\") pod \"keystone-bootstrap-mj282\" (UID: \"9347e7b4-1d53-4a28-9cbd-6a392b8145d8\") " pod="openstack/keystone-bootstrap-mj282" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.477793 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-72dzl\" (UniqueName: \"kubernetes.io/projected/9347e7b4-1d53-4a28-9cbd-6a392b8145d8-kube-api-access-72dzl\") pod \"keystone-bootstrap-mj282\" (UID: \"9347e7b4-1d53-4a28-9cbd-6a392b8145d8\") " pod="openstack/keystone-bootstrap-mj282" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.485767 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9347e7b4-1d53-4a28-9cbd-6a392b8145d8-credential-keys\") pod \"keystone-bootstrap-mj282\" (UID: \"9347e7b4-1d53-4a28-9cbd-6a392b8145d8\") " pod="openstack/keystone-bootstrap-mj282" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.498493 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9347e7b4-1d53-4a28-9cbd-6a392b8145d8-scripts\") pod \"keystone-bootstrap-mj282\" (UID: \"9347e7b4-1d53-4a28-9cbd-6a392b8145d8\") " pod="openstack/keystone-bootstrap-mj282" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.504913 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9347e7b4-1d53-4a28-9cbd-6a392b8145d8-fernet-keys\") pod \"keystone-bootstrap-mj282\" (UID: \"9347e7b4-1d53-4a28-9cbd-6a392b8145d8\") " pod="openstack/keystone-bootstrap-mj282" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.515860 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9347e7b4-1d53-4a28-9cbd-6a392b8145d8-combined-ca-bundle\") pod \"keystone-bootstrap-mj282\" (UID: \"9347e7b4-1d53-4a28-9cbd-6a392b8145d8\") " pod="openstack/keystone-bootstrap-mj282" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.517468 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9347e7b4-1d53-4a28-9cbd-6a392b8145d8-config-data\") pod \"keystone-bootstrap-mj282\" (UID: \"9347e7b4-1d53-4a28-9cbd-6a392b8145d8\") " pod="openstack/keystone-bootstrap-mj282" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.531962 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-72dzl\" (UniqueName: \"kubernetes.io/projected/9347e7b4-1d53-4a28-9cbd-6a392b8145d8-kube-api-access-72dzl\") pod \"keystone-bootstrap-mj282\" (UID: \"9347e7b4-1d53-4a28-9cbd-6a392b8145d8\") " pod="openstack/keystone-bootstrap-mj282" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.567710 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-xf24l"] Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.568754 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-xf24l" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.581665 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-ft6t6" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.581914 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.582043 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.582991 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/35248aca-4932-408d-a758-766cce4ae517-scripts\") pod \"horizon-79659d7df7-6dkj8\" (UID: \"35248aca-4932-408d-a758-766cce4ae517\") " pod="openstack/horizon-79659d7df7-6dkj8" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.583041 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/35248aca-4932-408d-a758-766cce4ae517-logs\") pod \"horizon-79659d7df7-6dkj8\" (UID: \"35248aca-4932-408d-a758-766cce4ae517\") " pod="openstack/horizon-79659d7df7-6dkj8" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.583060 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tm2rl\" (UniqueName: \"kubernetes.io/projected/35248aca-4932-408d-a758-766cce4ae517-kube-api-access-tm2rl\") pod \"horizon-79659d7df7-6dkj8\" (UID: \"35248aca-4932-408d-a758-766cce4ae517\") " pod="openstack/horizon-79659d7df7-6dkj8" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.583088 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/35248aca-4932-408d-a758-766cce4ae517-horizon-secret-key\") pod \"horizon-79659d7df7-6dkj8\" (UID: \"35248aca-4932-408d-a758-766cce4ae517\") " pod="openstack/horizon-79659d7df7-6dkj8" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.583110 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/35248aca-4932-408d-a758-766cce4ae517-config-data\") pod \"horizon-79659d7df7-6dkj8\" (UID: \"35248aca-4932-408d-a758-766cce4ae517\") " pod="openstack/horizon-79659d7df7-6dkj8" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.584379 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/35248aca-4932-408d-a758-766cce4ae517-config-data\") pod \"horizon-79659d7df7-6dkj8\" (UID: \"35248aca-4932-408d-a758-766cce4ae517\") " pod="openstack/horizon-79659d7df7-6dkj8" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.584827 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/35248aca-4932-408d-a758-766cce4ae517-scripts\") pod \"horizon-79659d7df7-6dkj8\" (UID: \"35248aca-4932-408d-a758-766cce4ae517\") " pod="openstack/horizon-79659d7df7-6dkj8" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.585045 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/35248aca-4932-408d-a758-766cce4ae517-logs\") pod \"horizon-79659d7df7-6dkj8\" (UID: \"35248aca-4932-408d-a758-766cce4ae517\") " pod="openstack/horizon-79659d7df7-6dkj8" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.593547 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55fff446b9-84t5x" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.595957 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-tjcz4"] Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.596763 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/35248aca-4932-408d-a758-766cce4ae517-horizon-secret-key\") pod \"horizon-79659d7df7-6dkj8\" (UID: \"35248aca-4932-408d-a758-766cce4ae517\") " pod="openstack/horizon-79659d7df7-6dkj8" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.597266 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-tjcz4" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.607767 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.608168 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-sdgb9" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.608374 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.617325 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-xf24l"] Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.636792 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tm2rl\" (UniqueName: \"kubernetes.io/projected/35248aca-4932-408d-a758-766cce4ae517-kube-api-access-tm2rl\") pod \"horizon-79659d7df7-6dkj8\" (UID: \"35248aca-4932-408d-a758-766cce4ae517\") " pod="openstack/horizon-79659d7df7-6dkj8" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.649353 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-tjcz4"] Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.659608 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-mj282" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.693022 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/c036a628-a241-45ea-bd2a-5d6d50804f61-config\") pod \"neutron-db-sync-xf24l\" (UID: \"c036a628-a241-45ea-bd2a-5d6d50804f61\") " pod="openstack/neutron-db-sync-xf24l" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.693073 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9c66bb6a-5a3c-4223-aba6-c5d9d0ea83ba-etc-machine-id\") pod \"cinder-db-sync-tjcz4\" (UID: \"9c66bb6a-5a3c-4223-aba6-c5d9d0ea83ba\") " pod="openstack/cinder-db-sync-tjcz4" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.693107 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fqdf4\" (UniqueName: \"kubernetes.io/projected/9c66bb6a-5a3c-4223-aba6-c5d9d0ea83ba-kube-api-access-fqdf4\") pod \"cinder-db-sync-tjcz4\" (UID: \"9c66bb6a-5a3c-4223-aba6-c5d9d0ea83ba\") " pod="openstack/cinder-db-sync-tjcz4" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.693135 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/9c66bb6a-5a3c-4223-aba6-c5d9d0ea83ba-db-sync-config-data\") pod \"cinder-db-sync-tjcz4\" (UID: \"9c66bb6a-5a3c-4223-aba6-c5d9d0ea83ba\") " pod="openstack/cinder-db-sync-tjcz4" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.693157 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4jkb4\" (UniqueName: \"kubernetes.io/projected/c036a628-a241-45ea-bd2a-5d6d50804f61-kube-api-access-4jkb4\") pod \"neutron-db-sync-xf24l\" (UID: \"c036a628-a241-45ea-bd2a-5d6d50804f61\") " pod="openstack/neutron-db-sync-xf24l" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.693184 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c036a628-a241-45ea-bd2a-5d6d50804f61-combined-ca-bundle\") pod \"neutron-db-sync-xf24l\" (UID: \"c036a628-a241-45ea-bd2a-5d6d50804f61\") " pod="openstack/neutron-db-sync-xf24l" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.693204 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c66bb6a-5a3c-4223-aba6-c5d9d0ea83ba-scripts\") pod \"cinder-db-sync-tjcz4\" (UID: \"9c66bb6a-5a3c-4223-aba6-c5d9d0ea83ba\") " pod="openstack/cinder-db-sync-tjcz4" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.693224 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c66bb6a-5a3c-4223-aba6-c5d9d0ea83ba-config-data\") pod \"cinder-db-sync-tjcz4\" (UID: \"9c66bb6a-5a3c-4223-aba6-c5d9d0ea83ba\") " pod="openstack/cinder-db-sync-tjcz4" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.693295 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c66bb6a-5a3c-4223-aba6-c5d9d0ea83ba-combined-ca-bundle\") pod \"cinder-db-sync-tjcz4\" (UID: \"9c66bb6a-5a3c-4223-aba6-c5d9d0ea83ba\") " pod="openstack/cinder-db-sync-tjcz4" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.724995 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-gp9st"] Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.745554 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-79659d7df7-6dkj8" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.749351 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-gp9st"] Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.749551 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-gp9st" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.763397 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.763618 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-2djz5" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.773237 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.802798 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.812416 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0feba5bf-20c5-4578-9cfc-f37d6511c2d6-combined-ca-bundle\") pod \"barbican-db-sync-gp9st\" (UID: \"0feba5bf-20c5-4578-9cfc-f37d6511c2d6\") " pod="openstack/barbican-db-sync-gp9st" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.812613 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tf2lz\" (UniqueName: \"kubernetes.io/projected/0feba5bf-20c5-4578-9cfc-f37d6511c2d6-kube-api-access-tf2lz\") pod \"barbican-db-sync-gp9st\" (UID: \"0feba5bf-20c5-4578-9cfc-f37d6511c2d6\") " pod="openstack/barbican-db-sync-gp9st" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.812730 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/c036a628-a241-45ea-bd2a-5d6d50804f61-config\") pod \"neutron-db-sync-xf24l\" (UID: \"c036a628-a241-45ea-bd2a-5d6d50804f61\") " pod="openstack/neutron-db-sync-xf24l" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.812871 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9c66bb6a-5a3c-4223-aba6-c5d9d0ea83ba-etc-machine-id\") pod \"cinder-db-sync-tjcz4\" (UID: \"9c66bb6a-5a3c-4223-aba6-c5d9d0ea83ba\") " pod="openstack/cinder-db-sync-tjcz4" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.812972 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fqdf4\" (UniqueName: \"kubernetes.io/projected/9c66bb6a-5a3c-4223-aba6-c5d9d0ea83ba-kube-api-access-fqdf4\") pod \"cinder-db-sync-tjcz4\" (UID: \"9c66bb6a-5a3c-4223-aba6-c5d9d0ea83ba\") " pod="openstack/cinder-db-sync-tjcz4" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.813073 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/9c66bb6a-5a3c-4223-aba6-c5d9d0ea83ba-db-sync-config-data\") pod \"cinder-db-sync-tjcz4\" (UID: \"9c66bb6a-5a3c-4223-aba6-c5d9d0ea83ba\") " pod="openstack/cinder-db-sync-tjcz4" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.813319 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4jkb4\" (UniqueName: \"kubernetes.io/projected/c036a628-a241-45ea-bd2a-5d6d50804f61-kube-api-access-4jkb4\") pod \"neutron-db-sync-xf24l\" (UID: \"c036a628-a241-45ea-bd2a-5d6d50804f61\") " pod="openstack/neutron-db-sync-xf24l" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.813447 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c036a628-a241-45ea-bd2a-5d6d50804f61-combined-ca-bundle\") pod \"neutron-db-sync-xf24l\" (UID: \"c036a628-a241-45ea-bd2a-5d6d50804f61\") " pod="openstack/neutron-db-sync-xf24l" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.813547 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c66bb6a-5a3c-4223-aba6-c5d9d0ea83ba-scripts\") pod \"cinder-db-sync-tjcz4\" (UID: \"9c66bb6a-5a3c-4223-aba6-c5d9d0ea83ba\") " pod="openstack/cinder-db-sync-tjcz4" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.813643 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c66bb6a-5a3c-4223-aba6-c5d9d0ea83ba-config-data\") pod \"cinder-db-sync-tjcz4\" (UID: \"9c66bb6a-5a3c-4223-aba6-c5d9d0ea83ba\") " pod="openstack/cinder-db-sync-tjcz4" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.813022 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9c66bb6a-5a3c-4223-aba6-c5d9d0ea83ba-etc-machine-id\") pod \"cinder-db-sync-tjcz4\" (UID: \"9c66bb6a-5a3c-4223-aba6-c5d9d0ea83ba\") " pod="openstack/cinder-db-sync-tjcz4" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.814466 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c66bb6a-5a3c-4223-aba6-c5d9d0ea83ba-combined-ca-bundle\") pod \"cinder-db-sync-tjcz4\" (UID: \"9c66bb6a-5a3c-4223-aba6-c5d9d0ea83ba\") " pod="openstack/cinder-db-sync-tjcz4" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.814595 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0feba5bf-20c5-4578-9cfc-f37d6511c2d6-db-sync-config-data\") pod \"barbican-db-sync-gp9st\" (UID: \"0feba5bf-20c5-4578-9cfc-f37d6511c2d6\") " pod="openstack/barbican-db-sync-gp9st" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.815414 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.827909 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55fff446b9-84t5x"] Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.828193 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.828357 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.845125 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/9c66bb6a-5a3c-4223-aba6-c5d9d0ea83ba-db-sync-config-data\") pod \"cinder-db-sync-tjcz4\" (UID: \"9c66bb6a-5a3c-4223-aba6-c5d9d0ea83ba\") " pod="openstack/cinder-db-sync-tjcz4" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.854729 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4jkb4\" (UniqueName: \"kubernetes.io/projected/c036a628-a241-45ea-bd2a-5d6d50804f61-kube-api-access-4jkb4\") pod \"neutron-db-sync-xf24l\" (UID: \"c036a628-a241-45ea-bd2a-5d6d50804f61\") " pod="openstack/neutron-db-sync-xf24l" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.855090 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/c036a628-a241-45ea-bd2a-5d6d50804f61-config\") pod \"neutron-db-sync-xf24l\" (UID: \"c036a628-a241-45ea-bd2a-5d6d50804f61\") " pod="openstack/neutron-db-sync-xf24l" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.856474 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c036a628-a241-45ea-bd2a-5d6d50804f61-combined-ca-bundle\") pod \"neutron-db-sync-xf24l\" (UID: \"c036a628-a241-45ea-bd2a-5d6d50804f61\") " pod="openstack/neutron-db-sync-xf24l" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.865562 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c66bb6a-5a3c-4223-aba6-c5d9d0ea83ba-combined-ca-bundle\") pod \"cinder-db-sync-tjcz4\" (UID: \"9c66bb6a-5a3c-4223-aba6-c5d9d0ea83ba\") " pod="openstack/cinder-db-sync-tjcz4" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.867906 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c66bb6a-5a3c-4223-aba6-c5d9d0ea83ba-config-data\") pod \"cinder-db-sync-tjcz4\" (UID: \"9c66bb6a-5a3c-4223-aba6-c5d9d0ea83ba\") " pod="openstack/cinder-db-sync-tjcz4" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.869575 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c66bb6a-5a3c-4223-aba6-c5d9d0ea83ba-scripts\") pod \"cinder-db-sync-tjcz4\" (UID: \"9c66bb6a-5a3c-4223-aba6-c5d9d0ea83ba\") " pod="openstack/cinder-db-sync-tjcz4" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.875347 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fqdf4\" (UniqueName: \"kubernetes.io/projected/9c66bb6a-5a3c-4223-aba6-c5d9d0ea83ba-kube-api-access-fqdf4\") pod \"cinder-db-sync-tjcz4\" (UID: \"9c66bb6a-5a3c-4223-aba6-c5d9d0ea83ba\") " pod="openstack/cinder-db-sync-tjcz4" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.880399 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-8558cb5fd9-8cr6r"] Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.881808 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-8558cb5fd9-8cr6r" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.908064 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-8558cb5fd9-8cr6r"] Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.916693 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8ad39937-75ce-468c-90df-b3abbaefe498-logs\") pod \"horizon-8558cb5fd9-8cr6r\" (UID: \"8ad39937-75ce-468c-90df-b3abbaefe498\") " pod="openstack/horizon-8558cb5fd9-8cr6r" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.916736 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8ad39937-75ce-468c-90df-b3abbaefe498-horizon-secret-key\") pod \"horizon-8558cb5fd9-8cr6r\" (UID: \"8ad39937-75ce-468c-90df-b3abbaefe498\") " pod="openstack/horizon-8558cb5fd9-8cr6r" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.916769 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0feba5bf-20c5-4578-9cfc-f37d6511c2d6-db-sync-config-data\") pod \"barbican-db-sync-gp9st\" (UID: \"0feba5bf-20c5-4578-9cfc-f37d6511c2d6\") " pod="openstack/barbican-db-sync-gp9st" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.916789 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9226a04e-0195-42a1-b651-2f5379b1e2a9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9226a04e-0195-42a1-b651-2f5379b1e2a9\") " pod="openstack/ceilometer-0" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.916813 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0feba5bf-20c5-4578-9cfc-f37d6511c2d6-combined-ca-bundle\") pod \"barbican-db-sync-gp9st\" (UID: \"0feba5bf-20c5-4578-9cfc-f37d6511c2d6\") " pod="openstack/barbican-db-sync-gp9st" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.916827 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9226a04e-0195-42a1-b651-2f5379b1e2a9-log-httpd\") pod \"ceilometer-0\" (UID: \"9226a04e-0195-42a1-b651-2f5379b1e2a9\") " pod="openstack/ceilometer-0" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.916841 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8ad39937-75ce-468c-90df-b3abbaefe498-config-data\") pod \"horizon-8558cb5fd9-8cr6r\" (UID: \"8ad39937-75ce-468c-90df-b3abbaefe498\") " pod="openstack/horizon-8558cb5fd9-8cr6r" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.916860 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9226a04e-0195-42a1-b651-2f5379b1e2a9-scripts\") pod \"ceilometer-0\" (UID: \"9226a04e-0195-42a1-b651-2f5379b1e2a9\") " pod="openstack/ceilometer-0" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.916879 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tf2lz\" (UniqueName: \"kubernetes.io/projected/0feba5bf-20c5-4578-9cfc-f37d6511c2d6-kube-api-access-tf2lz\") pod \"barbican-db-sync-gp9st\" (UID: \"0feba5bf-20c5-4578-9cfc-f37d6511c2d6\") " pod="openstack/barbican-db-sync-gp9st" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.916914 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9226a04e-0195-42a1-b651-2f5379b1e2a9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9226a04e-0195-42a1-b651-2f5379b1e2a9\") " pod="openstack/ceilometer-0" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.916944 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z2mzr\" (UniqueName: \"kubernetes.io/projected/9226a04e-0195-42a1-b651-2f5379b1e2a9-kube-api-access-z2mzr\") pod \"ceilometer-0\" (UID: \"9226a04e-0195-42a1-b651-2f5379b1e2a9\") " pod="openstack/ceilometer-0" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.916965 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8ad39937-75ce-468c-90df-b3abbaefe498-scripts\") pod \"horizon-8558cb5fd9-8cr6r\" (UID: \"8ad39937-75ce-468c-90df-b3abbaefe498\") " pod="openstack/horizon-8558cb5fd9-8cr6r" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.916987 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9226a04e-0195-42a1-b651-2f5379b1e2a9-run-httpd\") pod \"ceilometer-0\" (UID: \"9226a04e-0195-42a1-b651-2f5379b1e2a9\") " pod="openstack/ceilometer-0" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.917013 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x2mfc\" (UniqueName: \"kubernetes.io/projected/8ad39937-75ce-468c-90df-b3abbaefe498-kube-api-access-x2mfc\") pod \"horizon-8558cb5fd9-8cr6r\" (UID: \"8ad39937-75ce-468c-90df-b3abbaefe498\") " pod="openstack/horizon-8558cb5fd9-8cr6r" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.917042 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9226a04e-0195-42a1-b651-2f5379b1e2a9-config-data\") pod \"ceilometer-0\" (UID: \"9226a04e-0195-42a1-b651-2f5379b1e2a9\") " pod="openstack/ceilometer-0" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.917250 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-vj2tf"] Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.918442 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-vj2tf" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.927712 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-vj2tf"] Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.928065 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-xf24l" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.930176 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0feba5bf-20c5-4578-9cfc-f37d6511c2d6-combined-ca-bundle\") pod \"barbican-db-sync-gp9st\" (UID: \"0feba5bf-20c5-4578-9cfc-f37d6511c2d6\") " pod="openstack/barbican-db-sync-gp9st" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.934722 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0feba5bf-20c5-4578-9cfc-f37d6511c2d6-db-sync-config-data\") pod \"barbican-db-sync-gp9st\" (UID: \"0feba5bf-20c5-4578-9cfc-f37d6511c2d6\") " pod="openstack/barbican-db-sync-gp9st" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.936145 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.936342 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.936475 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-k6xjq" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.944361 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tf2lz\" (UniqueName: \"kubernetes.io/projected/0feba5bf-20c5-4578-9cfc-f37d6511c2d6-kube-api-access-tf2lz\") pod \"barbican-db-sync-gp9st\" (UID: \"0feba5bf-20c5-4578-9cfc-f37d6511c2d6\") " pod="openstack/barbican-db-sync-gp9st" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.965632 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-76fcf4b695-wpgst"] Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.967086 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76fcf4b695-wpgst" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.968184 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-tjcz4" Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.985349 4885 generic.go:334] "Generic (PLEG): container finished" podID="9fc15c77-b663-4395-ae92-922b1e4756ba" containerID="4b4ebce37f94f38134bb618a2002d6b8fb4bc1bec4fc68a0d3ab6a914e2c84a0" exitCode=0 Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.985388 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-x2nx7" event={"ID":"9fc15c77-b663-4395-ae92-922b1e4756ba","Type":"ContainerDied","Data":"4b4ebce37f94f38134bb618a2002d6b8fb4bc1bec4fc68a0d3ab6a914e2c84a0"} Oct 02 02:03:38 crc kubenswrapper[4885]: I1002 02:03:38.995400 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-76fcf4b695-wpgst"] Oct 02 02:03:39 crc kubenswrapper[4885]: I1002 02:03:39.023427 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9226a04e-0195-42a1-b651-2f5379b1e2a9-config-data\") pod \"ceilometer-0\" (UID: \"9226a04e-0195-42a1-b651-2f5379b1e2a9\") " pod="openstack/ceilometer-0" Oct 02 02:03:39 crc kubenswrapper[4885]: I1002 02:03:39.023482 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8fa66af9-7b4b-480c-8f96-bda850e2eef6-scripts\") pod \"placement-db-sync-vj2tf\" (UID: \"8fa66af9-7b4b-480c-8f96-bda850e2eef6\") " pod="openstack/placement-db-sync-vj2tf" Oct 02 02:03:39 crc kubenswrapper[4885]: I1002 02:03:39.023522 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5adbde10-fd09-4e06-bad3-7f5d7bdf5b41-dns-svc\") pod \"dnsmasq-dns-76fcf4b695-wpgst\" (UID: \"5adbde10-fd09-4e06-bad3-7f5d7bdf5b41\") " pod="openstack/dnsmasq-dns-76fcf4b695-wpgst" Oct 02 02:03:39 crc kubenswrapper[4885]: I1002 02:03:39.023546 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8ad39937-75ce-468c-90df-b3abbaefe498-logs\") pod \"horizon-8558cb5fd9-8cr6r\" (UID: \"8ad39937-75ce-468c-90df-b3abbaefe498\") " pod="openstack/horizon-8558cb5fd9-8cr6r" Oct 02 02:03:39 crc kubenswrapper[4885]: I1002 02:03:39.023573 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8fa66af9-7b4b-480c-8f96-bda850e2eef6-combined-ca-bundle\") pod \"placement-db-sync-vj2tf\" (UID: \"8fa66af9-7b4b-480c-8f96-bda850e2eef6\") " pod="openstack/placement-db-sync-vj2tf" Oct 02 02:03:39 crc kubenswrapper[4885]: I1002 02:03:39.023601 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8ad39937-75ce-468c-90df-b3abbaefe498-horizon-secret-key\") pod \"horizon-8558cb5fd9-8cr6r\" (UID: \"8ad39937-75ce-468c-90df-b3abbaefe498\") " pod="openstack/horizon-8558cb5fd9-8cr6r" Oct 02 02:03:39 crc kubenswrapper[4885]: I1002 02:03:39.023634 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9226a04e-0195-42a1-b651-2f5379b1e2a9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9226a04e-0195-42a1-b651-2f5379b1e2a9\") " pod="openstack/ceilometer-0" Oct 02 02:03:39 crc kubenswrapper[4885]: I1002 02:03:39.023663 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9226a04e-0195-42a1-b651-2f5379b1e2a9-log-httpd\") pod \"ceilometer-0\" (UID: \"9226a04e-0195-42a1-b651-2f5379b1e2a9\") " pod="openstack/ceilometer-0" Oct 02 02:03:39 crc kubenswrapper[4885]: I1002 02:03:39.023684 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8ad39937-75ce-468c-90df-b3abbaefe498-config-data\") pod \"horizon-8558cb5fd9-8cr6r\" (UID: \"8ad39937-75ce-468c-90df-b3abbaefe498\") " pod="openstack/horizon-8558cb5fd9-8cr6r" Oct 02 02:03:39 crc kubenswrapper[4885]: I1002 02:03:39.023707 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9226a04e-0195-42a1-b651-2f5379b1e2a9-scripts\") pod \"ceilometer-0\" (UID: \"9226a04e-0195-42a1-b651-2f5379b1e2a9\") " pod="openstack/ceilometer-0" Oct 02 02:03:39 crc kubenswrapper[4885]: I1002 02:03:39.023756 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-58fgd\" (UniqueName: \"kubernetes.io/projected/5adbde10-fd09-4e06-bad3-7f5d7bdf5b41-kube-api-access-58fgd\") pod \"dnsmasq-dns-76fcf4b695-wpgst\" (UID: \"5adbde10-fd09-4e06-bad3-7f5d7bdf5b41\") " pod="openstack/dnsmasq-dns-76fcf4b695-wpgst" Oct 02 02:03:39 crc kubenswrapper[4885]: I1002 02:03:39.023781 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fpmmt\" (UniqueName: \"kubernetes.io/projected/8fa66af9-7b4b-480c-8f96-bda850e2eef6-kube-api-access-fpmmt\") pod \"placement-db-sync-vj2tf\" (UID: \"8fa66af9-7b4b-480c-8f96-bda850e2eef6\") " pod="openstack/placement-db-sync-vj2tf" Oct 02 02:03:39 crc kubenswrapper[4885]: I1002 02:03:39.023804 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9226a04e-0195-42a1-b651-2f5379b1e2a9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9226a04e-0195-42a1-b651-2f5379b1e2a9\") " pod="openstack/ceilometer-0" Oct 02 02:03:39 crc kubenswrapper[4885]: I1002 02:03:39.023832 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8fa66af9-7b4b-480c-8f96-bda850e2eef6-logs\") pod \"placement-db-sync-vj2tf\" (UID: \"8fa66af9-7b4b-480c-8f96-bda850e2eef6\") " pod="openstack/placement-db-sync-vj2tf" Oct 02 02:03:39 crc kubenswrapper[4885]: I1002 02:03:39.023853 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5adbde10-fd09-4e06-bad3-7f5d7bdf5b41-dns-swift-storage-0\") pod \"dnsmasq-dns-76fcf4b695-wpgst\" (UID: \"5adbde10-fd09-4e06-bad3-7f5d7bdf5b41\") " pod="openstack/dnsmasq-dns-76fcf4b695-wpgst" Oct 02 02:03:39 crc kubenswrapper[4885]: I1002 02:03:39.023891 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z2mzr\" (UniqueName: \"kubernetes.io/projected/9226a04e-0195-42a1-b651-2f5379b1e2a9-kube-api-access-z2mzr\") pod \"ceilometer-0\" (UID: \"9226a04e-0195-42a1-b651-2f5379b1e2a9\") " pod="openstack/ceilometer-0" Oct 02 02:03:39 crc kubenswrapper[4885]: I1002 02:03:39.023913 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8fa66af9-7b4b-480c-8f96-bda850e2eef6-config-data\") pod \"placement-db-sync-vj2tf\" (UID: \"8fa66af9-7b4b-480c-8f96-bda850e2eef6\") " pod="openstack/placement-db-sync-vj2tf" Oct 02 02:03:39 crc kubenswrapper[4885]: I1002 02:03:39.023942 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5adbde10-fd09-4e06-bad3-7f5d7bdf5b41-ovsdbserver-sb\") pod \"dnsmasq-dns-76fcf4b695-wpgst\" (UID: \"5adbde10-fd09-4e06-bad3-7f5d7bdf5b41\") " pod="openstack/dnsmasq-dns-76fcf4b695-wpgst" Oct 02 02:03:39 crc kubenswrapper[4885]: I1002 02:03:39.023972 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8ad39937-75ce-468c-90df-b3abbaefe498-scripts\") pod \"horizon-8558cb5fd9-8cr6r\" (UID: \"8ad39937-75ce-468c-90df-b3abbaefe498\") " pod="openstack/horizon-8558cb5fd9-8cr6r" Oct 02 02:03:39 crc kubenswrapper[4885]: I1002 02:03:39.024005 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9226a04e-0195-42a1-b651-2f5379b1e2a9-run-httpd\") pod \"ceilometer-0\" (UID: \"9226a04e-0195-42a1-b651-2f5379b1e2a9\") " pod="openstack/ceilometer-0" Oct 02 02:03:39 crc kubenswrapper[4885]: I1002 02:03:39.024042 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x2mfc\" (UniqueName: \"kubernetes.io/projected/8ad39937-75ce-468c-90df-b3abbaefe498-kube-api-access-x2mfc\") pod \"horizon-8558cb5fd9-8cr6r\" (UID: \"8ad39937-75ce-468c-90df-b3abbaefe498\") " pod="openstack/horizon-8558cb5fd9-8cr6r" Oct 02 02:03:39 crc kubenswrapper[4885]: I1002 02:03:39.024067 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5adbde10-fd09-4e06-bad3-7f5d7bdf5b41-config\") pod \"dnsmasq-dns-76fcf4b695-wpgst\" (UID: \"5adbde10-fd09-4e06-bad3-7f5d7bdf5b41\") " pod="openstack/dnsmasq-dns-76fcf4b695-wpgst" Oct 02 02:03:39 crc kubenswrapper[4885]: I1002 02:03:39.024558 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8ad39937-75ce-468c-90df-b3abbaefe498-logs\") pod \"horizon-8558cb5fd9-8cr6r\" (UID: \"8ad39937-75ce-468c-90df-b3abbaefe498\") " pod="openstack/horizon-8558cb5fd9-8cr6r" Oct 02 02:03:39 crc kubenswrapper[4885]: I1002 02:03:39.025055 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8ad39937-75ce-468c-90df-b3abbaefe498-scripts\") pod \"horizon-8558cb5fd9-8cr6r\" (UID: \"8ad39937-75ce-468c-90df-b3abbaefe498\") " pod="openstack/horizon-8558cb5fd9-8cr6r" Oct 02 02:03:39 crc kubenswrapper[4885]: I1002 02:03:39.027416 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9226a04e-0195-42a1-b651-2f5379b1e2a9-run-httpd\") pod \"ceilometer-0\" (UID: \"9226a04e-0195-42a1-b651-2f5379b1e2a9\") " pod="openstack/ceilometer-0" Oct 02 02:03:39 crc kubenswrapper[4885]: I1002 02:03:39.027804 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9226a04e-0195-42a1-b651-2f5379b1e2a9-log-httpd\") pod \"ceilometer-0\" (UID: \"9226a04e-0195-42a1-b651-2f5379b1e2a9\") " pod="openstack/ceilometer-0" Oct 02 02:03:39 crc kubenswrapper[4885]: I1002 02:03:39.029434 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8ad39937-75ce-468c-90df-b3abbaefe498-config-data\") pod \"horizon-8558cb5fd9-8cr6r\" (UID: \"8ad39937-75ce-468c-90df-b3abbaefe498\") " pod="openstack/horizon-8558cb5fd9-8cr6r" Oct 02 02:03:39 crc kubenswrapper[4885]: I1002 02:03:39.029513 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5adbde10-fd09-4e06-bad3-7f5d7bdf5b41-ovsdbserver-nb\") pod \"dnsmasq-dns-76fcf4b695-wpgst\" (UID: \"5adbde10-fd09-4e06-bad3-7f5d7bdf5b41\") " pod="openstack/dnsmasq-dns-76fcf4b695-wpgst" Oct 02 02:03:39 crc kubenswrapper[4885]: I1002 02:03:39.032299 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9226a04e-0195-42a1-b651-2f5379b1e2a9-config-data\") pod \"ceilometer-0\" (UID: \"9226a04e-0195-42a1-b651-2f5379b1e2a9\") " pod="openstack/ceilometer-0" Oct 02 02:03:39 crc kubenswrapper[4885]: I1002 02:03:39.046718 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9226a04e-0195-42a1-b651-2f5379b1e2a9-scripts\") pod \"ceilometer-0\" (UID: \"9226a04e-0195-42a1-b651-2f5379b1e2a9\") " pod="openstack/ceilometer-0" Oct 02 02:03:39 crc kubenswrapper[4885]: I1002 02:03:39.048084 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z2mzr\" (UniqueName: \"kubernetes.io/projected/9226a04e-0195-42a1-b651-2f5379b1e2a9-kube-api-access-z2mzr\") pod \"ceilometer-0\" (UID: \"9226a04e-0195-42a1-b651-2f5379b1e2a9\") " pod="openstack/ceilometer-0" Oct 02 02:03:39 crc kubenswrapper[4885]: I1002 02:03:39.052567 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9226a04e-0195-42a1-b651-2f5379b1e2a9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9226a04e-0195-42a1-b651-2f5379b1e2a9\") " pod="openstack/ceilometer-0" Oct 02 02:03:39 crc kubenswrapper[4885]: I1002 02:03:39.053807 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77585f5f8c-x2nx7" Oct 02 02:03:39 crc kubenswrapper[4885]: I1002 02:03:39.055458 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8ad39937-75ce-468c-90df-b3abbaefe498-horizon-secret-key\") pod \"horizon-8558cb5fd9-8cr6r\" (UID: \"8ad39937-75ce-468c-90df-b3abbaefe498\") " pod="openstack/horizon-8558cb5fd9-8cr6r" Oct 02 02:03:39 crc kubenswrapper[4885]: I1002 02:03:39.056737 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x2mfc\" (UniqueName: \"kubernetes.io/projected/8ad39937-75ce-468c-90df-b3abbaefe498-kube-api-access-x2mfc\") pod \"horizon-8558cb5fd9-8cr6r\" (UID: \"8ad39937-75ce-468c-90df-b3abbaefe498\") " pod="openstack/horizon-8558cb5fd9-8cr6r" Oct 02 02:03:39 crc kubenswrapper[4885]: I1002 02:03:39.058106 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9226a04e-0195-42a1-b651-2f5379b1e2a9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9226a04e-0195-42a1-b651-2f5379b1e2a9\") " pod="openstack/ceilometer-0" Oct 02 02:03:39 crc kubenswrapper[4885]: I1002 02:03:39.130511 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9fc15c77-b663-4395-ae92-922b1e4756ba-ovsdbserver-sb\") pod \"9fc15c77-b663-4395-ae92-922b1e4756ba\" (UID: \"9fc15c77-b663-4395-ae92-922b1e4756ba\") " Oct 02 02:03:39 crc kubenswrapper[4885]: I1002 02:03:39.130584 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9fc15c77-b663-4395-ae92-922b1e4756ba-dns-svc\") pod \"9fc15c77-b663-4395-ae92-922b1e4756ba\" (UID: \"9fc15c77-b663-4395-ae92-922b1e4756ba\") " Oct 02 02:03:39 crc kubenswrapper[4885]: I1002 02:03:39.130606 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9fc15c77-b663-4395-ae92-922b1e4756ba-ovsdbserver-nb\") pod \"9fc15c77-b663-4395-ae92-922b1e4756ba\" (UID: \"9fc15c77-b663-4395-ae92-922b1e4756ba\") " Oct 02 02:03:39 crc kubenswrapper[4885]: I1002 02:03:39.130643 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9fc15c77-b663-4395-ae92-922b1e4756ba-dns-swift-storage-0\") pod \"9fc15c77-b663-4395-ae92-922b1e4756ba\" (UID: \"9fc15c77-b663-4395-ae92-922b1e4756ba\") " Oct 02 02:03:39 crc kubenswrapper[4885]: I1002 02:03:39.130797 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9fc15c77-b663-4395-ae92-922b1e4756ba-config\") pod \"9fc15c77-b663-4395-ae92-922b1e4756ba\" (UID: \"9fc15c77-b663-4395-ae92-922b1e4756ba\") " Oct 02 02:03:39 crc kubenswrapper[4885]: I1002 02:03:39.130971 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ht7d9\" (UniqueName: \"kubernetes.io/projected/9fc15c77-b663-4395-ae92-922b1e4756ba-kube-api-access-ht7d9\") pod \"9fc15c77-b663-4395-ae92-922b1e4756ba\" (UID: \"9fc15c77-b663-4395-ae92-922b1e4756ba\") " Oct 02 02:03:39 crc kubenswrapper[4885]: I1002 02:03:39.131403 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-58fgd\" (UniqueName: \"kubernetes.io/projected/5adbde10-fd09-4e06-bad3-7f5d7bdf5b41-kube-api-access-58fgd\") pod \"dnsmasq-dns-76fcf4b695-wpgst\" (UID: \"5adbde10-fd09-4e06-bad3-7f5d7bdf5b41\") " pod="openstack/dnsmasq-dns-76fcf4b695-wpgst" Oct 02 02:03:39 crc kubenswrapper[4885]: I1002 02:03:39.131433 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fpmmt\" (UniqueName: \"kubernetes.io/projected/8fa66af9-7b4b-480c-8f96-bda850e2eef6-kube-api-access-fpmmt\") pod \"placement-db-sync-vj2tf\" (UID: \"8fa66af9-7b4b-480c-8f96-bda850e2eef6\") " pod="openstack/placement-db-sync-vj2tf" Oct 02 02:03:39 crc kubenswrapper[4885]: I1002 02:03:39.131482 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8fa66af9-7b4b-480c-8f96-bda850e2eef6-logs\") pod \"placement-db-sync-vj2tf\" (UID: \"8fa66af9-7b4b-480c-8f96-bda850e2eef6\") " pod="openstack/placement-db-sync-vj2tf" Oct 02 02:03:39 crc kubenswrapper[4885]: I1002 02:03:39.131501 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5adbde10-fd09-4e06-bad3-7f5d7bdf5b41-dns-swift-storage-0\") pod \"dnsmasq-dns-76fcf4b695-wpgst\" (UID: \"5adbde10-fd09-4e06-bad3-7f5d7bdf5b41\") " pod="openstack/dnsmasq-dns-76fcf4b695-wpgst" Oct 02 02:03:39 crc kubenswrapper[4885]: I1002 02:03:39.131551 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8fa66af9-7b4b-480c-8f96-bda850e2eef6-config-data\") pod \"placement-db-sync-vj2tf\" (UID: \"8fa66af9-7b4b-480c-8f96-bda850e2eef6\") " pod="openstack/placement-db-sync-vj2tf" Oct 02 02:03:39 crc kubenswrapper[4885]: I1002 02:03:39.131572 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5adbde10-fd09-4e06-bad3-7f5d7bdf5b41-ovsdbserver-sb\") pod \"dnsmasq-dns-76fcf4b695-wpgst\" (UID: \"5adbde10-fd09-4e06-bad3-7f5d7bdf5b41\") " pod="openstack/dnsmasq-dns-76fcf4b695-wpgst" Oct 02 02:03:39 crc kubenswrapper[4885]: I1002 02:03:39.131642 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5adbde10-fd09-4e06-bad3-7f5d7bdf5b41-config\") pod \"dnsmasq-dns-76fcf4b695-wpgst\" (UID: \"5adbde10-fd09-4e06-bad3-7f5d7bdf5b41\") " pod="openstack/dnsmasq-dns-76fcf4b695-wpgst" Oct 02 02:03:39 crc kubenswrapper[4885]: I1002 02:03:39.131668 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5adbde10-fd09-4e06-bad3-7f5d7bdf5b41-ovsdbserver-nb\") pod \"dnsmasq-dns-76fcf4b695-wpgst\" (UID: \"5adbde10-fd09-4e06-bad3-7f5d7bdf5b41\") " pod="openstack/dnsmasq-dns-76fcf4b695-wpgst" Oct 02 02:03:39 crc kubenswrapper[4885]: I1002 02:03:39.131715 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8fa66af9-7b4b-480c-8f96-bda850e2eef6-scripts\") pod \"placement-db-sync-vj2tf\" (UID: \"8fa66af9-7b4b-480c-8f96-bda850e2eef6\") " pod="openstack/placement-db-sync-vj2tf" Oct 02 02:03:39 crc kubenswrapper[4885]: I1002 02:03:39.131749 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5adbde10-fd09-4e06-bad3-7f5d7bdf5b41-dns-svc\") pod \"dnsmasq-dns-76fcf4b695-wpgst\" (UID: \"5adbde10-fd09-4e06-bad3-7f5d7bdf5b41\") " pod="openstack/dnsmasq-dns-76fcf4b695-wpgst" Oct 02 02:03:39 crc kubenswrapper[4885]: I1002 02:03:39.131788 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8fa66af9-7b4b-480c-8f96-bda850e2eef6-combined-ca-bundle\") pod \"placement-db-sync-vj2tf\" (UID: \"8fa66af9-7b4b-480c-8f96-bda850e2eef6\") " pod="openstack/placement-db-sync-vj2tf" Oct 02 02:03:39 crc kubenswrapper[4885]: I1002 02:03:39.134231 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5adbde10-fd09-4e06-bad3-7f5d7bdf5b41-dns-swift-storage-0\") pod \"dnsmasq-dns-76fcf4b695-wpgst\" (UID: \"5adbde10-fd09-4e06-bad3-7f5d7bdf5b41\") " pod="openstack/dnsmasq-dns-76fcf4b695-wpgst" Oct 02 02:03:39 crc kubenswrapper[4885]: I1002 02:03:39.136207 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5adbde10-fd09-4e06-bad3-7f5d7bdf5b41-ovsdbserver-sb\") pod \"dnsmasq-dns-76fcf4b695-wpgst\" (UID: \"5adbde10-fd09-4e06-bad3-7f5d7bdf5b41\") " pod="openstack/dnsmasq-dns-76fcf4b695-wpgst" Oct 02 02:03:39 crc kubenswrapper[4885]: I1002 02:03:39.136467 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5adbde10-fd09-4e06-bad3-7f5d7bdf5b41-dns-svc\") pod \"dnsmasq-dns-76fcf4b695-wpgst\" (UID: \"5adbde10-fd09-4e06-bad3-7f5d7bdf5b41\") " pod="openstack/dnsmasq-dns-76fcf4b695-wpgst" Oct 02 02:03:39 crc kubenswrapper[4885]: I1002 02:03:39.136814 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5adbde10-fd09-4e06-bad3-7f5d7bdf5b41-ovsdbserver-nb\") pod \"dnsmasq-dns-76fcf4b695-wpgst\" (UID: \"5adbde10-fd09-4e06-bad3-7f5d7bdf5b41\") " pod="openstack/dnsmasq-dns-76fcf4b695-wpgst" Oct 02 02:03:39 crc kubenswrapper[4885]: I1002 02:03:39.138038 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8fa66af9-7b4b-480c-8f96-bda850e2eef6-logs\") pod \"placement-db-sync-vj2tf\" (UID: \"8fa66af9-7b4b-480c-8f96-bda850e2eef6\") " pod="openstack/placement-db-sync-vj2tf" Oct 02 02:03:39 crc kubenswrapper[4885]: I1002 02:03:39.139983 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8fa66af9-7b4b-480c-8f96-bda850e2eef6-combined-ca-bundle\") pod \"placement-db-sync-vj2tf\" (UID: \"8fa66af9-7b4b-480c-8f96-bda850e2eef6\") " pod="openstack/placement-db-sync-vj2tf" Oct 02 02:03:39 crc kubenswrapper[4885]: I1002 02:03:39.142173 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5adbde10-fd09-4e06-bad3-7f5d7bdf5b41-config\") pod \"dnsmasq-dns-76fcf4b695-wpgst\" (UID: \"5adbde10-fd09-4e06-bad3-7f5d7bdf5b41\") " pod="openstack/dnsmasq-dns-76fcf4b695-wpgst" Oct 02 02:03:39 crc kubenswrapper[4885]: I1002 02:03:39.145061 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-gp9st" Oct 02 02:03:39 crc kubenswrapper[4885]: I1002 02:03:39.153276 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8fa66af9-7b4b-480c-8f96-bda850e2eef6-scripts\") pod \"placement-db-sync-vj2tf\" (UID: \"8fa66af9-7b4b-480c-8f96-bda850e2eef6\") " pod="openstack/placement-db-sync-vj2tf" Oct 02 02:03:39 crc kubenswrapper[4885]: I1002 02:03:39.154219 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8fa66af9-7b4b-480c-8f96-bda850e2eef6-config-data\") pod \"placement-db-sync-vj2tf\" (UID: \"8fa66af9-7b4b-480c-8f96-bda850e2eef6\") " pod="openstack/placement-db-sync-vj2tf" Oct 02 02:03:39 crc kubenswrapper[4885]: I1002 02:03:39.154374 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9fc15c77-b663-4395-ae92-922b1e4756ba-kube-api-access-ht7d9" (OuterVolumeSpecName: "kube-api-access-ht7d9") pod "9fc15c77-b663-4395-ae92-922b1e4756ba" (UID: "9fc15c77-b663-4395-ae92-922b1e4756ba"). InnerVolumeSpecName "kube-api-access-ht7d9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:03:39 crc kubenswrapper[4885]: I1002 02:03:39.173612 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fpmmt\" (UniqueName: \"kubernetes.io/projected/8fa66af9-7b4b-480c-8f96-bda850e2eef6-kube-api-access-fpmmt\") pod \"placement-db-sync-vj2tf\" (UID: \"8fa66af9-7b4b-480c-8f96-bda850e2eef6\") " pod="openstack/placement-db-sync-vj2tf" Oct 02 02:03:39 crc kubenswrapper[4885]: I1002 02:03:39.173933 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 02:03:39 crc kubenswrapper[4885]: I1002 02:03:39.196708 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-58fgd\" (UniqueName: \"kubernetes.io/projected/5adbde10-fd09-4e06-bad3-7f5d7bdf5b41-kube-api-access-58fgd\") pod \"dnsmasq-dns-76fcf4b695-wpgst\" (UID: \"5adbde10-fd09-4e06-bad3-7f5d7bdf5b41\") " pod="openstack/dnsmasq-dns-76fcf4b695-wpgst" Oct 02 02:03:39 crc kubenswrapper[4885]: I1002 02:03:39.230038 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9fc15c77-b663-4395-ae92-922b1e4756ba-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "9fc15c77-b663-4395-ae92-922b1e4756ba" (UID: "9fc15c77-b663-4395-ae92-922b1e4756ba"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:03:39 crc kubenswrapper[4885]: I1002 02:03:39.232900 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ht7d9\" (UniqueName: \"kubernetes.io/projected/9fc15c77-b663-4395-ae92-922b1e4756ba-kube-api-access-ht7d9\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:39 crc kubenswrapper[4885]: I1002 02:03:39.232927 4885 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9fc15c77-b663-4395-ae92-922b1e4756ba-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:39 crc kubenswrapper[4885]: I1002 02:03:39.236824 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9fc15c77-b663-4395-ae92-922b1e4756ba-config" (OuterVolumeSpecName: "config") pod "9fc15c77-b663-4395-ae92-922b1e4756ba" (UID: "9fc15c77-b663-4395-ae92-922b1e4756ba"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:03:39 crc kubenswrapper[4885]: I1002 02:03:39.238312 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9fc15c77-b663-4395-ae92-922b1e4756ba-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "9fc15c77-b663-4395-ae92-922b1e4756ba" (UID: "9fc15c77-b663-4395-ae92-922b1e4756ba"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:03:39 crc kubenswrapper[4885]: I1002 02:03:39.239119 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9fc15c77-b663-4395-ae92-922b1e4756ba-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "9fc15c77-b663-4395-ae92-922b1e4756ba" (UID: "9fc15c77-b663-4395-ae92-922b1e4756ba"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:03:39 crc kubenswrapper[4885]: I1002 02:03:39.245251 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-8558cb5fd9-8cr6r" Oct 02 02:03:39 crc kubenswrapper[4885]: I1002 02:03:39.252672 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9fc15c77-b663-4395-ae92-922b1e4756ba-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9fc15c77-b663-4395-ae92-922b1e4756ba" (UID: "9fc15c77-b663-4395-ae92-922b1e4756ba"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:03:39 crc kubenswrapper[4885]: I1002 02:03:39.257791 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-vj2tf" Oct 02 02:03:39 crc kubenswrapper[4885]: I1002 02:03:39.337848 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76fcf4b695-wpgst" Oct 02 02:03:39 crc kubenswrapper[4885]: I1002 02:03:39.339781 4885 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9fc15c77-b663-4395-ae92-922b1e4756ba-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:39 crc kubenswrapper[4885]: I1002 02:03:39.339796 4885 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9fc15c77-b663-4395-ae92-922b1e4756ba-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:39 crc kubenswrapper[4885]: I1002 02:03:39.339805 4885 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9fc15c77-b663-4395-ae92-922b1e4756ba-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:39 crc kubenswrapper[4885]: I1002 02:03:39.339814 4885 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9fc15c77-b663-4395-ae92-922b1e4756ba-config\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:39 crc kubenswrapper[4885]: I1002 02:03:39.398110 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55fff446b9-84t5x"] Oct 02 02:03:39 crc kubenswrapper[4885]: I1002 02:03:39.480191 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-79659d7df7-6dkj8"] Oct 02 02:03:39 crc kubenswrapper[4885]: I1002 02:03:39.497053 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-mj282"] Oct 02 02:03:39 crc kubenswrapper[4885]: W1002 02:03:39.518186 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod35248aca_4932_408d_a758_766cce4ae517.slice/crio-a508a5af806d151a21550ebba73caf8ae175cddf2a7941e341c4d67e97c3af77 WatchSource:0}: Error finding container a508a5af806d151a21550ebba73caf8ae175cddf2a7941e341c4d67e97c3af77: Status 404 returned error can't find the container with id a508a5af806d151a21550ebba73caf8ae175cddf2a7941e341c4d67e97c3af77 Oct 02 02:03:39 crc kubenswrapper[4885]: I1002 02:03:39.680009 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-tjcz4"] Oct 02 02:03:39 crc kubenswrapper[4885]: I1002 02:03:39.688147 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-xf24l"] Oct 02 02:03:39 crc kubenswrapper[4885]: W1002 02:03:39.697442 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc036a628_a241_45ea_bd2a_5d6d50804f61.slice/crio-a90523b46bfc0032de65212bd4351879dd04017043fc5b44bd795c65402bec04 WatchSource:0}: Error finding container a90523b46bfc0032de65212bd4351879dd04017043fc5b44bd795c65402bec04: Status 404 returned error can't find the container with id a90523b46bfc0032de65212bd4351879dd04017043fc5b44bd795c65402bec04 Oct 02 02:03:39 crc kubenswrapper[4885]: W1002 02:03:39.700038 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9c66bb6a_5a3c_4223_aba6_c5d9d0ea83ba.slice/crio-a265f07b2a178becf2d4f64243b607392df0bcc0cd1797f70e9daa3dd886a014 WatchSource:0}: Error finding container a265f07b2a178becf2d4f64243b607392df0bcc0cd1797f70e9daa3dd886a014: Status 404 returned error can't find the container with id a265f07b2a178becf2d4f64243b607392df0bcc0cd1797f70e9daa3dd886a014 Oct 02 02:03:39 crc kubenswrapper[4885]: I1002 02:03:39.850446 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 02:03:39 crc kubenswrapper[4885]: I1002 02:03:39.864514 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-gp9st"] Oct 02 02:03:39 crc kubenswrapper[4885]: I1002 02:03:39.955777 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-8558cb5fd9-8cr6r"] Oct 02 02:03:39 crc kubenswrapper[4885]: W1002 02:03:39.957878 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8ad39937_75ce_468c_90df_b3abbaefe498.slice/crio-6f02b41e568cd4c191d3ac36f5fd8d6536883727af2c4417485c66f8afc7b87f WatchSource:0}: Error finding container 6f02b41e568cd4c191d3ac36f5fd8d6536883727af2c4417485c66f8afc7b87f: Status 404 returned error can't find the container with id 6f02b41e568cd4c191d3ac36f5fd8d6536883727af2c4417485c66f8afc7b87f Oct 02 02:03:39 crc kubenswrapper[4885]: I1002 02:03:39.968949 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-76fcf4b695-wpgst"] Oct 02 02:03:39 crc kubenswrapper[4885]: I1002 02:03:39.992732 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-vj2tf"] Oct 02 02:03:39 crc kubenswrapper[4885]: W1002 02:03:39.992915 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8fa66af9_7b4b_480c_8f96_bda850e2eef6.slice/crio-55414b8c8eb0471f8443ae7434426584d5315281a20ef9fd8540ffc929bb4f50 WatchSource:0}: Error finding container 55414b8c8eb0471f8443ae7434426584d5315281a20ef9fd8540ffc929bb4f50: Status 404 returned error can't find the container with id 55414b8c8eb0471f8443ae7434426584d5315281a20ef9fd8540ffc929bb4f50 Oct 02 02:03:39 crc kubenswrapper[4885]: I1002 02:03:39.994615 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-79659d7df7-6dkj8" event={"ID":"35248aca-4932-408d-a758-766cce4ae517","Type":"ContainerStarted","Data":"a508a5af806d151a21550ebba73caf8ae175cddf2a7941e341c4d67e97c3af77"} Oct 02 02:03:39 crc kubenswrapper[4885]: I1002 02:03:39.996642 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-gp9st" event={"ID":"0feba5bf-20c5-4578-9cfc-f37d6511c2d6","Type":"ContainerStarted","Data":"31f95e8399c06106eae6157550f827c372a5d14208ea2c35ccb560d237e7e9dc"} Oct 02 02:03:39 crc kubenswrapper[4885]: I1002 02:03:39.998942 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76fcf4b695-wpgst" event={"ID":"5adbde10-fd09-4e06-bad3-7f5d7bdf5b41","Type":"ContainerStarted","Data":"11a4f72ffa16280cdb1784abaa6fcfc17a2f8881f392cb3b30731dde19a72f5a"} Oct 02 02:03:40 crc kubenswrapper[4885]: I1002 02:03:40.000237 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-xf24l" event={"ID":"c036a628-a241-45ea-bd2a-5d6d50804f61","Type":"ContainerStarted","Data":"1a1f0197c0399a742a3103b0db978338c9e66ae73dd5516c960bfa34cfe0a60a"} Oct 02 02:03:40 crc kubenswrapper[4885]: I1002 02:03:40.000294 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-xf24l" event={"ID":"c036a628-a241-45ea-bd2a-5d6d50804f61","Type":"ContainerStarted","Data":"a90523b46bfc0032de65212bd4351879dd04017043fc5b44bd795c65402bec04"} Oct 02 02:03:40 crc kubenswrapper[4885]: I1002 02:03:40.003974 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-tjcz4" event={"ID":"9c66bb6a-5a3c-4223-aba6-c5d9d0ea83ba","Type":"ContainerStarted","Data":"a265f07b2a178becf2d4f64243b607392df0bcc0cd1797f70e9daa3dd886a014"} Oct 02 02:03:40 crc kubenswrapper[4885]: I1002 02:03:40.004682 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-8558cb5fd9-8cr6r" event={"ID":"8ad39937-75ce-468c-90df-b3abbaefe498","Type":"ContainerStarted","Data":"6f02b41e568cd4c191d3ac36f5fd8d6536883727af2c4417485c66f8afc7b87f"} Oct 02 02:03:40 crc kubenswrapper[4885]: I1002 02:03:40.005696 4885 generic.go:334] "Generic (PLEG): container finished" podID="afa17394-b736-4acb-bf4f-ca79fc4908ff" containerID="9882ed5a9b9309139db854a7ed1f8ce637522b02626ea1512ec7bb751c504d85" exitCode=0 Oct 02 02:03:40 crc kubenswrapper[4885]: I1002 02:03:40.005747 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55fff446b9-84t5x" event={"ID":"afa17394-b736-4acb-bf4f-ca79fc4908ff","Type":"ContainerDied","Data":"9882ed5a9b9309139db854a7ed1f8ce637522b02626ea1512ec7bb751c504d85"} Oct 02 02:03:40 crc kubenswrapper[4885]: I1002 02:03:40.006182 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55fff446b9-84t5x" event={"ID":"afa17394-b736-4acb-bf4f-ca79fc4908ff","Type":"ContainerStarted","Data":"ad2990806d8c8db8828ee57f6f04ee72f7500fc49d4b91fd8ba0353c375d72e2"} Oct 02 02:03:40 crc kubenswrapper[4885]: I1002 02:03:40.013018 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-xf24l" podStartSLOduration=2.013000037 podStartE2EDuration="2.013000037s" podCreationTimestamp="2025-10-02 02:03:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:03:40.011579583 +0000 UTC m=+1008.823326972" watchObservedRunningTime="2025-10-02 02:03:40.013000037 +0000 UTC m=+1008.824747446" Oct 02 02:03:40 crc kubenswrapper[4885]: I1002 02:03:40.016918 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-x2nx7" event={"ID":"9fc15c77-b663-4395-ae92-922b1e4756ba","Type":"ContainerDied","Data":"859f4908b20a9cf03af8c87905ea30ffa64e16851d1dade87d6d05b7f1599118"} Oct 02 02:03:40 crc kubenswrapper[4885]: I1002 02:03:40.016977 4885 scope.go:117] "RemoveContainer" containerID="4b4ebce37f94f38134bb618a2002d6b8fb4bc1bec4fc68a0d3ab6a914e2c84a0" Oct 02 02:03:40 crc kubenswrapper[4885]: I1002 02:03:40.017211 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77585f5f8c-x2nx7" Oct 02 02:03:40 crc kubenswrapper[4885]: I1002 02:03:40.022922 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9226a04e-0195-42a1-b651-2f5379b1e2a9","Type":"ContainerStarted","Data":"471e6c4e91ce2890981d6a2361c21e1b7a21510bbff4072c0385cb579d915bbf"} Oct 02 02:03:40 crc kubenswrapper[4885]: I1002 02:03:40.033231 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-mj282" event={"ID":"9347e7b4-1d53-4a28-9cbd-6a392b8145d8","Type":"ContainerStarted","Data":"ffe87ad35a1124753436f265a479f54b35ecbb206685610a84f40e705bc72cdd"} Oct 02 02:03:40 crc kubenswrapper[4885]: I1002 02:03:40.035017 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-mj282" event={"ID":"9347e7b4-1d53-4a28-9cbd-6a392b8145d8","Type":"ContainerStarted","Data":"e62fba6b80b17d6e49407c762ec104bb0499fdb60a8faee16778d517cc1fd5ed"} Oct 02 02:03:40 crc kubenswrapper[4885]: I1002 02:03:40.064792 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-mj282" podStartSLOduration=2.064129475 podStartE2EDuration="2.064129475s" podCreationTimestamp="2025-10-02 02:03:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:03:40.053695579 +0000 UTC m=+1008.865442978" watchObservedRunningTime="2025-10-02 02:03:40.064129475 +0000 UTC m=+1008.875876884" Oct 02 02:03:40 crc kubenswrapper[4885]: I1002 02:03:40.085870 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-x2nx7"] Oct 02 02:03:40 crc kubenswrapper[4885]: I1002 02:03:40.092953 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-x2nx7"] Oct 02 02:03:40 crc kubenswrapper[4885]: I1002 02:03:40.101076 4885 scope.go:117] "RemoveContainer" containerID="fe1dec9f2387d5c95aec1c710183ab2006a76467906b5a9001f650f6227ed39d" Oct 02 02:03:40 crc kubenswrapper[4885]: I1002 02:03:40.267609 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55fff446b9-84t5x" Oct 02 02:03:40 crc kubenswrapper[4885]: I1002 02:03:40.378367 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/afa17394-b736-4acb-bf4f-ca79fc4908ff-dns-swift-storage-0\") pod \"afa17394-b736-4acb-bf4f-ca79fc4908ff\" (UID: \"afa17394-b736-4acb-bf4f-ca79fc4908ff\") " Oct 02 02:03:40 crc kubenswrapper[4885]: I1002 02:03:40.378411 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/afa17394-b736-4acb-bf4f-ca79fc4908ff-ovsdbserver-sb\") pod \"afa17394-b736-4acb-bf4f-ca79fc4908ff\" (UID: \"afa17394-b736-4acb-bf4f-ca79fc4908ff\") " Oct 02 02:03:40 crc kubenswrapper[4885]: I1002 02:03:40.378431 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/afa17394-b736-4acb-bf4f-ca79fc4908ff-config\") pod \"afa17394-b736-4acb-bf4f-ca79fc4908ff\" (UID: \"afa17394-b736-4acb-bf4f-ca79fc4908ff\") " Oct 02 02:03:40 crc kubenswrapper[4885]: I1002 02:03:40.378458 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/afa17394-b736-4acb-bf4f-ca79fc4908ff-dns-svc\") pod \"afa17394-b736-4acb-bf4f-ca79fc4908ff\" (UID: \"afa17394-b736-4acb-bf4f-ca79fc4908ff\") " Oct 02 02:03:40 crc kubenswrapper[4885]: I1002 02:03:40.378601 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/afa17394-b736-4acb-bf4f-ca79fc4908ff-ovsdbserver-nb\") pod \"afa17394-b736-4acb-bf4f-ca79fc4908ff\" (UID: \"afa17394-b736-4acb-bf4f-ca79fc4908ff\") " Oct 02 02:03:40 crc kubenswrapper[4885]: I1002 02:03:40.378618 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vx58g\" (UniqueName: \"kubernetes.io/projected/afa17394-b736-4acb-bf4f-ca79fc4908ff-kube-api-access-vx58g\") pod \"afa17394-b736-4acb-bf4f-ca79fc4908ff\" (UID: \"afa17394-b736-4acb-bf4f-ca79fc4908ff\") " Oct 02 02:03:40 crc kubenswrapper[4885]: I1002 02:03:40.387456 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/afa17394-b736-4acb-bf4f-ca79fc4908ff-kube-api-access-vx58g" (OuterVolumeSpecName: "kube-api-access-vx58g") pod "afa17394-b736-4acb-bf4f-ca79fc4908ff" (UID: "afa17394-b736-4acb-bf4f-ca79fc4908ff"). InnerVolumeSpecName "kube-api-access-vx58g". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:03:40 crc kubenswrapper[4885]: I1002 02:03:40.406363 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/afa17394-b736-4acb-bf4f-ca79fc4908ff-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "afa17394-b736-4acb-bf4f-ca79fc4908ff" (UID: "afa17394-b736-4acb-bf4f-ca79fc4908ff"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:03:40 crc kubenswrapper[4885]: I1002 02:03:40.409420 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/afa17394-b736-4acb-bf4f-ca79fc4908ff-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "afa17394-b736-4acb-bf4f-ca79fc4908ff" (UID: "afa17394-b736-4acb-bf4f-ca79fc4908ff"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:03:40 crc kubenswrapper[4885]: I1002 02:03:40.413003 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/afa17394-b736-4acb-bf4f-ca79fc4908ff-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "afa17394-b736-4acb-bf4f-ca79fc4908ff" (UID: "afa17394-b736-4acb-bf4f-ca79fc4908ff"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:03:40 crc kubenswrapper[4885]: I1002 02:03:40.419620 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/afa17394-b736-4acb-bf4f-ca79fc4908ff-config" (OuterVolumeSpecName: "config") pod "afa17394-b736-4acb-bf4f-ca79fc4908ff" (UID: "afa17394-b736-4acb-bf4f-ca79fc4908ff"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:03:40 crc kubenswrapper[4885]: I1002 02:03:40.419846 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/afa17394-b736-4acb-bf4f-ca79fc4908ff-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "afa17394-b736-4acb-bf4f-ca79fc4908ff" (UID: "afa17394-b736-4acb-bf4f-ca79fc4908ff"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:03:40 crc kubenswrapper[4885]: I1002 02:03:40.480114 4885 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/afa17394-b736-4acb-bf4f-ca79fc4908ff-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:40 crc kubenswrapper[4885]: I1002 02:03:40.480141 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vx58g\" (UniqueName: \"kubernetes.io/projected/afa17394-b736-4acb-bf4f-ca79fc4908ff-kube-api-access-vx58g\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:40 crc kubenswrapper[4885]: I1002 02:03:40.480154 4885 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/afa17394-b736-4acb-bf4f-ca79fc4908ff-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:40 crc kubenswrapper[4885]: I1002 02:03:40.480163 4885 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/afa17394-b736-4acb-bf4f-ca79fc4908ff-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:40 crc kubenswrapper[4885]: I1002 02:03:40.480173 4885 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/afa17394-b736-4acb-bf4f-ca79fc4908ff-config\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:40 crc kubenswrapper[4885]: I1002 02:03:40.480181 4885 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/afa17394-b736-4acb-bf4f-ca79fc4908ff-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:41 crc kubenswrapper[4885]: I1002 02:03:41.050785 4885 generic.go:334] "Generic (PLEG): container finished" podID="5adbde10-fd09-4e06-bad3-7f5d7bdf5b41" containerID="c6a5a15dcf6954c5e3e2d30d5a3cf3359566f0e2eaed48ec5e4bd65bef682e54" exitCode=0 Oct 02 02:03:41 crc kubenswrapper[4885]: I1002 02:03:41.050846 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76fcf4b695-wpgst" event={"ID":"5adbde10-fd09-4e06-bad3-7f5d7bdf5b41","Type":"ContainerDied","Data":"c6a5a15dcf6954c5e3e2d30d5a3cf3359566f0e2eaed48ec5e4bd65bef682e54"} Oct 02 02:03:41 crc kubenswrapper[4885]: I1002 02:03:41.052029 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-vj2tf" event={"ID":"8fa66af9-7b4b-480c-8f96-bda850e2eef6","Type":"ContainerStarted","Data":"55414b8c8eb0471f8443ae7434426584d5315281a20ef9fd8540ffc929bb4f50"} Oct 02 02:03:41 crc kubenswrapper[4885]: I1002 02:03:41.055158 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55fff446b9-84t5x" event={"ID":"afa17394-b736-4acb-bf4f-ca79fc4908ff","Type":"ContainerDied","Data":"ad2990806d8c8db8828ee57f6f04ee72f7500fc49d4b91fd8ba0353c375d72e2"} Oct 02 02:03:41 crc kubenswrapper[4885]: I1002 02:03:41.055199 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55fff446b9-84t5x" Oct 02 02:03:41 crc kubenswrapper[4885]: I1002 02:03:41.055211 4885 scope.go:117] "RemoveContainer" containerID="9882ed5a9b9309139db854a7ed1f8ce637522b02626ea1512ec7bb751c504d85" Oct 02 02:03:41 crc kubenswrapper[4885]: I1002 02:03:41.202769 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55fff446b9-84t5x"] Oct 02 02:03:41 crc kubenswrapper[4885]: I1002 02:03:41.204875 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-55fff446b9-84t5x"] Oct 02 02:03:42 crc kubenswrapper[4885]: I1002 02:03:42.064483 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9fc15c77-b663-4395-ae92-922b1e4756ba" path="/var/lib/kubelet/pods/9fc15c77-b663-4395-ae92-922b1e4756ba/volumes" Oct 02 02:03:42 crc kubenswrapper[4885]: I1002 02:03:42.065351 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="afa17394-b736-4acb-bf4f-ca79fc4908ff" path="/var/lib/kubelet/pods/afa17394-b736-4acb-bf4f-ca79fc4908ff/volumes" Oct 02 02:03:42 crc kubenswrapper[4885]: I1002 02:03:42.082419 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 02:03:42 crc kubenswrapper[4885]: I1002 02:03:42.083392 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76fcf4b695-wpgst" event={"ID":"5adbde10-fd09-4e06-bad3-7f5d7bdf5b41","Type":"ContainerStarted","Data":"aff3d2a87f29168594522e9b4f20c0cbbd55852066fe579d0b5a475463828e49"} Oct 02 02:03:42 crc kubenswrapper[4885]: I1002 02:03:42.084180 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-76fcf4b695-wpgst" Oct 02 02:03:42 crc kubenswrapper[4885]: I1002 02:03:42.133786 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-79659d7df7-6dkj8"] Oct 02 02:03:42 crc kubenswrapper[4885]: I1002 02:03:42.192427 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-5686549c7f-j9kgc"] Oct 02 02:03:42 crc kubenswrapper[4885]: E1002 02:03:42.192818 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9fc15c77-b663-4395-ae92-922b1e4756ba" containerName="init" Oct 02 02:03:42 crc kubenswrapper[4885]: I1002 02:03:42.192830 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="9fc15c77-b663-4395-ae92-922b1e4756ba" containerName="init" Oct 02 02:03:42 crc kubenswrapper[4885]: E1002 02:03:42.192843 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afa17394-b736-4acb-bf4f-ca79fc4908ff" containerName="init" Oct 02 02:03:42 crc kubenswrapper[4885]: I1002 02:03:42.192849 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="afa17394-b736-4acb-bf4f-ca79fc4908ff" containerName="init" Oct 02 02:03:42 crc kubenswrapper[4885]: E1002 02:03:42.192862 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9fc15c77-b663-4395-ae92-922b1e4756ba" containerName="dnsmasq-dns" Oct 02 02:03:42 crc kubenswrapper[4885]: I1002 02:03:42.192869 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="9fc15c77-b663-4395-ae92-922b1e4756ba" containerName="dnsmasq-dns" Oct 02 02:03:42 crc kubenswrapper[4885]: I1002 02:03:42.193023 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="afa17394-b736-4acb-bf4f-ca79fc4908ff" containerName="init" Oct 02 02:03:42 crc kubenswrapper[4885]: I1002 02:03:42.193040 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="9fc15c77-b663-4395-ae92-922b1e4756ba" containerName="dnsmasq-dns" Oct 02 02:03:42 crc kubenswrapper[4885]: I1002 02:03:42.193903 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5686549c7f-j9kgc" Oct 02 02:03:42 crc kubenswrapper[4885]: I1002 02:03:42.201934 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5686549c7f-j9kgc"] Oct 02 02:03:42 crc kubenswrapper[4885]: I1002 02:03:42.261763 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-76fcf4b695-wpgst" podStartSLOduration=4.261740536 podStartE2EDuration="4.261740536s" podCreationTimestamp="2025-10-02 02:03:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:03:42.255441906 +0000 UTC m=+1011.067189305" watchObservedRunningTime="2025-10-02 02:03:42.261740536 +0000 UTC m=+1011.073487935" Oct 02 02:03:42 crc kubenswrapper[4885]: I1002 02:03:42.347111 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5056f944-c3cf-45e1-99f5-f93132717b93-scripts\") pod \"horizon-5686549c7f-j9kgc\" (UID: \"5056f944-c3cf-45e1-99f5-f93132717b93\") " pod="openstack/horizon-5686549c7f-j9kgc" Oct 02 02:03:42 crc kubenswrapper[4885]: I1002 02:03:42.347160 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5056f944-c3cf-45e1-99f5-f93132717b93-config-data\") pod \"horizon-5686549c7f-j9kgc\" (UID: \"5056f944-c3cf-45e1-99f5-f93132717b93\") " pod="openstack/horizon-5686549c7f-j9kgc" Oct 02 02:03:42 crc kubenswrapper[4885]: I1002 02:03:42.347196 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5056f944-c3cf-45e1-99f5-f93132717b93-logs\") pod \"horizon-5686549c7f-j9kgc\" (UID: \"5056f944-c3cf-45e1-99f5-f93132717b93\") " pod="openstack/horizon-5686549c7f-j9kgc" Oct 02 02:03:42 crc kubenswrapper[4885]: I1002 02:03:42.347240 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/5056f944-c3cf-45e1-99f5-f93132717b93-horizon-secret-key\") pod \"horizon-5686549c7f-j9kgc\" (UID: \"5056f944-c3cf-45e1-99f5-f93132717b93\") " pod="openstack/horizon-5686549c7f-j9kgc" Oct 02 02:03:42 crc kubenswrapper[4885]: I1002 02:03:42.347318 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gx8x4\" (UniqueName: \"kubernetes.io/projected/5056f944-c3cf-45e1-99f5-f93132717b93-kube-api-access-gx8x4\") pod \"horizon-5686549c7f-j9kgc\" (UID: \"5056f944-c3cf-45e1-99f5-f93132717b93\") " pod="openstack/horizon-5686549c7f-j9kgc" Oct 02 02:03:42 crc kubenswrapper[4885]: I1002 02:03:42.448342 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5056f944-c3cf-45e1-99f5-f93132717b93-scripts\") pod \"horizon-5686549c7f-j9kgc\" (UID: \"5056f944-c3cf-45e1-99f5-f93132717b93\") " pod="openstack/horizon-5686549c7f-j9kgc" Oct 02 02:03:42 crc kubenswrapper[4885]: I1002 02:03:42.448645 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5056f944-c3cf-45e1-99f5-f93132717b93-config-data\") pod \"horizon-5686549c7f-j9kgc\" (UID: \"5056f944-c3cf-45e1-99f5-f93132717b93\") " pod="openstack/horizon-5686549c7f-j9kgc" Oct 02 02:03:42 crc kubenswrapper[4885]: I1002 02:03:42.448687 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5056f944-c3cf-45e1-99f5-f93132717b93-logs\") pod \"horizon-5686549c7f-j9kgc\" (UID: \"5056f944-c3cf-45e1-99f5-f93132717b93\") " pod="openstack/horizon-5686549c7f-j9kgc" Oct 02 02:03:42 crc kubenswrapper[4885]: I1002 02:03:42.448728 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/5056f944-c3cf-45e1-99f5-f93132717b93-horizon-secret-key\") pod \"horizon-5686549c7f-j9kgc\" (UID: \"5056f944-c3cf-45e1-99f5-f93132717b93\") " pod="openstack/horizon-5686549c7f-j9kgc" Oct 02 02:03:42 crc kubenswrapper[4885]: I1002 02:03:42.448788 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gx8x4\" (UniqueName: \"kubernetes.io/projected/5056f944-c3cf-45e1-99f5-f93132717b93-kube-api-access-gx8x4\") pod \"horizon-5686549c7f-j9kgc\" (UID: \"5056f944-c3cf-45e1-99f5-f93132717b93\") " pod="openstack/horizon-5686549c7f-j9kgc" Oct 02 02:03:42 crc kubenswrapper[4885]: I1002 02:03:42.449100 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5056f944-c3cf-45e1-99f5-f93132717b93-scripts\") pod \"horizon-5686549c7f-j9kgc\" (UID: \"5056f944-c3cf-45e1-99f5-f93132717b93\") " pod="openstack/horizon-5686549c7f-j9kgc" Oct 02 02:03:42 crc kubenswrapper[4885]: I1002 02:03:42.449385 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5056f944-c3cf-45e1-99f5-f93132717b93-logs\") pod \"horizon-5686549c7f-j9kgc\" (UID: \"5056f944-c3cf-45e1-99f5-f93132717b93\") " pod="openstack/horizon-5686549c7f-j9kgc" Oct 02 02:03:42 crc kubenswrapper[4885]: I1002 02:03:42.450468 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5056f944-c3cf-45e1-99f5-f93132717b93-config-data\") pod \"horizon-5686549c7f-j9kgc\" (UID: \"5056f944-c3cf-45e1-99f5-f93132717b93\") " pod="openstack/horizon-5686549c7f-j9kgc" Oct 02 02:03:42 crc kubenswrapper[4885]: I1002 02:03:42.466933 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/5056f944-c3cf-45e1-99f5-f93132717b93-horizon-secret-key\") pod \"horizon-5686549c7f-j9kgc\" (UID: \"5056f944-c3cf-45e1-99f5-f93132717b93\") " pod="openstack/horizon-5686549c7f-j9kgc" Oct 02 02:03:42 crc kubenswrapper[4885]: I1002 02:03:42.471855 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gx8x4\" (UniqueName: \"kubernetes.io/projected/5056f944-c3cf-45e1-99f5-f93132717b93-kube-api-access-gx8x4\") pod \"horizon-5686549c7f-j9kgc\" (UID: \"5056f944-c3cf-45e1-99f5-f93132717b93\") " pod="openstack/horizon-5686549c7f-j9kgc" Oct 02 02:03:42 crc kubenswrapper[4885]: I1002 02:03:42.512559 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5686549c7f-j9kgc" Oct 02 02:03:43 crc kubenswrapper[4885]: I1002 02:03:43.110336 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5686549c7f-j9kgc"] Oct 02 02:03:43 crc kubenswrapper[4885]: W1002 02:03:43.113190 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5056f944_c3cf_45e1_99f5_f93132717b93.slice/crio-ec729febcc65ceb7baec51b7003eb6b379eac5a90a92b9c59151858053988768 WatchSource:0}: Error finding container ec729febcc65ceb7baec51b7003eb6b379eac5a90a92b9c59151858053988768: Status 404 returned error can't find the container with id ec729febcc65ceb7baec51b7003eb6b379eac5a90a92b9c59151858053988768 Oct 02 02:03:44 crc kubenswrapper[4885]: I1002 02:03:44.109016 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5686549c7f-j9kgc" event={"ID":"5056f944-c3cf-45e1-99f5-f93132717b93","Type":"ContainerStarted","Data":"ec729febcc65ceb7baec51b7003eb6b379eac5a90a92b9c59151858053988768"} Oct 02 02:03:47 crc kubenswrapper[4885]: I1002 02:03:47.620792 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-8558cb5fd9-8cr6r"] Oct 02 02:03:47 crc kubenswrapper[4885]: I1002 02:03:47.649845 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7c49b4cffb-brdz9"] Oct 02 02:03:47 crc kubenswrapper[4885]: I1002 02:03:47.651174 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7c49b4cffb-brdz9" Oct 02 02:03:47 crc kubenswrapper[4885]: I1002 02:03:47.654888 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Oct 02 02:03:47 crc kubenswrapper[4885]: I1002 02:03:47.667173 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7c49b4cffb-brdz9"] Oct 02 02:03:47 crc kubenswrapper[4885]: I1002 02:03:47.738805 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5686549c7f-j9kgc"] Oct 02 02:03:47 crc kubenswrapper[4885]: I1002 02:03:47.751181 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-6d7c47744-vjcbs"] Oct 02 02:03:47 crc kubenswrapper[4885]: I1002 02:03:47.759106 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b0fddcb3-3649-48f4-aa8f-85dd1c8137cf-scripts\") pod \"horizon-7c49b4cffb-brdz9\" (UID: \"b0fddcb3-3649-48f4-aa8f-85dd1c8137cf\") " pod="openstack/horizon-7c49b4cffb-brdz9" Oct 02 02:03:47 crc kubenswrapper[4885]: I1002 02:03:47.759157 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/b0fddcb3-3649-48f4-aa8f-85dd1c8137cf-horizon-tls-certs\") pod \"horizon-7c49b4cffb-brdz9\" (UID: \"b0fddcb3-3649-48f4-aa8f-85dd1c8137cf\") " pod="openstack/horizon-7c49b4cffb-brdz9" Oct 02 02:03:47 crc kubenswrapper[4885]: I1002 02:03:47.759182 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pmfgm\" (UniqueName: \"kubernetes.io/projected/b0fddcb3-3649-48f4-aa8f-85dd1c8137cf-kube-api-access-pmfgm\") pod \"horizon-7c49b4cffb-brdz9\" (UID: \"b0fddcb3-3649-48f4-aa8f-85dd1c8137cf\") " pod="openstack/horizon-7c49b4cffb-brdz9" Oct 02 02:03:47 crc kubenswrapper[4885]: I1002 02:03:47.759199 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0fddcb3-3649-48f4-aa8f-85dd1c8137cf-combined-ca-bundle\") pod \"horizon-7c49b4cffb-brdz9\" (UID: \"b0fddcb3-3649-48f4-aa8f-85dd1c8137cf\") " pod="openstack/horizon-7c49b4cffb-brdz9" Oct 02 02:03:47 crc kubenswrapper[4885]: I1002 02:03:47.759239 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b0fddcb3-3649-48f4-aa8f-85dd1c8137cf-logs\") pod \"horizon-7c49b4cffb-brdz9\" (UID: \"b0fddcb3-3649-48f4-aa8f-85dd1c8137cf\") " pod="openstack/horizon-7c49b4cffb-brdz9" Oct 02 02:03:47 crc kubenswrapper[4885]: I1002 02:03:47.759289 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b0fddcb3-3649-48f4-aa8f-85dd1c8137cf-config-data\") pod \"horizon-7c49b4cffb-brdz9\" (UID: \"b0fddcb3-3649-48f4-aa8f-85dd1c8137cf\") " pod="openstack/horizon-7c49b4cffb-brdz9" Oct 02 02:03:47 crc kubenswrapper[4885]: I1002 02:03:47.759317 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/b0fddcb3-3649-48f4-aa8f-85dd1c8137cf-horizon-secret-key\") pod \"horizon-7c49b4cffb-brdz9\" (UID: \"b0fddcb3-3649-48f4-aa8f-85dd1c8137cf\") " pod="openstack/horizon-7c49b4cffb-brdz9" Oct 02 02:03:47 crc kubenswrapper[4885]: I1002 02:03:47.761876 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6d7c47744-vjcbs"] Oct 02 02:03:47 crc kubenswrapper[4885]: I1002 02:03:47.761983 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6d7c47744-vjcbs" Oct 02 02:03:47 crc kubenswrapper[4885]: I1002 02:03:47.860269 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b0fddcb3-3649-48f4-aa8f-85dd1c8137cf-logs\") pod \"horizon-7c49b4cffb-brdz9\" (UID: \"b0fddcb3-3649-48f4-aa8f-85dd1c8137cf\") " pod="openstack/horizon-7c49b4cffb-brdz9" Oct 02 02:03:47 crc kubenswrapper[4885]: I1002 02:03:47.860311 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/607de89d-b68e-49e8-beb4-7664a37c6105-logs\") pod \"horizon-6d7c47744-vjcbs\" (UID: \"607de89d-b68e-49e8-beb4-7664a37c6105\") " pod="openstack/horizon-6d7c47744-vjcbs" Oct 02 02:03:47 crc kubenswrapper[4885]: I1002 02:03:47.860352 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b0fddcb3-3649-48f4-aa8f-85dd1c8137cf-config-data\") pod \"horizon-7c49b4cffb-brdz9\" (UID: \"b0fddcb3-3649-48f4-aa8f-85dd1c8137cf\") " pod="openstack/horizon-7c49b4cffb-brdz9" Oct 02 02:03:47 crc kubenswrapper[4885]: I1002 02:03:47.860382 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/b0fddcb3-3649-48f4-aa8f-85dd1c8137cf-horizon-secret-key\") pod \"horizon-7c49b4cffb-brdz9\" (UID: \"b0fddcb3-3649-48f4-aa8f-85dd1c8137cf\") " pod="openstack/horizon-7c49b4cffb-brdz9" Oct 02 02:03:47 crc kubenswrapper[4885]: I1002 02:03:47.860405 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vtmrn\" (UniqueName: \"kubernetes.io/projected/607de89d-b68e-49e8-beb4-7664a37c6105-kube-api-access-vtmrn\") pod \"horizon-6d7c47744-vjcbs\" (UID: \"607de89d-b68e-49e8-beb4-7664a37c6105\") " pod="openstack/horizon-6d7c47744-vjcbs" Oct 02 02:03:47 crc kubenswrapper[4885]: I1002 02:03:47.860444 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/607de89d-b68e-49e8-beb4-7664a37c6105-combined-ca-bundle\") pod \"horizon-6d7c47744-vjcbs\" (UID: \"607de89d-b68e-49e8-beb4-7664a37c6105\") " pod="openstack/horizon-6d7c47744-vjcbs" Oct 02 02:03:47 crc kubenswrapper[4885]: I1002 02:03:47.860461 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/607de89d-b68e-49e8-beb4-7664a37c6105-horizon-tls-certs\") pod \"horizon-6d7c47744-vjcbs\" (UID: \"607de89d-b68e-49e8-beb4-7664a37c6105\") " pod="openstack/horizon-6d7c47744-vjcbs" Oct 02 02:03:47 crc kubenswrapper[4885]: I1002 02:03:47.860482 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/607de89d-b68e-49e8-beb4-7664a37c6105-horizon-secret-key\") pod \"horizon-6d7c47744-vjcbs\" (UID: \"607de89d-b68e-49e8-beb4-7664a37c6105\") " pod="openstack/horizon-6d7c47744-vjcbs" Oct 02 02:03:47 crc kubenswrapper[4885]: I1002 02:03:47.860499 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/607de89d-b68e-49e8-beb4-7664a37c6105-scripts\") pod \"horizon-6d7c47744-vjcbs\" (UID: \"607de89d-b68e-49e8-beb4-7664a37c6105\") " pod="openstack/horizon-6d7c47744-vjcbs" Oct 02 02:03:47 crc kubenswrapper[4885]: I1002 02:03:47.860520 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/607de89d-b68e-49e8-beb4-7664a37c6105-config-data\") pod \"horizon-6d7c47744-vjcbs\" (UID: \"607de89d-b68e-49e8-beb4-7664a37c6105\") " pod="openstack/horizon-6d7c47744-vjcbs" Oct 02 02:03:47 crc kubenswrapper[4885]: I1002 02:03:47.860541 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b0fddcb3-3649-48f4-aa8f-85dd1c8137cf-scripts\") pod \"horizon-7c49b4cffb-brdz9\" (UID: \"b0fddcb3-3649-48f4-aa8f-85dd1c8137cf\") " pod="openstack/horizon-7c49b4cffb-brdz9" Oct 02 02:03:47 crc kubenswrapper[4885]: I1002 02:03:47.860563 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/b0fddcb3-3649-48f4-aa8f-85dd1c8137cf-horizon-tls-certs\") pod \"horizon-7c49b4cffb-brdz9\" (UID: \"b0fddcb3-3649-48f4-aa8f-85dd1c8137cf\") " pod="openstack/horizon-7c49b4cffb-brdz9" Oct 02 02:03:47 crc kubenswrapper[4885]: I1002 02:03:47.860581 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pmfgm\" (UniqueName: \"kubernetes.io/projected/b0fddcb3-3649-48f4-aa8f-85dd1c8137cf-kube-api-access-pmfgm\") pod \"horizon-7c49b4cffb-brdz9\" (UID: \"b0fddcb3-3649-48f4-aa8f-85dd1c8137cf\") " pod="openstack/horizon-7c49b4cffb-brdz9" Oct 02 02:03:47 crc kubenswrapper[4885]: I1002 02:03:47.860603 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0fddcb3-3649-48f4-aa8f-85dd1c8137cf-combined-ca-bundle\") pod \"horizon-7c49b4cffb-brdz9\" (UID: \"b0fddcb3-3649-48f4-aa8f-85dd1c8137cf\") " pod="openstack/horizon-7c49b4cffb-brdz9" Oct 02 02:03:47 crc kubenswrapper[4885]: I1002 02:03:47.864709 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b0fddcb3-3649-48f4-aa8f-85dd1c8137cf-logs\") pod \"horizon-7c49b4cffb-brdz9\" (UID: \"b0fddcb3-3649-48f4-aa8f-85dd1c8137cf\") " pod="openstack/horizon-7c49b4cffb-brdz9" Oct 02 02:03:47 crc kubenswrapper[4885]: I1002 02:03:47.865739 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b0fddcb3-3649-48f4-aa8f-85dd1c8137cf-config-data\") pod \"horizon-7c49b4cffb-brdz9\" (UID: \"b0fddcb3-3649-48f4-aa8f-85dd1c8137cf\") " pod="openstack/horizon-7c49b4cffb-brdz9" Oct 02 02:03:47 crc kubenswrapper[4885]: I1002 02:03:47.866176 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b0fddcb3-3649-48f4-aa8f-85dd1c8137cf-scripts\") pod \"horizon-7c49b4cffb-brdz9\" (UID: \"b0fddcb3-3649-48f4-aa8f-85dd1c8137cf\") " pod="openstack/horizon-7c49b4cffb-brdz9" Oct 02 02:03:47 crc kubenswrapper[4885]: I1002 02:03:47.870433 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/b0fddcb3-3649-48f4-aa8f-85dd1c8137cf-horizon-secret-key\") pod \"horizon-7c49b4cffb-brdz9\" (UID: \"b0fddcb3-3649-48f4-aa8f-85dd1c8137cf\") " pod="openstack/horizon-7c49b4cffb-brdz9" Oct 02 02:03:47 crc kubenswrapper[4885]: I1002 02:03:47.872093 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/b0fddcb3-3649-48f4-aa8f-85dd1c8137cf-horizon-tls-certs\") pod \"horizon-7c49b4cffb-brdz9\" (UID: \"b0fddcb3-3649-48f4-aa8f-85dd1c8137cf\") " pod="openstack/horizon-7c49b4cffb-brdz9" Oct 02 02:03:47 crc kubenswrapper[4885]: I1002 02:03:47.880250 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pmfgm\" (UniqueName: \"kubernetes.io/projected/b0fddcb3-3649-48f4-aa8f-85dd1c8137cf-kube-api-access-pmfgm\") pod \"horizon-7c49b4cffb-brdz9\" (UID: \"b0fddcb3-3649-48f4-aa8f-85dd1c8137cf\") " pod="openstack/horizon-7c49b4cffb-brdz9" Oct 02 02:03:47 crc kubenswrapper[4885]: I1002 02:03:47.894036 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0fddcb3-3649-48f4-aa8f-85dd1c8137cf-combined-ca-bundle\") pod \"horizon-7c49b4cffb-brdz9\" (UID: \"b0fddcb3-3649-48f4-aa8f-85dd1c8137cf\") " pod="openstack/horizon-7c49b4cffb-brdz9" Oct 02 02:03:47 crc kubenswrapper[4885]: I1002 02:03:47.961682 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/607de89d-b68e-49e8-beb4-7664a37c6105-horizon-tls-certs\") pod \"horizon-6d7c47744-vjcbs\" (UID: \"607de89d-b68e-49e8-beb4-7664a37c6105\") " pod="openstack/horizon-6d7c47744-vjcbs" Oct 02 02:03:47 crc kubenswrapper[4885]: I1002 02:03:47.961747 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/607de89d-b68e-49e8-beb4-7664a37c6105-horizon-secret-key\") pod \"horizon-6d7c47744-vjcbs\" (UID: \"607de89d-b68e-49e8-beb4-7664a37c6105\") " pod="openstack/horizon-6d7c47744-vjcbs" Oct 02 02:03:47 crc kubenswrapper[4885]: I1002 02:03:47.961776 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/607de89d-b68e-49e8-beb4-7664a37c6105-scripts\") pod \"horizon-6d7c47744-vjcbs\" (UID: \"607de89d-b68e-49e8-beb4-7664a37c6105\") " pod="openstack/horizon-6d7c47744-vjcbs" Oct 02 02:03:47 crc kubenswrapper[4885]: I1002 02:03:47.961813 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/607de89d-b68e-49e8-beb4-7664a37c6105-config-data\") pod \"horizon-6d7c47744-vjcbs\" (UID: \"607de89d-b68e-49e8-beb4-7664a37c6105\") " pod="openstack/horizon-6d7c47744-vjcbs" Oct 02 02:03:47 crc kubenswrapper[4885]: I1002 02:03:47.961894 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/607de89d-b68e-49e8-beb4-7664a37c6105-logs\") pod \"horizon-6d7c47744-vjcbs\" (UID: \"607de89d-b68e-49e8-beb4-7664a37c6105\") " pod="openstack/horizon-6d7c47744-vjcbs" Oct 02 02:03:47 crc kubenswrapper[4885]: I1002 02:03:47.961966 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vtmrn\" (UniqueName: \"kubernetes.io/projected/607de89d-b68e-49e8-beb4-7664a37c6105-kube-api-access-vtmrn\") pod \"horizon-6d7c47744-vjcbs\" (UID: \"607de89d-b68e-49e8-beb4-7664a37c6105\") " pod="openstack/horizon-6d7c47744-vjcbs" Oct 02 02:03:47 crc kubenswrapper[4885]: I1002 02:03:47.962014 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/607de89d-b68e-49e8-beb4-7664a37c6105-combined-ca-bundle\") pod \"horizon-6d7c47744-vjcbs\" (UID: \"607de89d-b68e-49e8-beb4-7664a37c6105\") " pod="openstack/horizon-6d7c47744-vjcbs" Oct 02 02:03:47 crc kubenswrapper[4885]: I1002 02:03:47.962930 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/607de89d-b68e-49e8-beb4-7664a37c6105-scripts\") pod \"horizon-6d7c47744-vjcbs\" (UID: \"607de89d-b68e-49e8-beb4-7664a37c6105\") " pod="openstack/horizon-6d7c47744-vjcbs" Oct 02 02:03:47 crc kubenswrapper[4885]: I1002 02:03:47.964005 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/607de89d-b68e-49e8-beb4-7664a37c6105-config-data\") pod \"horizon-6d7c47744-vjcbs\" (UID: \"607de89d-b68e-49e8-beb4-7664a37c6105\") " pod="openstack/horizon-6d7c47744-vjcbs" Oct 02 02:03:47 crc kubenswrapper[4885]: I1002 02:03:47.964251 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/607de89d-b68e-49e8-beb4-7664a37c6105-logs\") pod \"horizon-6d7c47744-vjcbs\" (UID: \"607de89d-b68e-49e8-beb4-7664a37c6105\") " pod="openstack/horizon-6d7c47744-vjcbs" Oct 02 02:03:47 crc kubenswrapper[4885]: I1002 02:03:47.965538 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/607de89d-b68e-49e8-beb4-7664a37c6105-horizon-secret-key\") pod \"horizon-6d7c47744-vjcbs\" (UID: \"607de89d-b68e-49e8-beb4-7664a37c6105\") " pod="openstack/horizon-6d7c47744-vjcbs" Oct 02 02:03:47 crc kubenswrapper[4885]: I1002 02:03:47.966239 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/607de89d-b68e-49e8-beb4-7664a37c6105-horizon-tls-certs\") pod \"horizon-6d7c47744-vjcbs\" (UID: \"607de89d-b68e-49e8-beb4-7664a37c6105\") " pod="openstack/horizon-6d7c47744-vjcbs" Oct 02 02:03:47 crc kubenswrapper[4885]: I1002 02:03:47.966793 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/607de89d-b68e-49e8-beb4-7664a37c6105-combined-ca-bundle\") pod \"horizon-6d7c47744-vjcbs\" (UID: \"607de89d-b68e-49e8-beb4-7664a37c6105\") " pod="openstack/horizon-6d7c47744-vjcbs" Oct 02 02:03:47 crc kubenswrapper[4885]: I1002 02:03:47.983933 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7c49b4cffb-brdz9" Oct 02 02:03:47 crc kubenswrapper[4885]: I1002 02:03:47.984210 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vtmrn\" (UniqueName: \"kubernetes.io/projected/607de89d-b68e-49e8-beb4-7664a37c6105-kube-api-access-vtmrn\") pod \"horizon-6d7c47744-vjcbs\" (UID: \"607de89d-b68e-49e8-beb4-7664a37c6105\") " pod="openstack/horizon-6d7c47744-vjcbs" Oct 02 02:03:48 crc kubenswrapper[4885]: I1002 02:03:48.082365 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6d7c47744-vjcbs" Oct 02 02:03:49 crc kubenswrapper[4885]: I1002 02:03:49.353853 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-76fcf4b695-wpgst" Oct 02 02:03:49 crc kubenswrapper[4885]: I1002 02:03:49.436226 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-qkx2j"] Oct 02 02:03:49 crc kubenswrapper[4885]: I1002 02:03:49.436458 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-698758b865-qkx2j" podUID="52b80336-4c74-4236-9600-7ea4606e5d6f" containerName="dnsmasq-dns" containerID="cri-o://13d632d093ec39287feac2e2cd715160da8bd3673c8eb579a58d27664223a705" gracePeriod=10 Oct 02 02:03:49 crc kubenswrapper[4885]: I1002 02:03:49.949461 4885 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-698758b865-qkx2j" podUID="52b80336-4c74-4236-9600-7ea4606e5d6f" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.114:5353: connect: connection refused" Oct 02 02:03:50 crc kubenswrapper[4885]: I1002 02:03:50.180111 4885 generic.go:334] "Generic (PLEG): container finished" podID="9347e7b4-1d53-4a28-9cbd-6a392b8145d8" containerID="ffe87ad35a1124753436f265a479f54b35ecbb206685610a84f40e705bc72cdd" exitCode=0 Oct 02 02:03:50 crc kubenswrapper[4885]: I1002 02:03:50.180209 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-mj282" event={"ID":"9347e7b4-1d53-4a28-9cbd-6a392b8145d8","Type":"ContainerDied","Data":"ffe87ad35a1124753436f265a479f54b35ecbb206685610a84f40e705bc72cdd"} Oct 02 02:03:50 crc kubenswrapper[4885]: I1002 02:03:50.182010 4885 generic.go:334] "Generic (PLEG): container finished" podID="52b80336-4c74-4236-9600-7ea4606e5d6f" containerID="13d632d093ec39287feac2e2cd715160da8bd3673c8eb579a58d27664223a705" exitCode=0 Oct 02 02:03:50 crc kubenswrapper[4885]: I1002 02:03:50.182041 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-qkx2j" event={"ID":"52b80336-4c74-4236-9600-7ea4606e5d6f","Type":"ContainerDied","Data":"13d632d093ec39287feac2e2cd715160da8bd3673c8eb579a58d27664223a705"} Oct 02 02:03:54 crc kubenswrapper[4885]: I1002 02:03:54.949502 4885 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-698758b865-qkx2j" podUID="52b80336-4c74-4236-9600-7ea4606e5d6f" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.114:5353: connect: connection refused" Oct 02 02:03:58 crc kubenswrapper[4885]: I1002 02:03:58.280154 4885 generic.go:334] "Generic (PLEG): container finished" podID="6a9ed972-b8c7-4ff7-af7b-899842ecc19c" containerID="8d716973aace109fb672bfa98c5dffccc3dd5fa5f26180d6b4abf24b760c199c" exitCode=0 Oct 02 02:03:58 crc kubenswrapper[4885]: I1002 02:03:58.280704 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-spxxt" event={"ID":"6a9ed972-b8c7-4ff7-af7b-899842ecc19c","Type":"ContainerDied","Data":"8d716973aace109fb672bfa98c5dffccc3dd5fa5f26180d6b4abf24b760c199c"} Oct 02 02:03:59 crc kubenswrapper[4885]: I1002 02:03:59.949754 4885 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-698758b865-qkx2j" podUID="52b80336-4c74-4236-9600-7ea4606e5d6f" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.114:5353: connect: connection refused" Oct 02 02:03:59 crc kubenswrapper[4885]: I1002 02:03:59.950395 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-698758b865-qkx2j" Oct 02 02:04:02 crc kubenswrapper[4885]: E1002 02:04:02.831433 4885 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-placement-api:current-podified" Oct 02 02:04:02 crc kubenswrapper[4885]: E1002 02:04:02.832038 4885 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:placement-db-sync,Image:quay.io/podified-antelope-centos9/openstack-placement-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/placement,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:placement-dbsync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fpmmt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42482,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-db-sync-vj2tf_openstack(8fa66af9-7b4b-480c-8f96-bda850e2eef6): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 02:04:02 crc kubenswrapper[4885]: E1002 02:04:02.833990 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/placement-db-sync-vj2tf" podUID="8fa66af9-7b4b-480c-8f96-bda850e2eef6" Oct 02 02:04:03 crc kubenswrapper[4885]: E1002 02:04:03.343425 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-placement-api:current-podified\\\"\"" pod="openstack/placement-db-sync-vj2tf" podUID="8fa66af9-7b4b-480c-8f96-bda850e2eef6" Oct 02 02:04:04 crc kubenswrapper[4885]: I1002 02:04:04.478857 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-mj282" Oct 02 02:04:04 crc kubenswrapper[4885]: I1002 02:04:04.612512 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9347e7b4-1d53-4a28-9cbd-6a392b8145d8-config-data\") pod \"9347e7b4-1d53-4a28-9cbd-6a392b8145d8\" (UID: \"9347e7b4-1d53-4a28-9cbd-6a392b8145d8\") " Oct 02 02:04:04 crc kubenswrapper[4885]: I1002 02:04:04.612557 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9347e7b4-1d53-4a28-9cbd-6a392b8145d8-fernet-keys\") pod \"9347e7b4-1d53-4a28-9cbd-6a392b8145d8\" (UID: \"9347e7b4-1d53-4a28-9cbd-6a392b8145d8\") " Oct 02 02:04:04 crc kubenswrapper[4885]: I1002 02:04:04.612623 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9347e7b4-1d53-4a28-9cbd-6a392b8145d8-combined-ca-bundle\") pod \"9347e7b4-1d53-4a28-9cbd-6a392b8145d8\" (UID: \"9347e7b4-1d53-4a28-9cbd-6a392b8145d8\") " Oct 02 02:04:04 crc kubenswrapper[4885]: I1002 02:04:04.612707 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-72dzl\" (UniqueName: \"kubernetes.io/projected/9347e7b4-1d53-4a28-9cbd-6a392b8145d8-kube-api-access-72dzl\") pod \"9347e7b4-1d53-4a28-9cbd-6a392b8145d8\" (UID: \"9347e7b4-1d53-4a28-9cbd-6a392b8145d8\") " Oct 02 02:04:04 crc kubenswrapper[4885]: I1002 02:04:04.612732 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9347e7b4-1d53-4a28-9cbd-6a392b8145d8-scripts\") pod \"9347e7b4-1d53-4a28-9cbd-6a392b8145d8\" (UID: \"9347e7b4-1d53-4a28-9cbd-6a392b8145d8\") " Oct 02 02:04:04 crc kubenswrapper[4885]: I1002 02:04:04.612777 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9347e7b4-1d53-4a28-9cbd-6a392b8145d8-credential-keys\") pod \"9347e7b4-1d53-4a28-9cbd-6a392b8145d8\" (UID: \"9347e7b4-1d53-4a28-9cbd-6a392b8145d8\") " Oct 02 02:04:04 crc kubenswrapper[4885]: I1002 02:04:04.623417 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9347e7b4-1d53-4a28-9cbd-6a392b8145d8-scripts" (OuterVolumeSpecName: "scripts") pod "9347e7b4-1d53-4a28-9cbd-6a392b8145d8" (UID: "9347e7b4-1d53-4a28-9cbd-6a392b8145d8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:04:04 crc kubenswrapper[4885]: I1002 02:04:04.624428 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9347e7b4-1d53-4a28-9cbd-6a392b8145d8-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "9347e7b4-1d53-4a28-9cbd-6a392b8145d8" (UID: "9347e7b4-1d53-4a28-9cbd-6a392b8145d8"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:04:04 crc kubenswrapper[4885]: I1002 02:04:04.624438 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9347e7b4-1d53-4a28-9cbd-6a392b8145d8-kube-api-access-72dzl" (OuterVolumeSpecName: "kube-api-access-72dzl") pod "9347e7b4-1d53-4a28-9cbd-6a392b8145d8" (UID: "9347e7b4-1d53-4a28-9cbd-6a392b8145d8"). InnerVolumeSpecName "kube-api-access-72dzl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:04:04 crc kubenswrapper[4885]: I1002 02:04:04.628624 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9347e7b4-1d53-4a28-9cbd-6a392b8145d8-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "9347e7b4-1d53-4a28-9cbd-6a392b8145d8" (UID: "9347e7b4-1d53-4a28-9cbd-6a392b8145d8"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:04:04 crc kubenswrapper[4885]: I1002 02:04:04.642371 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9347e7b4-1d53-4a28-9cbd-6a392b8145d8-config-data" (OuterVolumeSpecName: "config-data") pod "9347e7b4-1d53-4a28-9cbd-6a392b8145d8" (UID: "9347e7b4-1d53-4a28-9cbd-6a392b8145d8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:04:04 crc kubenswrapper[4885]: I1002 02:04:04.662801 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9347e7b4-1d53-4a28-9cbd-6a392b8145d8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9347e7b4-1d53-4a28-9cbd-6a392b8145d8" (UID: "9347e7b4-1d53-4a28-9cbd-6a392b8145d8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:04:04 crc kubenswrapper[4885]: I1002 02:04:04.714788 4885 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9347e7b4-1d53-4a28-9cbd-6a392b8145d8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:04:04 crc kubenswrapper[4885]: I1002 02:04:04.714891 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-72dzl\" (UniqueName: \"kubernetes.io/projected/9347e7b4-1d53-4a28-9cbd-6a392b8145d8-kube-api-access-72dzl\") on node \"crc\" DevicePath \"\"" Oct 02 02:04:04 crc kubenswrapper[4885]: I1002 02:04:04.714905 4885 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9347e7b4-1d53-4a28-9cbd-6a392b8145d8-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 02:04:04 crc kubenswrapper[4885]: I1002 02:04:04.714916 4885 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9347e7b4-1d53-4a28-9cbd-6a392b8145d8-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 02 02:04:04 crc kubenswrapper[4885]: I1002 02:04:04.714924 4885 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9347e7b4-1d53-4a28-9cbd-6a392b8145d8-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 02:04:04 crc kubenswrapper[4885]: I1002 02:04:04.714932 4885 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9347e7b4-1d53-4a28-9cbd-6a392b8145d8-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 02 02:04:04 crc kubenswrapper[4885]: I1002 02:04:04.948822 4885 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-698758b865-qkx2j" podUID="52b80336-4c74-4236-9600-7ea4606e5d6f" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.114:5353: connect: connection refused" Oct 02 02:04:05 crc kubenswrapper[4885]: I1002 02:04:05.362898 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-mj282" event={"ID":"9347e7b4-1d53-4a28-9cbd-6a392b8145d8","Type":"ContainerDied","Data":"e62fba6b80b17d6e49407c762ec104bb0499fdb60a8faee16778d517cc1fd5ed"} Oct 02 02:04:05 crc kubenswrapper[4885]: I1002 02:04:05.362995 4885 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e62fba6b80b17d6e49407c762ec104bb0499fdb60a8faee16778d517cc1fd5ed" Oct 02 02:04:05 crc kubenswrapper[4885]: I1002 02:04:05.363009 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-mj282" Oct 02 02:04:05 crc kubenswrapper[4885]: I1002 02:04:05.584614 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-mj282"] Oct 02 02:04:05 crc kubenswrapper[4885]: I1002 02:04:05.591573 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-mj282"] Oct 02 02:04:05 crc kubenswrapper[4885]: I1002 02:04:05.675242 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-dr95w"] Oct 02 02:04:05 crc kubenswrapper[4885]: E1002 02:04:05.675744 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9347e7b4-1d53-4a28-9cbd-6a392b8145d8" containerName="keystone-bootstrap" Oct 02 02:04:05 crc kubenswrapper[4885]: I1002 02:04:05.675774 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="9347e7b4-1d53-4a28-9cbd-6a392b8145d8" containerName="keystone-bootstrap" Oct 02 02:04:05 crc kubenswrapper[4885]: I1002 02:04:05.676102 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="9347e7b4-1d53-4a28-9cbd-6a392b8145d8" containerName="keystone-bootstrap" Oct 02 02:04:05 crc kubenswrapper[4885]: I1002 02:04:05.677882 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-dr95w" Oct 02 02:04:05 crc kubenswrapper[4885]: I1002 02:04:05.681013 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 02 02:04:05 crc kubenswrapper[4885]: I1002 02:04:05.681583 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 02 02:04:05 crc kubenswrapper[4885]: I1002 02:04:05.681847 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 02 02:04:05 crc kubenswrapper[4885]: I1002 02:04:05.682025 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-h4g4g" Oct 02 02:04:05 crc kubenswrapper[4885]: I1002 02:04:05.693839 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-dr95w"] Oct 02 02:04:05 crc kubenswrapper[4885]: I1002 02:04:05.734793 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d4a98a2-1f96-47a2-bacc-a2dd42bc4119-scripts\") pod \"keystone-bootstrap-dr95w\" (UID: \"1d4a98a2-1f96-47a2-bacc-a2dd42bc4119\") " pod="openstack/keystone-bootstrap-dr95w" Oct 02 02:04:05 crc kubenswrapper[4885]: I1002 02:04:05.734892 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1d4a98a2-1f96-47a2-bacc-a2dd42bc4119-credential-keys\") pod \"keystone-bootstrap-dr95w\" (UID: \"1d4a98a2-1f96-47a2-bacc-a2dd42bc4119\") " pod="openstack/keystone-bootstrap-dr95w" Oct 02 02:04:05 crc kubenswrapper[4885]: I1002 02:04:05.734933 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d4a98a2-1f96-47a2-bacc-a2dd42bc4119-combined-ca-bundle\") pod \"keystone-bootstrap-dr95w\" (UID: \"1d4a98a2-1f96-47a2-bacc-a2dd42bc4119\") " pod="openstack/keystone-bootstrap-dr95w" Oct 02 02:04:05 crc kubenswrapper[4885]: I1002 02:04:05.735098 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jzrlb\" (UniqueName: \"kubernetes.io/projected/1d4a98a2-1f96-47a2-bacc-a2dd42bc4119-kube-api-access-jzrlb\") pod \"keystone-bootstrap-dr95w\" (UID: \"1d4a98a2-1f96-47a2-bacc-a2dd42bc4119\") " pod="openstack/keystone-bootstrap-dr95w" Oct 02 02:04:05 crc kubenswrapper[4885]: I1002 02:04:05.735172 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d4a98a2-1f96-47a2-bacc-a2dd42bc4119-config-data\") pod \"keystone-bootstrap-dr95w\" (UID: \"1d4a98a2-1f96-47a2-bacc-a2dd42bc4119\") " pod="openstack/keystone-bootstrap-dr95w" Oct 02 02:04:05 crc kubenswrapper[4885]: I1002 02:04:05.735231 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1d4a98a2-1f96-47a2-bacc-a2dd42bc4119-fernet-keys\") pod \"keystone-bootstrap-dr95w\" (UID: \"1d4a98a2-1f96-47a2-bacc-a2dd42bc4119\") " pod="openstack/keystone-bootstrap-dr95w" Oct 02 02:04:05 crc kubenswrapper[4885]: I1002 02:04:05.836555 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jzrlb\" (UniqueName: \"kubernetes.io/projected/1d4a98a2-1f96-47a2-bacc-a2dd42bc4119-kube-api-access-jzrlb\") pod \"keystone-bootstrap-dr95w\" (UID: \"1d4a98a2-1f96-47a2-bacc-a2dd42bc4119\") " pod="openstack/keystone-bootstrap-dr95w" Oct 02 02:04:05 crc kubenswrapper[4885]: I1002 02:04:05.836631 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d4a98a2-1f96-47a2-bacc-a2dd42bc4119-config-data\") pod \"keystone-bootstrap-dr95w\" (UID: \"1d4a98a2-1f96-47a2-bacc-a2dd42bc4119\") " pod="openstack/keystone-bootstrap-dr95w" Oct 02 02:04:05 crc kubenswrapper[4885]: I1002 02:04:05.836672 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1d4a98a2-1f96-47a2-bacc-a2dd42bc4119-fernet-keys\") pod \"keystone-bootstrap-dr95w\" (UID: \"1d4a98a2-1f96-47a2-bacc-a2dd42bc4119\") " pod="openstack/keystone-bootstrap-dr95w" Oct 02 02:04:05 crc kubenswrapper[4885]: I1002 02:04:05.836711 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d4a98a2-1f96-47a2-bacc-a2dd42bc4119-scripts\") pod \"keystone-bootstrap-dr95w\" (UID: \"1d4a98a2-1f96-47a2-bacc-a2dd42bc4119\") " pod="openstack/keystone-bootstrap-dr95w" Oct 02 02:04:05 crc kubenswrapper[4885]: I1002 02:04:05.836738 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1d4a98a2-1f96-47a2-bacc-a2dd42bc4119-credential-keys\") pod \"keystone-bootstrap-dr95w\" (UID: \"1d4a98a2-1f96-47a2-bacc-a2dd42bc4119\") " pod="openstack/keystone-bootstrap-dr95w" Oct 02 02:04:05 crc kubenswrapper[4885]: I1002 02:04:05.836767 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d4a98a2-1f96-47a2-bacc-a2dd42bc4119-combined-ca-bundle\") pod \"keystone-bootstrap-dr95w\" (UID: \"1d4a98a2-1f96-47a2-bacc-a2dd42bc4119\") " pod="openstack/keystone-bootstrap-dr95w" Oct 02 02:04:05 crc kubenswrapper[4885]: I1002 02:04:05.841336 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d4a98a2-1f96-47a2-bacc-a2dd42bc4119-scripts\") pod \"keystone-bootstrap-dr95w\" (UID: \"1d4a98a2-1f96-47a2-bacc-a2dd42bc4119\") " pod="openstack/keystone-bootstrap-dr95w" Oct 02 02:04:05 crc kubenswrapper[4885]: I1002 02:04:05.843539 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d4a98a2-1f96-47a2-bacc-a2dd42bc4119-combined-ca-bundle\") pod \"keystone-bootstrap-dr95w\" (UID: \"1d4a98a2-1f96-47a2-bacc-a2dd42bc4119\") " pod="openstack/keystone-bootstrap-dr95w" Oct 02 02:04:05 crc kubenswrapper[4885]: I1002 02:04:05.843952 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1d4a98a2-1f96-47a2-bacc-a2dd42bc4119-credential-keys\") pod \"keystone-bootstrap-dr95w\" (UID: \"1d4a98a2-1f96-47a2-bacc-a2dd42bc4119\") " pod="openstack/keystone-bootstrap-dr95w" Oct 02 02:04:05 crc kubenswrapper[4885]: I1002 02:04:05.844061 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1d4a98a2-1f96-47a2-bacc-a2dd42bc4119-fernet-keys\") pod \"keystone-bootstrap-dr95w\" (UID: \"1d4a98a2-1f96-47a2-bacc-a2dd42bc4119\") " pod="openstack/keystone-bootstrap-dr95w" Oct 02 02:04:05 crc kubenswrapper[4885]: I1002 02:04:05.851522 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d4a98a2-1f96-47a2-bacc-a2dd42bc4119-config-data\") pod \"keystone-bootstrap-dr95w\" (UID: \"1d4a98a2-1f96-47a2-bacc-a2dd42bc4119\") " pod="openstack/keystone-bootstrap-dr95w" Oct 02 02:04:05 crc kubenswrapper[4885]: I1002 02:04:05.868043 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jzrlb\" (UniqueName: \"kubernetes.io/projected/1d4a98a2-1f96-47a2-bacc-a2dd42bc4119-kube-api-access-jzrlb\") pod \"keystone-bootstrap-dr95w\" (UID: \"1d4a98a2-1f96-47a2-bacc-a2dd42bc4119\") " pod="openstack/keystone-bootstrap-dr95w" Oct 02 02:04:06 crc kubenswrapper[4885]: I1002 02:04:06.000188 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-dr95w" Oct 02 02:04:06 crc kubenswrapper[4885]: I1002 02:04:06.063808 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9347e7b4-1d53-4a28-9cbd-6a392b8145d8" path="/var/lib/kubelet/pods/9347e7b4-1d53-4a28-9cbd-6a392b8145d8/volumes" Oct 02 02:04:07 crc kubenswrapper[4885]: E1002 02:04:07.428819 4885 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Oct 02 02:04:07 crc kubenswrapper[4885]: E1002 02:04:07.429367 4885 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fqdf4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-tjcz4_openstack(9c66bb6a-5a3c-4223-aba6-c5d9d0ea83ba): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 02:04:07 crc kubenswrapper[4885]: E1002 02:04:07.430667 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-tjcz4" podUID="9c66bb6a-5a3c-4223-aba6-c5d9d0ea83ba" Oct 02 02:04:07 crc kubenswrapper[4885]: E1002 02:04:07.927779 4885 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified" Oct 02 02:04:07 crc kubenswrapper[4885]: E1002 02:04:07.928207 4885 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tf2lz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-gp9st_openstack(0feba5bf-20c5-4578-9cfc-f37d6511c2d6): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 02:04:07 crc kubenswrapper[4885]: E1002 02:04:07.929449 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-gp9st" podUID="0feba5bf-20c5-4578-9cfc-f37d6511c2d6" Oct 02 02:04:07 crc kubenswrapper[4885]: I1002 02:04:07.984616 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-spxxt" Oct 02 02:04:08 crc kubenswrapper[4885]: I1002 02:04:08.101136 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk2sb\" (UniqueName: \"kubernetes.io/projected/6a9ed972-b8c7-4ff7-af7b-899842ecc19c-kube-api-access-tk2sb\") pod \"6a9ed972-b8c7-4ff7-af7b-899842ecc19c\" (UID: \"6a9ed972-b8c7-4ff7-af7b-899842ecc19c\") " Oct 02 02:04:08 crc kubenswrapper[4885]: I1002 02:04:08.101222 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6a9ed972-b8c7-4ff7-af7b-899842ecc19c-db-sync-config-data\") pod \"6a9ed972-b8c7-4ff7-af7b-899842ecc19c\" (UID: \"6a9ed972-b8c7-4ff7-af7b-899842ecc19c\") " Oct 02 02:04:08 crc kubenswrapper[4885]: I1002 02:04:08.101519 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a9ed972-b8c7-4ff7-af7b-899842ecc19c-config-data\") pod \"6a9ed972-b8c7-4ff7-af7b-899842ecc19c\" (UID: \"6a9ed972-b8c7-4ff7-af7b-899842ecc19c\") " Oct 02 02:04:08 crc kubenswrapper[4885]: I1002 02:04:08.101622 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a9ed972-b8c7-4ff7-af7b-899842ecc19c-combined-ca-bundle\") pod \"6a9ed972-b8c7-4ff7-af7b-899842ecc19c\" (UID: \"6a9ed972-b8c7-4ff7-af7b-899842ecc19c\") " Oct 02 02:04:08 crc kubenswrapper[4885]: I1002 02:04:08.110036 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a9ed972-b8c7-4ff7-af7b-899842ecc19c-kube-api-access-tk2sb" (OuterVolumeSpecName: "kube-api-access-tk2sb") pod "6a9ed972-b8c7-4ff7-af7b-899842ecc19c" (UID: "6a9ed972-b8c7-4ff7-af7b-899842ecc19c"). InnerVolumeSpecName "kube-api-access-tk2sb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:04:08 crc kubenswrapper[4885]: I1002 02:04:08.113317 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a9ed972-b8c7-4ff7-af7b-899842ecc19c-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "6a9ed972-b8c7-4ff7-af7b-899842ecc19c" (UID: "6a9ed972-b8c7-4ff7-af7b-899842ecc19c"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:04:08 crc kubenswrapper[4885]: I1002 02:04:08.131835 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a9ed972-b8c7-4ff7-af7b-899842ecc19c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6a9ed972-b8c7-4ff7-af7b-899842ecc19c" (UID: "6a9ed972-b8c7-4ff7-af7b-899842ecc19c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:04:08 crc kubenswrapper[4885]: I1002 02:04:08.158833 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a9ed972-b8c7-4ff7-af7b-899842ecc19c-config-data" (OuterVolumeSpecName: "config-data") pod "6a9ed972-b8c7-4ff7-af7b-899842ecc19c" (UID: "6a9ed972-b8c7-4ff7-af7b-899842ecc19c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:04:08 crc kubenswrapper[4885]: I1002 02:04:08.204166 4885 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a9ed972-b8c7-4ff7-af7b-899842ecc19c-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 02:04:08 crc kubenswrapper[4885]: I1002 02:04:08.204201 4885 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a9ed972-b8c7-4ff7-af7b-899842ecc19c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:04:08 crc kubenswrapper[4885]: I1002 02:04:08.204214 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk2sb\" (UniqueName: \"kubernetes.io/projected/6a9ed972-b8c7-4ff7-af7b-899842ecc19c-kube-api-access-tk2sb\") on node \"crc\" DevicePath \"\"" Oct 02 02:04:08 crc kubenswrapper[4885]: I1002 02:04:08.204500 4885 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6a9ed972-b8c7-4ff7-af7b-899842ecc19c-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 02:04:08 crc kubenswrapper[4885]: E1002 02:04:08.372737 4885 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified" Oct 02 02:04:08 crc kubenswrapper[4885]: E1002 02:04:08.373005 4885 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n57bh88h55hdch554h554h698h695hcbh5cchd7h654h66ch574hbdh4h5dch547h57h5b9hdbh64h658h5b8h87h58fh647h67bh577h76h5b7h64bq,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-z2mzr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(9226a04e-0195-42a1-b651-2f5379b1e2a9): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 02:04:08 crc kubenswrapper[4885]: I1002 02:04:08.411227 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-spxxt" event={"ID":"6a9ed972-b8c7-4ff7-af7b-899842ecc19c","Type":"ContainerDied","Data":"c4c16af334ec7909627f304bcf3404a3f330aae644750d65ad8deafc948d085b"} Oct 02 02:04:08 crc kubenswrapper[4885]: I1002 02:04:08.411558 4885 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c4c16af334ec7909627f304bcf3404a3f330aae644750d65ad8deafc948d085b" Oct 02 02:04:08 crc kubenswrapper[4885]: I1002 02:04:08.411317 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-spxxt" Oct 02 02:04:08 crc kubenswrapper[4885]: I1002 02:04:08.414995 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-qkx2j" event={"ID":"52b80336-4c74-4236-9600-7ea4606e5d6f","Type":"ContainerDied","Data":"e4d5fb452419cdbec9b0adfdcfef9e875939e09cde369b3ae1b2665922ca85d6"} Oct 02 02:04:08 crc kubenswrapper[4885]: I1002 02:04:08.415064 4885 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e4d5fb452419cdbec9b0adfdcfef9e875939e09cde369b3ae1b2665922ca85d6" Oct 02 02:04:08 crc kubenswrapper[4885]: I1002 02:04:08.451038 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-qkx2j" Oct 02 02:04:08 crc kubenswrapper[4885]: E1002 02:04:08.458983 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-tjcz4" podUID="9c66bb6a-5a3c-4223-aba6-c5d9d0ea83ba" Oct 02 02:04:08 crc kubenswrapper[4885]: E1002 02:04:08.459010 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified\\\"\"" pod="openstack/barbican-db-sync-gp9st" podUID="0feba5bf-20c5-4578-9cfc-f37d6511c2d6" Oct 02 02:04:08 crc kubenswrapper[4885]: I1002 02:04:08.508548 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/52b80336-4c74-4236-9600-7ea4606e5d6f-dns-svc\") pod \"52b80336-4c74-4236-9600-7ea4606e5d6f\" (UID: \"52b80336-4c74-4236-9600-7ea4606e5d6f\") " Oct 02 02:04:08 crc kubenswrapper[4885]: I1002 02:04:08.509487 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/52b80336-4c74-4236-9600-7ea4606e5d6f-ovsdbserver-sb\") pod \"52b80336-4c74-4236-9600-7ea4606e5d6f\" (UID: \"52b80336-4c74-4236-9600-7ea4606e5d6f\") " Oct 02 02:04:08 crc kubenswrapper[4885]: I1002 02:04:08.509698 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/52b80336-4c74-4236-9600-7ea4606e5d6f-ovsdbserver-nb\") pod \"52b80336-4c74-4236-9600-7ea4606e5d6f\" (UID: \"52b80336-4c74-4236-9600-7ea4606e5d6f\") " Oct 02 02:04:08 crc kubenswrapper[4885]: I1002 02:04:08.509842 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-br595\" (UniqueName: \"kubernetes.io/projected/52b80336-4c74-4236-9600-7ea4606e5d6f-kube-api-access-br595\") pod \"52b80336-4c74-4236-9600-7ea4606e5d6f\" (UID: \"52b80336-4c74-4236-9600-7ea4606e5d6f\") " Oct 02 02:04:08 crc kubenswrapper[4885]: I1002 02:04:08.511122 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/52b80336-4c74-4236-9600-7ea4606e5d6f-config\") pod \"52b80336-4c74-4236-9600-7ea4606e5d6f\" (UID: \"52b80336-4c74-4236-9600-7ea4606e5d6f\") " Oct 02 02:04:08 crc kubenswrapper[4885]: I1002 02:04:08.514545 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/52b80336-4c74-4236-9600-7ea4606e5d6f-kube-api-access-br595" (OuterVolumeSpecName: "kube-api-access-br595") pod "52b80336-4c74-4236-9600-7ea4606e5d6f" (UID: "52b80336-4c74-4236-9600-7ea4606e5d6f"). InnerVolumeSpecName "kube-api-access-br595". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:04:08 crc kubenswrapper[4885]: I1002 02:04:08.585778 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/52b80336-4c74-4236-9600-7ea4606e5d6f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "52b80336-4c74-4236-9600-7ea4606e5d6f" (UID: "52b80336-4c74-4236-9600-7ea4606e5d6f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:04:08 crc kubenswrapper[4885]: I1002 02:04:08.586491 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/52b80336-4c74-4236-9600-7ea4606e5d6f-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "52b80336-4c74-4236-9600-7ea4606e5d6f" (UID: "52b80336-4c74-4236-9600-7ea4606e5d6f"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:04:08 crc kubenswrapper[4885]: I1002 02:04:08.592202 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/52b80336-4c74-4236-9600-7ea4606e5d6f-config" (OuterVolumeSpecName: "config") pod "52b80336-4c74-4236-9600-7ea4606e5d6f" (UID: "52b80336-4c74-4236-9600-7ea4606e5d6f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:04:08 crc kubenswrapper[4885]: I1002 02:04:08.606676 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/52b80336-4c74-4236-9600-7ea4606e5d6f-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "52b80336-4c74-4236-9600-7ea4606e5d6f" (UID: "52b80336-4c74-4236-9600-7ea4606e5d6f"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:04:08 crc kubenswrapper[4885]: I1002 02:04:08.613905 4885 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/52b80336-4c74-4236-9600-7ea4606e5d6f-config\") on node \"crc\" DevicePath \"\"" Oct 02 02:04:08 crc kubenswrapper[4885]: I1002 02:04:08.613931 4885 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/52b80336-4c74-4236-9600-7ea4606e5d6f-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 02:04:08 crc kubenswrapper[4885]: I1002 02:04:08.613941 4885 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/52b80336-4c74-4236-9600-7ea4606e5d6f-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 02:04:08 crc kubenswrapper[4885]: I1002 02:04:08.613951 4885 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/52b80336-4c74-4236-9600-7ea4606e5d6f-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 02:04:08 crc kubenswrapper[4885]: I1002 02:04:08.613961 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-br595\" (UniqueName: \"kubernetes.io/projected/52b80336-4c74-4236-9600-7ea4606e5d6f-kube-api-access-br595\") on node \"crc\" DevicePath \"\"" Oct 02 02:04:08 crc kubenswrapper[4885]: I1002 02:04:08.906868 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6d7c47744-vjcbs"] Oct 02 02:04:08 crc kubenswrapper[4885]: W1002 02:04:08.915139 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod607de89d_b68e_49e8_beb4_7664a37c6105.slice/crio-eab1d34dba2aa4a3f6cc8d14dae192f6f94c0dbb1615c9dc41c5428a91642845 WatchSource:0}: Error finding container eab1d34dba2aa4a3f6cc8d14dae192f6f94c0dbb1615c9dc41c5428a91642845: Status 404 returned error can't find the container with id eab1d34dba2aa4a3f6cc8d14dae192f6f94c0dbb1615c9dc41c5428a91642845 Oct 02 02:04:08 crc kubenswrapper[4885]: I1002 02:04:08.994635 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-dr95w"] Oct 02 02:04:09 crc kubenswrapper[4885]: I1002 02:04:09.021996 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7c49b4cffb-brdz9"] Oct 02 02:04:09 crc kubenswrapper[4885]: I1002 02:04:09.341936 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-x6dgq"] Oct 02 02:04:09 crc kubenswrapper[4885]: E1002 02:04:09.343183 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52b80336-4c74-4236-9600-7ea4606e5d6f" containerName="dnsmasq-dns" Oct 02 02:04:09 crc kubenswrapper[4885]: I1002 02:04:09.343206 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="52b80336-4c74-4236-9600-7ea4606e5d6f" containerName="dnsmasq-dns" Oct 02 02:04:09 crc kubenswrapper[4885]: E1002 02:04:09.343233 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a9ed972-b8c7-4ff7-af7b-899842ecc19c" containerName="glance-db-sync" Oct 02 02:04:09 crc kubenswrapper[4885]: I1002 02:04:09.343240 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a9ed972-b8c7-4ff7-af7b-899842ecc19c" containerName="glance-db-sync" Oct 02 02:04:09 crc kubenswrapper[4885]: E1002 02:04:09.343268 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52b80336-4c74-4236-9600-7ea4606e5d6f" containerName="init" Oct 02 02:04:09 crc kubenswrapper[4885]: I1002 02:04:09.343274 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="52b80336-4c74-4236-9600-7ea4606e5d6f" containerName="init" Oct 02 02:04:09 crc kubenswrapper[4885]: I1002 02:04:09.343418 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="52b80336-4c74-4236-9600-7ea4606e5d6f" containerName="dnsmasq-dns" Oct 02 02:04:09 crc kubenswrapper[4885]: I1002 02:04:09.343437 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a9ed972-b8c7-4ff7-af7b-899842ecc19c" containerName="glance-db-sync" Oct 02 02:04:09 crc kubenswrapper[4885]: I1002 02:04:09.344337 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-x6dgq" Oct 02 02:04:09 crc kubenswrapper[4885]: I1002 02:04:09.356627 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-x6dgq"] Oct 02 02:04:09 crc kubenswrapper[4885]: I1002 02:04:09.439864 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-79659d7df7-6dkj8" event={"ID":"35248aca-4932-408d-a758-766cce4ae517","Type":"ContainerStarted","Data":"ef22dce64ab597d1b805bb0b4263ecaa74301822abb44a47719aa7375f39e18f"} Oct 02 02:04:09 crc kubenswrapper[4885]: I1002 02:04:09.439906 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-79659d7df7-6dkj8" event={"ID":"35248aca-4932-408d-a758-766cce4ae517","Type":"ContainerStarted","Data":"8d94317efe50870f3aa4725f3ee16894d0d9d196ed65560002f5e1578770a242"} Oct 02 02:04:09 crc kubenswrapper[4885]: I1002 02:04:09.440033 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-79659d7df7-6dkj8" podUID="35248aca-4932-408d-a758-766cce4ae517" containerName="horizon-log" containerID="cri-o://8d94317efe50870f3aa4725f3ee16894d0d9d196ed65560002f5e1578770a242" gracePeriod=30 Oct 02 02:04:09 crc kubenswrapper[4885]: I1002 02:04:09.440315 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-79659d7df7-6dkj8" podUID="35248aca-4932-408d-a758-766cce4ae517" containerName="horizon" containerID="cri-o://ef22dce64ab597d1b805bb0b4263ecaa74301822abb44a47719aa7375f39e18f" gracePeriod=30 Oct 02 02:04:09 crc kubenswrapper[4885]: I1002 02:04:09.449235 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d055d017-5640-4a3d-b684-6ce2755eab74-ovsdbserver-sb\") pod \"dnsmasq-dns-8b5c85b87-x6dgq\" (UID: \"d055d017-5640-4a3d-b684-6ce2755eab74\") " pod="openstack/dnsmasq-dns-8b5c85b87-x6dgq" Oct 02 02:04:09 crc kubenswrapper[4885]: I1002 02:04:09.449320 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d055d017-5640-4a3d-b684-6ce2755eab74-ovsdbserver-nb\") pod \"dnsmasq-dns-8b5c85b87-x6dgq\" (UID: \"d055d017-5640-4a3d-b684-6ce2755eab74\") " pod="openstack/dnsmasq-dns-8b5c85b87-x6dgq" Oct 02 02:04:09 crc kubenswrapper[4885]: I1002 02:04:09.449360 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d055d017-5640-4a3d-b684-6ce2755eab74-dns-swift-storage-0\") pod \"dnsmasq-dns-8b5c85b87-x6dgq\" (UID: \"d055d017-5640-4a3d-b684-6ce2755eab74\") " pod="openstack/dnsmasq-dns-8b5c85b87-x6dgq" Oct 02 02:04:09 crc kubenswrapper[4885]: I1002 02:04:09.449383 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d055d017-5640-4a3d-b684-6ce2755eab74-config\") pod \"dnsmasq-dns-8b5c85b87-x6dgq\" (UID: \"d055d017-5640-4a3d-b684-6ce2755eab74\") " pod="openstack/dnsmasq-dns-8b5c85b87-x6dgq" Oct 02 02:04:09 crc kubenswrapper[4885]: I1002 02:04:09.449401 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d055d017-5640-4a3d-b684-6ce2755eab74-dns-svc\") pod \"dnsmasq-dns-8b5c85b87-x6dgq\" (UID: \"d055d017-5640-4a3d-b684-6ce2755eab74\") " pod="openstack/dnsmasq-dns-8b5c85b87-x6dgq" Oct 02 02:04:09 crc kubenswrapper[4885]: I1002 02:04:09.449472 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jz9t7\" (UniqueName: \"kubernetes.io/projected/d055d017-5640-4a3d-b684-6ce2755eab74-kube-api-access-jz9t7\") pod \"dnsmasq-dns-8b5c85b87-x6dgq\" (UID: \"d055d017-5640-4a3d-b684-6ce2755eab74\") " pod="openstack/dnsmasq-dns-8b5c85b87-x6dgq" Oct 02 02:04:09 crc kubenswrapper[4885]: I1002 02:04:09.455847 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7c49b4cffb-brdz9" event={"ID":"b0fddcb3-3649-48f4-aa8f-85dd1c8137cf","Type":"ContainerStarted","Data":"9a7c1423da72336ecfa2cfe108bb2a287a4349ab5629c7bfb13486150d01b56a"} Oct 02 02:04:09 crc kubenswrapper[4885]: I1002 02:04:09.455888 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7c49b4cffb-brdz9" event={"ID":"b0fddcb3-3649-48f4-aa8f-85dd1c8137cf","Type":"ContainerStarted","Data":"2f5362e018a9e3621dd43a09aeb9e15a34e904554a2392e255e4061c6a46d062"} Oct 02 02:04:09 crc kubenswrapper[4885]: I1002 02:04:09.474721 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5686549c7f-j9kgc" event={"ID":"5056f944-c3cf-45e1-99f5-f93132717b93","Type":"ContainerStarted","Data":"41549198f0212f18def5c424af45e5fc7686634cca8e0d8a4bff946650eb6e5e"} Oct 02 02:04:09 crc kubenswrapper[4885]: I1002 02:04:09.474761 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5686549c7f-j9kgc" event={"ID":"5056f944-c3cf-45e1-99f5-f93132717b93","Type":"ContainerStarted","Data":"62c198b211a751154fad0eba8680b9e2bb3fd8d45f10cc75b898a36e966de946"} Oct 02 02:04:09 crc kubenswrapper[4885]: I1002 02:04:09.474876 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-5686549c7f-j9kgc" podUID="5056f944-c3cf-45e1-99f5-f93132717b93" containerName="horizon-log" containerID="cri-o://62c198b211a751154fad0eba8680b9e2bb3fd8d45f10cc75b898a36e966de946" gracePeriod=30 Oct 02 02:04:09 crc kubenswrapper[4885]: I1002 02:04:09.475748 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-5686549c7f-j9kgc" podUID="5056f944-c3cf-45e1-99f5-f93132717b93" containerName="horizon" containerID="cri-o://41549198f0212f18def5c424af45e5fc7686634cca8e0d8a4bff946650eb6e5e" gracePeriod=30 Oct 02 02:04:09 crc kubenswrapper[4885]: I1002 02:04:09.479106 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-79659d7df7-6dkj8" podStartSLOduration=2.498404124 podStartE2EDuration="31.479087834s" podCreationTimestamp="2025-10-02 02:03:38 +0000 UTC" firstStartedPulling="2025-10-02 02:03:39.524363703 +0000 UTC m=+1008.336111102" lastFinishedPulling="2025-10-02 02:04:08.505047403 +0000 UTC m=+1037.316794812" observedRunningTime="2025-10-02 02:04:09.467404119 +0000 UTC m=+1038.279151518" watchObservedRunningTime="2025-10-02 02:04:09.479087834 +0000 UTC m=+1038.290835233" Oct 02 02:04:09 crc kubenswrapper[4885]: I1002 02:04:09.512811 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6d7c47744-vjcbs" event={"ID":"607de89d-b68e-49e8-beb4-7664a37c6105","Type":"ContainerStarted","Data":"cfcd0a6bf9eeda848ac25c17468dc6142192f1dffd31f6f6e9d6baff7f27a9a2"} Oct 02 02:04:09 crc kubenswrapper[4885]: I1002 02:04:09.512859 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6d7c47744-vjcbs" event={"ID":"607de89d-b68e-49e8-beb4-7664a37c6105","Type":"ContainerStarted","Data":"eab1d34dba2aa4a3f6cc8d14dae192f6f94c0dbb1615c9dc41c5428a91642845"} Oct 02 02:04:09 crc kubenswrapper[4885]: I1002 02:04:09.522149 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-5686549c7f-j9kgc" podStartSLOduration=2.132947678 podStartE2EDuration="27.522131548s" podCreationTimestamp="2025-10-02 02:03:42 +0000 UTC" firstStartedPulling="2025-10-02 02:03:43.114672797 +0000 UTC m=+1011.926420216" lastFinishedPulling="2025-10-02 02:04:08.503856677 +0000 UTC m=+1037.315604086" observedRunningTime="2025-10-02 02:04:09.513093844 +0000 UTC m=+1038.324841243" watchObservedRunningTime="2025-10-02 02:04:09.522131548 +0000 UTC m=+1038.333878947" Oct 02 02:04:09 crc kubenswrapper[4885]: I1002 02:04:09.526742 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-8558cb5fd9-8cr6r" event={"ID":"8ad39937-75ce-468c-90df-b3abbaefe498","Type":"ContainerStarted","Data":"1ebc0d49cd4350c0748387677aa9ada287742667a4fa2985825e62867011aa82"} Oct 02 02:04:09 crc kubenswrapper[4885]: I1002 02:04:09.526782 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-8558cb5fd9-8cr6r" event={"ID":"8ad39937-75ce-468c-90df-b3abbaefe498","Type":"ContainerStarted","Data":"39adfd0e77c1141d8aa33b2b5ea71968edba0ddc693a26b57ab6d5ff366e56a5"} Oct 02 02:04:09 crc kubenswrapper[4885]: I1002 02:04:09.526832 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-8558cb5fd9-8cr6r" podUID="8ad39937-75ce-468c-90df-b3abbaefe498" containerName="horizon-log" containerID="cri-o://39adfd0e77c1141d8aa33b2b5ea71968edba0ddc693a26b57ab6d5ff366e56a5" gracePeriod=30 Oct 02 02:04:09 crc kubenswrapper[4885]: I1002 02:04:09.526926 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-8558cb5fd9-8cr6r" podUID="8ad39937-75ce-468c-90df-b3abbaefe498" containerName="horizon" containerID="cri-o://1ebc0d49cd4350c0748387677aa9ada287742667a4fa2985825e62867011aa82" gracePeriod=30 Oct 02 02:04:09 crc kubenswrapper[4885]: I1002 02:04:09.534910 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-qkx2j" Oct 02 02:04:09 crc kubenswrapper[4885]: I1002 02:04:09.535577 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-dr95w" event={"ID":"1d4a98a2-1f96-47a2-bacc-a2dd42bc4119","Type":"ContainerStarted","Data":"c6c205ead834e44b7376e00c80be9ca0db099be3ec3ef0e4d6aa2a17921241ce"} Oct 02 02:04:09 crc kubenswrapper[4885]: I1002 02:04:09.535624 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-dr95w" event={"ID":"1d4a98a2-1f96-47a2-bacc-a2dd42bc4119","Type":"ContainerStarted","Data":"15942f3d3388cc1972ff0c9b83d7728fba10a2f5975a6c727868882c69b37f18"} Oct 02 02:04:09 crc kubenswrapper[4885]: I1002 02:04:09.551962 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d055d017-5640-4a3d-b684-6ce2755eab74-config\") pod \"dnsmasq-dns-8b5c85b87-x6dgq\" (UID: \"d055d017-5640-4a3d-b684-6ce2755eab74\") " pod="openstack/dnsmasq-dns-8b5c85b87-x6dgq" Oct 02 02:04:09 crc kubenswrapper[4885]: I1002 02:04:09.552009 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d055d017-5640-4a3d-b684-6ce2755eab74-dns-svc\") pod \"dnsmasq-dns-8b5c85b87-x6dgq\" (UID: \"d055d017-5640-4a3d-b684-6ce2755eab74\") " pod="openstack/dnsmasq-dns-8b5c85b87-x6dgq" Oct 02 02:04:09 crc kubenswrapper[4885]: I1002 02:04:09.552093 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jz9t7\" (UniqueName: \"kubernetes.io/projected/d055d017-5640-4a3d-b684-6ce2755eab74-kube-api-access-jz9t7\") pod \"dnsmasq-dns-8b5c85b87-x6dgq\" (UID: \"d055d017-5640-4a3d-b684-6ce2755eab74\") " pod="openstack/dnsmasq-dns-8b5c85b87-x6dgq" Oct 02 02:04:09 crc kubenswrapper[4885]: I1002 02:04:09.552174 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d055d017-5640-4a3d-b684-6ce2755eab74-ovsdbserver-sb\") pod \"dnsmasq-dns-8b5c85b87-x6dgq\" (UID: \"d055d017-5640-4a3d-b684-6ce2755eab74\") " pod="openstack/dnsmasq-dns-8b5c85b87-x6dgq" Oct 02 02:04:09 crc kubenswrapper[4885]: I1002 02:04:09.552218 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d055d017-5640-4a3d-b684-6ce2755eab74-ovsdbserver-nb\") pod \"dnsmasq-dns-8b5c85b87-x6dgq\" (UID: \"d055d017-5640-4a3d-b684-6ce2755eab74\") " pod="openstack/dnsmasq-dns-8b5c85b87-x6dgq" Oct 02 02:04:09 crc kubenswrapper[4885]: I1002 02:04:09.552241 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d055d017-5640-4a3d-b684-6ce2755eab74-dns-swift-storage-0\") pod \"dnsmasq-dns-8b5c85b87-x6dgq\" (UID: \"d055d017-5640-4a3d-b684-6ce2755eab74\") " pod="openstack/dnsmasq-dns-8b5c85b87-x6dgq" Oct 02 02:04:09 crc kubenswrapper[4885]: I1002 02:04:09.552977 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d055d017-5640-4a3d-b684-6ce2755eab74-dns-swift-storage-0\") pod \"dnsmasq-dns-8b5c85b87-x6dgq\" (UID: \"d055d017-5640-4a3d-b684-6ce2755eab74\") " pod="openstack/dnsmasq-dns-8b5c85b87-x6dgq" Oct 02 02:04:09 crc kubenswrapper[4885]: I1002 02:04:09.553535 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d055d017-5640-4a3d-b684-6ce2755eab74-config\") pod \"dnsmasq-dns-8b5c85b87-x6dgq\" (UID: \"d055d017-5640-4a3d-b684-6ce2755eab74\") " pod="openstack/dnsmasq-dns-8b5c85b87-x6dgq" Oct 02 02:04:09 crc kubenswrapper[4885]: I1002 02:04:09.554205 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d055d017-5640-4a3d-b684-6ce2755eab74-ovsdbserver-sb\") pod \"dnsmasq-dns-8b5c85b87-x6dgq\" (UID: \"d055d017-5640-4a3d-b684-6ce2755eab74\") " pod="openstack/dnsmasq-dns-8b5c85b87-x6dgq" Oct 02 02:04:09 crc kubenswrapper[4885]: I1002 02:04:09.559880 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d055d017-5640-4a3d-b684-6ce2755eab74-dns-svc\") pod \"dnsmasq-dns-8b5c85b87-x6dgq\" (UID: \"d055d017-5640-4a3d-b684-6ce2755eab74\") " pod="openstack/dnsmasq-dns-8b5c85b87-x6dgq" Oct 02 02:04:09 crc kubenswrapper[4885]: I1002 02:04:09.566604 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d055d017-5640-4a3d-b684-6ce2755eab74-ovsdbserver-nb\") pod \"dnsmasq-dns-8b5c85b87-x6dgq\" (UID: \"d055d017-5640-4a3d-b684-6ce2755eab74\") " pod="openstack/dnsmasq-dns-8b5c85b87-x6dgq" Oct 02 02:04:09 crc kubenswrapper[4885]: I1002 02:04:09.571746 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-8558cb5fd9-8cr6r" podStartSLOduration=3.1672579770000002 podStartE2EDuration="31.57173008s" podCreationTimestamp="2025-10-02 02:03:38 +0000 UTC" firstStartedPulling="2025-10-02 02:03:39.974993065 +0000 UTC m=+1008.786740474" lastFinishedPulling="2025-10-02 02:04:08.379465178 +0000 UTC m=+1037.191212577" observedRunningTime="2025-10-02 02:04:09.546805914 +0000 UTC m=+1038.358553313" watchObservedRunningTime="2025-10-02 02:04:09.57173008 +0000 UTC m=+1038.383477479" Oct 02 02:04:09 crc kubenswrapper[4885]: I1002 02:04:09.573997 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jz9t7\" (UniqueName: \"kubernetes.io/projected/d055d017-5640-4a3d-b684-6ce2755eab74-kube-api-access-jz9t7\") pod \"dnsmasq-dns-8b5c85b87-x6dgq\" (UID: \"d055d017-5640-4a3d-b684-6ce2755eab74\") " pod="openstack/dnsmasq-dns-8b5c85b87-x6dgq" Oct 02 02:04:09 crc kubenswrapper[4885]: I1002 02:04:09.593244 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-dr95w" podStartSLOduration=4.593226921 podStartE2EDuration="4.593226921s" podCreationTimestamp="2025-10-02 02:04:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:04:09.570287976 +0000 UTC m=+1038.382035375" watchObservedRunningTime="2025-10-02 02:04:09.593226921 +0000 UTC m=+1038.404974320" Oct 02 02:04:09 crc kubenswrapper[4885]: I1002 02:04:09.609409 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-qkx2j"] Oct 02 02:04:09 crc kubenswrapper[4885]: I1002 02:04:09.627414 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-698758b865-qkx2j"] Oct 02 02:04:09 crc kubenswrapper[4885]: I1002 02:04:09.670579 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-x6dgq" Oct 02 02:04:10 crc kubenswrapper[4885]: I1002 02:04:10.055641 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="52b80336-4c74-4236-9600-7ea4606e5d6f" path="/var/lib/kubelet/pods/52b80336-4c74-4236-9600-7ea4606e5d6f/volumes" Oct 02 02:04:10 crc kubenswrapper[4885]: I1002 02:04:10.176175 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-x6dgq"] Oct 02 02:04:10 crc kubenswrapper[4885]: I1002 02:04:10.282812 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 02:04:10 crc kubenswrapper[4885]: I1002 02:04:10.285130 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 02 02:04:10 crc kubenswrapper[4885]: I1002 02:04:10.289639 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 02:04:10 crc kubenswrapper[4885]: I1002 02:04:10.290122 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Oct 02 02:04:10 crc kubenswrapper[4885]: I1002 02:04:10.290389 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 02 02:04:10 crc kubenswrapper[4885]: I1002 02:04:10.290577 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-fkkck" Oct 02 02:04:10 crc kubenswrapper[4885]: I1002 02:04:10.369937 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d091410b-e1bf-423e-929f-18d2a9c65891-logs\") pod \"glance-default-external-api-0\" (UID: \"d091410b-e1bf-423e-929f-18d2a9c65891\") " pod="openstack/glance-default-external-api-0" Oct 02 02:04:10 crc kubenswrapper[4885]: I1002 02:04:10.369988 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d091410b-e1bf-423e-929f-18d2a9c65891-config-data\") pod \"glance-default-external-api-0\" (UID: \"d091410b-e1bf-423e-929f-18d2a9c65891\") " pod="openstack/glance-default-external-api-0" Oct 02 02:04:10 crc kubenswrapper[4885]: I1002 02:04:10.370012 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d091410b-e1bf-423e-929f-18d2a9c65891-scripts\") pod \"glance-default-external-api-0\" (UID: \"d091410b-e1bf-423e-929f-18d2a9c65891\") " pod="openstack/glance-default-external-api-0" Oct 02 02:04:10 crc kubenswrapper[4885]: I1002 02:04:10.370089 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s72zw\" (UniqueName: \"kubernetes.io/projected/d091410b-e1bf-423e-929f-18d2a9c65891-kube-api-access-s72zw\") pod \"glance-default-external-api-0\" (UID: \"d091410b-e1bf-423e-929f-18d2a9c65891\") " pod="openstack/glance-default-external-api-0" Oct 02 02:04:10 crc kubenswrapper[4885]: I1002 02:04:10.370130 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d091410b-e1bf-423e-929f-18d2a9c65891-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"d091410b-e1bf-423e-929f-18d2a9c65891\") " pod="openstack/glance-default-external-api-0" Oct 02 02:04:10 crc kubenswrapper[4885]: I1002 02:04:10.370178 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d091410b-e1bf-423e-929f-18d2a9c65891-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d091410b-e1bf-423e-929f-18d2a9c65891\") " pod="openstack/glance-default-external-api-0" Oct 02 02:04:10 crc kubenswrapper[4885]: I1002 02:04:10.370204 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"d091410b-e1bf-423e-929f-18d2a9c65891\") " pod="openstack/glance-default-external-api-0" Oct 02 02:04:10 crc kubenswrapper[4885]: I1002 02:04:10.467345 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 02:04:10 crc kubenswrapper[4885]: I1002 02:04:10.469368 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 02 02:04:10 crc kubenswrapper[4885]: I1002 02:04:10.473616 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 02 02:04:10 crc kubenswrapper[4885]: I1002 02:04:10.474847 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d091410b-e1bf-423e-929f-18d2a9c65891-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"d091410b-e1bf-423e-929f-18d2a9c65891\") " pod="openstack/glance-default-external-api-0" Oct 02 02:04:10 crc kubenswrapper[4885]: I1002 02:04:10.474946 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d091410b-e1bf-423e-929f-18d2a9c65891-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d091410b-e1bf-423e-929f-18d2a9c65891\") " pod="openstack/glance-default-external-api-0" Oct 02 02:04:10 crc kubenswrapper[4885]: I1002 02:04:10.474989 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"d091410b-e1bf-423e-929f-18d2a9c65891\") " pod="openstack/glance-default-external-api-0" Oct 02 02:04:10 crc kubenswrapper[4885]: I1002 02:04:10.475079 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d091410b-e1bf-423e-929f-18d2a9c65891-logs\") pod \"glance-default-external-api-0\" (UID: \"d091410b-e1bf-423e-929f-18d2a9c65891\") " pod="openstack/glance-default-external-api-0" Oct 02 02:04:10 crc kubenswrapper[4885]: I1002 02:04:10.475127 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d091410b-e1bf-423e-929f-18d2a9c65891-config-data\") pod \"glance-default-external-api-0\" (UID: \"d091410b-e1bf-423e-929f-18d2a9c65891\") " pod="openstack/glance-default-external-api-0" Oct 02 02:04:10 crc kubenswrapper[4885]: I1002 02:04:10.475159 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d091410b-e1bf-423e-929f-18d2a9c65891-scripts\") pod \"glance-default-external-api-0\" (UID: \"d091410b-e1bf-423e-929f-18d2a9c65891\") " pod="openstack/glance-default-external-api-0" Oct 02 02:04:10 crc kubenswrapper[4885]: I1002 02:04:10.475219 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s72zw\" (UniqueName: \"kubernetes.io/projected/d091410b-e1bf-423e-929f-18d2a9c65891-kube-api-access-s72zw\") pod \"glance-default-external-api-0\" (UID: \"d091410b-e1bf-423e-929f-18d2a9c65891\") " pod="openstack/glance-default-external-api-0" Oct 02 02:04:10 crc kubenswrapper[4885]: I1002 02:04:10.476749 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d091410b-e1bf-423e-929f-18d2a9c65891-logs\") pod \"glance-default-external-api-0\" (UID: \"d091410b-e1bf-423e-929f-18d2a9c65891\") " pod="openstack/glance-default-external-api-0" Oct 02 02:04:10 crc kubenswrapper[4885]: I1002 02:04:10.477039 4885 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"d091410b-e1bf-423e-929f-18d2a9c65891\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/glance-default-external-api-0" Oct 02 02:04:10 crc kubenswrapper[4885]: I1002 02:04:10.477701 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d091410b-e1bf-423e-929f-18d2a9c65891-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d091410b-e1bf-423e-929f-18d2a9c65891\") " pod="openstack/glance-default-external-api-0" Oct 02 02:04:10 crc kubenswrapper[4885]: I1002 02:04:10.483957 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d091410b-e1bf-423e-929f-18d2a9c65891-scripts\") pod \"glance-default-external-api-0\" (UID: \"d091410b-e1bf-423e-929f-18d2a9c65891\") " pod="openstack/glance-default-external-api-0" Oct 02 02:04:10 crc kubenswrapper[4885]: I1002 02:04:10.504659 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s72zw\" (UniqueName: \"kubernetes.io/projected/d091410b-e1bf-423e-929f-18d2a9c65891-kube-api-access-s72zw\") pod \"glance-default-external-api-0\" (UID: \"d091410b-e1bf-423e-929f-18d2a9c65891\") " pod="openstack/glance-default-external-api-0" Oct 02 02:04:10 crc kubenswrapper[4885]: I1002 02:04:10.509311 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d091410b-e1bf-423e-929f-18d2a9c65891-config-data\") pod \"glance-default-external-api-0\" (UID: \"d091410b-e1bf-423e-929f-18d2a9c65891\") " pod="openstack/glance-default-external-api-0" Oct 02 02:04:10 crc kubenswrapper[4885]: I1002 02:04:10.511270 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d091410b-e1bf-423e-929f-18d2a9c65891-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"d091410b-e1bf-423e-929f-18d2a9c65891\") " pod="openstack/glance-default-external-api-0" Oct 02 02:04:10 crc kubenswrapper[4885]: I1002 02:04:10.513463 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"d091410b-e1bf-423e-929f-18d2a9c65891\") " pod="openstack/glance-default-external-api-0" Oct 02 02:04:10 crc kubenswrapper[4885]: I1002 02:04:10.529654 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 02:04:10 crc kubenswrapper[4885]: W1002 02:04:10.563011 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd055d017_5640_4a3d_b684_6ce2755eab74.slice/crio-7863f6254cf6fea29e7ccfd47a19709da3e70fdd1b07e3f178f579f164208669 WatchSource:0}: Error finding container 7863f6254cf6fea29e7ccfd47a19709da3e70fdd1b07e3f178f579f164208669: Status 404 returned error can't find the container with id 7863f6254cf6fea29e7ccfd47a19709da3e70fdd1b07e3f178f579f164208669 Oct 02 02:04:10 crc kubenswrapper[4885]: I1002 02:04:10.564169 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7c49b4cffb-brdz9" event={"ID":"b0fddcb3-3649-48f4-aa8f-85dd1c8137cf","Type":"ContainerStarted","Data":"87f582c0838b65442115c622d40cc23454484a73c597b4939299160d6e09873a"} Oct 02 02:04:10 crc kubenswrapper[4885]: I1002 02:04:10.577307 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6458faf-0340-43da-b4cf-4fa3b6384909-config-data\") pod \"glance-default-internal-api-0\" (UID: \"b6458faf-0340-43da-b4cf-4fa3b6384909\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:04:10 crc kubenswrapper[4885]: I1002 02:04:10.577428 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b6458faf-0340-43da-b4cf-4fa3b6384909-logs\") pod \"glance-default-internal-api-0\" (UID: \"b6458faf-0340-43da-b4cf-4fa3b6384909\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:04:10 crc kubenswrapper[4885]: I1002 02:04:10.577492 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b6458faf-0340-43da-b4cf-4fa3b6384909-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"b6458faf-0340-43da-b4cf-4fa3b6384909\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:04:10 crc kubenswrapper[4885]: I1002 02:04:10.577640 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b6458faf-0340-43da-b4cf-4fa3b6384909-scripts\") pod \"glance-default-internal-api-0\" (UID: \"b6458faf-0340-43da-b4cf-4fa3b6384909\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:04:10 crc kubenswrapper[4885]: I1002 02:04:10.577776 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"b6458faf-0340-43da-b4cf-4fa3b6384909\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:04:10 crc kubenswrapper[4885]: I1002 02:04:10.577810 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6458faf-0340-43da-b4cf-4fa3b6384909-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"b6458faf-0340-43da-b4cf-4fa3b6384909\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:04:10 crc kubenswrapper[4885]: I1002 02:04:10.577841 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qxwt6\" (UniqueName: \"kubernetes.io/projected/b6458faf-0340-43da-b4cf-4fa3b6384909-kube-api-access-qxwt6\") pod \"glance-default-internal-api-0\" (UID: \"b6458faf-0340-43da-b4cf-4fa3b6384909\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:04:10 crc kubenswrapper[4885]: I1002 02:04:10.582311 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6d7c47744-vjcbs" event={"ID":"607de89d-b68e-49e8-beb4-7664a37c6105","Type":"ContainerStarted","Data":"5fc3a6a95be092491f167c630f4d2f93cffd70f619f6e39fcee3ffb12230ba33"} Oct 02 02:04:10 crc kubenswrapper[4885]: I1002 02:04:10.589309 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-7c49b4cffb-brdz9" podStartSLOduration=23.589293529 podStartE2EDuration="23.589293529s" podCreationTimestamp="2025-10-02 02:03:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:04:10.589115013 +0000 UTC m=+1039.400862402" watchObservedRunningTime="2025-10-02 02:04:10.589293529 +0000 UTC m=+1039.401040928" Oct 02 02:04:10 crc kubenswrapper[4885]: I1002 02:04:10.614495 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 02 02:04:10 crc kubenswrapper[4885]: I1002 02:04:10.699850 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"b6458faf-0340-43da-b4cf-4fa3b6384909\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:04:10 crc kubenswrapper[4885]: I1002 02:04:10.700164 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6458faf-0340-43da-b4cf-4fa3b6384909-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"b6458faf-0340-43da-b4cf-4fa3b6384909\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:04:10 crc kubenswrapper[4885]: I1002 02:04:10.700192 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qxwt6\" (UniqueName: \"kubernetes.io/projected/b6458faf-0340-43da-b4cf-4fa3b6384909-kube-api-access-qxwt6\") pod \"glance-default-internal-api-0\" (UID: \"b6458faf-0340-43da-b4cf-4fa3b6384909\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:04:10 crc kubenswrapper[4885]: I1002 02:04:10.700284 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6458faf-0340-43da-b4cf-4fa3b6384909-config-data\") pod \"glance-default-internal-api-0\" (UID: \"b6458faf-0340-43da-b4cf-4fa3b6384909\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:04:10 crc kubenswrapper[4885]: I1002 02:04:10.700249 4885 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"b6458faf-0340-43da-b4cf-4fa3b6384909\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/glance-default-internal-api-0" Oct 02 02:04:10 crc kubenswrapper[4885]: I1002 02:04:10.702422 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b6458faf-0340-43da-b4cf-4fa3b6384909-logs\") pod \"glance-default-internal-api-0\" (UID: \"b6458faf-0340-43da-b4cf-4fa3b6384909\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:04:10 crc kubenswrapper[4885]: I1002 02:04:10.702525 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b6458faf-0340-43da-b4cf-4fa3b6384909-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"b6458faf-0340-43da-b4cf-4fa3b6384909\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:04:10 crc kubenswrapper[4885]: I1002 02:04:10.702751 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b6458faf-0340-43da-b4cf-4fa3b6384909-scripts\") pod \"glance-default-internal-api-0\" (UID: \"b6458faf-0340-43da-b4cf-4fa3b6384909\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:04:10 crc kubenswrapper[4885]: I1002 02:04:10.702861 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b6458faf-0340-43da-b4cf-4fa3b6384909-logs\") pod \"glance-default-internal-api-0\" (UID: \"b6458faf-0340-43da-b4cf-4fa3b6384909\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:04:10 crc kubenswrapper[4885]: I1002 02:04:10.703657 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b6458faf-0340-43da-b4cf-4fa3b6384909-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"b6458faf-0340-43da-b4cf-4fa3b6384909\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:04:10 crc kubenswrapper[4885]: I1002 02:04:10.708939 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b6458faf-0340-43da-b4cf-4fa3b6384909-scripts\") pod \"glance-default-internal-api-0\" (UID: \"b6458faf-0340-43da-b4cf-4fa3b6384909\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:04:10 crc kubenswrapper[4885]: I1002 02:04:10.717422 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6458faf-0340-43da-b4cf-4fa3b6384909-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"b6458faf-0340-43da-b4cf-4fa3b6384909\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:04:10 crc kubenswrapper[4885]: I1002 02:04:10.721937 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qxwt6\" (UniqueName: \"kubernetes.io/projected/b6458faf-0340-43da-b4cf-4fa3b6384909-kube-api-access-qxwt6\") pod \"glance-default-internal-api-0\" (UID: \"b6458faf-0340-43da-b4cf-4fa3b6384909\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:04:10 crc kubenswrapper[4885]: I1002 02:04:10.742430 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6458faf-0340-43da-b4cf-4fa3b6384909-config-data\") pod \"glance-default-internal-api-0\" (UID: \"b6458faf-0340-43da-b4cf-4fa3b6384909\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:04:10 crc kubenswrapper[4885]: I1002 02:04:10.776146 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"b6458faf-0340-43da-b4cf-4fa3b6384909\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:04:10 crc kubenswrapper[4885]: I1002 02:04:10.969413 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 02 02:04:11 crc kubenswrapper[4885]: I1002 02:04:11.202424 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-6d7c47744-vjcbs" podStartSLOduration=24.202407014 podStartE2EDuration="24.202407014s" podCreationTimestamp="2025-10-02 02:03:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:04:10.614378639 +0000 UTC m=+1039.426126028" watchObservedRunningTime="2025-10-02 02:04:11.202407014 +0000 UTC m=+1040.014154413" Oct 02 02:04:11 crc kubenswrapper[4885]: I1002 02:04:11.203782 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 02:04:11 crc kubenswrapper[4885]: I1002 02:04:11.555049 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 02:04:11 crc kubenswrapper[4885]: W1002 02:04:11.562482 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb6458faf_0340_43da_b4cf_4fa3b6384909.slice/crio-3cb09135a306c0546056c6e5a8611cd3876b37e41499f2cd78db8dadd2310e00 WatchSource:0}: Error finding container 3cb09135a306c0546056c6e5a8611cd3876b37e41499f2cd78db8dadd2310e00: Status 404 returned error can't find the container with id 3cb09135a306c0546056c6e5a8611cd3876b37e41499f2cd78db8dadd2310e00 Oct 02 02:04:11 crc kubenswrapper[4885]: I1002 02:04:11.601066 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d091410b-e1bf-423e-929f-18d2a9c65891","Type":"ContainerStarted","Data":"585f5e96d87df6f4e4087afa3df8bf0be66da056f6a5252ea24694ddc51ebdc4"} Oct 02 02:04:11 crc kubenswrapper[4885]: I1002 02:04:11.602806 4885 generic.go:334] "Generic (PLEG): container finished" podID="d055d017-5640-4a3d-b684-6ce2755eab74" containerID="f10b5d12a6ffdc5055f34cf17444d9b03b28e4b17439d618d94a61c06b28736c" exitCode=0 Oct 02 02:04:11 crc kubenswrapper[4885]: I1002 02:04:11.602867 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-x6dgq" event={"ID":"d055d017-5640-4a3d-b684-6ce2755eab74","Type":"ContainerDied","Data":"f10b5d12a6ffdc5055f34cf17444d9b03b28e4b17439d618d94a61c06b28736c"} Oct 02 02:04:11 crc kubenswrapper[4885]: I1002 02:04:11.602888 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-x6dgq" event={"ID":"d055d017-5640-4a3d-b684-6ce2755eab74","Type":"ContainerStarted","Data":"7863f6254cf6fea29e7ccfd47a19709da3e70fdd1b07e3f178f579f164208669"} Oct 02 02:04:11 crc kubenswrapper[4885]: I1002 02:04:11.612868 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9226a04e-0195-42a1-b651-2f5379b1e2a9","Type":"ContainerStarted","Data":"bc5a69f8f4d26b9f7c36647a448c3943fedcadcf35e24b1c46b0c1cb51a86fef"} Oct 02 02:04:11 crc kubenswrapper[4885]: I1002 02:04:11.615050 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b6458faf-0340-43da-b4cf-4fa3b6384909","Type":"ContainerStarted","Data":"3cb09135a306c0546056c6e5a8611cd3876b37e41499f2cd78db8dadd2310e00"} Oct 02 02:04:12 crc kubenswrapper[4885]: I1002 02:04:12.235807 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 02:04:12 crc kubenswrapper[4885]: I1002 02:04:12.305780 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 02:04:12 crc kubenswrapper[4885]: I1002 02:04:12.513307 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-5686549c7f-j9kgc" Oct 02 02:04:12 crc kubenswrapper[4885]: I1002 02:04:12.622340 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b6458faf-0340-43da-b4cf-4fa3b6384909","Type":"ContainerStarted","Data":"b4cfd3f7c1947afe37521744cf9538f5d916bf4e5eebeae8ba4abf7aba4b817f"} Oct 02 02:04:12 crc kubenswrapper[4885]: I1002 02:04:12.626750 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d091410b-e1bf-423e-929f-18d2a9c65891","Type":"ContainerStarted","Data":"775dde44ca3ce75c6bc52c6fbbe3f6e7cdbf8f9c303eb816d831b82b8e22f86c"} Oct 02 02:04:12 crc kubenswrapper[4885]: I1002 02:04:12.630358 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-x6dgq" event={"ID":"d055d017-5640-4a3d-b684-6ce2755eab74","Type":"ContainerStarted","Data":"64f15f37e85e8ff6d8c80cc01cf3fed9b5f658a2dccbcf79ffaccdbed931c8d5"} Oct 02 02:04:12 crc kubenswrapper[4885]: I1002 02:04:12.631752 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8b5c85b87-x6dgq" Oct 02 02:04:14 crc kubenswrapper[4885]: I1002 02:04:14.076553 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8b5c85b87-x6dgq" podStartSLOduration=5.076528281 podStartE2EDuration="5.076528281s" podCreationTimestamp="2025-10-02 02:04:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:04:12.648741963 +0000 UTC m=+1041.460489362" watchObservedRunningTime="2025-10-02 02:04:14.076528281 +0000 UTC m=+1042.888275680" Oct 02 02:04:14 crc kubenswrapper[4885]: I1002 02:04:14.662225 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b6458faf-0340-43da-b4cf-4fa3b6384909","Type":"ContainerStarted","Data":"e7966a9c7055b70c3575805bff8da5acbb90942af542595e888c8c2db22587f4"} Oct 02 02:04:14 crc kubenswrapper[4885]: I1002 02:04:14.662751 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="b6458faf-0340-43da-b4cf-4fa3b6384909" containerName="glance-log" containerID="cri-o://b4cfd3f7c1947afe37521744cf9538f5d916bf4e5eebeae8ba4abf7aba4b817f" gracePeriod=30 Oct 02 02:04:14 crc kubenswrapper[4885]: I1002 02:04:14.663290 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="b6458faf-0340-43da-b4cf-4fa3b6384909" containerName="glance-httpd" containerID="cri-o://e7966a9c7055b70c3575805bff8da5acbb90942af542595e888c8c2db22587f4" gracePeriod=30 Oct 02 02:04:14 crc kubenswrapper[4885]: I1002 02:04:14.673788 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d091410b-e1bf-423e-929f-18d2a9c65891","Type":"ContainerStarted","Data":"70c7aba074612eaff2e55f0f7d4ebed65c6d37d359fa30ef18084ec9df97e1aa"} Oct 02 02:04:14 crc kubenswrapper[4885]: I1002 02:04:14.674016 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="d091410b-e1bf-423e-929f-18d2a9c65891" containerName="glance-httpd" containerID="cri-o://70c7aba074612eaff2e55f0f7d4ebed65c6d37d359fa30ef18084ec9df97e1aa" gracePeriod=30 Oct 02 02:04:14 crc kubenswrapper[4885]: I1002 02:04:14.674142 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="d091410b-e1bf-423e-929f-18d2a9c65891" containerName="glance-log" containerID="cri-o://775dde44ca3ce75c6bc52c6fbbe3f6e7cdbf8f9c303eb816d831b82b8e22f86c" gracePeriod=30 Oct 02 02:04:14 crc kubenswrapper[4885]: I1002 02:04:14.682376 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=5.682356435 podStartE2EDuration="5.682356435s" podCreationTimestamp="2025-10-02 02:04:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:04:14.67854124 +0000 UTC m=+1043.490288639" watchObservedRunningTime="2025-10-02 02:04:14.682356435 +0000 UTC m=+1043.494103834" Oct 02 02:04:14 crc kubenswrapper[4885]: I1002 02:04:14.696768 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=5.696749602 podStartE2EDuration="5.696749602s" podCreationTimestamp="2025-10-02 02:04:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:04:14.696181254 +0000 UTC m=+1043.507928653" watchObservedRunningTime="2025-10-02 02:04:14.696749602 +0000 UTC m=+1043.508497001" Oct 02 02:04:15 crc kubenswrapper[4885]: I1002 02:04:15.686920 4885 generic.go:334] "Generic (PLEG): container finished" podID="d091410b-e1bf-423e-929f-18d2a9c65891" containerID="70c7aba074612eaff2e55f0f7d4ebed65c6d37d359fa30ef18084ec9df97e1aa" exitCode=0 Oct 02 02:04:15 crc kubenswrapper[4885]: I1002 02:04:15.686945 4885 generic.go:334] "Generic (PLEG): container finished" podID="d091410b-e1bf-423e-929f-18d2a9c65891" containerID="775dde44ca3ce75c6bc52c6fbbe3f6e7cdbf8f9c303eb816d831b82b8e22f86c" exitCode=143 Oct 02 02:04:15 crc kubenswrapper[4885]: I1002 02:04:15.686977 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d091410b-e1bf-423e-929f-18d2a9c65891","Type":"ContainerDied","Data":"70c7aba074612eaff2e55f0f7d4ebed65c6d37d359fa30ef18084ec9df97e1aa"} Oct 02 02:04:15 crc kubenswrapper[4885]: I1002 02:04:15.687001 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d091410b-e1bf-423e-929f-18d2a9c65891","Type":"ContainerDied","Data":"775dde44ca3ce75c6bc52c6fbbe3f6e7cdbf8f9c303eb816d831b82b8e22f86c"} Oct 02 02:04:15 crc kubenswrapper[4885]: I1002 02:04:15.690576 4885 generic.go:334] "Generic (PLEG): container finished" podID="b6458faf-0340-43da-b4cf-4fa3b6384909" containerID="b4cfd3f7c1947afe37521744cf9538f5d916bf4e5eebeae8ba4abf7aba4b817f" exitCode=143 Oct 02 02:04:15 crc kubenswrapper[4885]: I1002 02:04:15.690604 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b6458faf-0340-43da-b4cf-4fa3b6384909","Type":"ContainerDied","Data":"b4cfd3f7c1947afe37521744cf9538f5d916bf4e5eebeae8ba4abf7aba4b817f"} Oct 02 02:04:16 crc kubenswrapper[4885]: I1002 02:04:16.698366 4885 generic.go:334] "Generic (PLEG): container finished" podID="b6458faf-0340-43da-b4cf-4fa3b6384909" containerID="e7966a9c7055b70c3575805bff8da5acbb90942af542595e888c8c2db22587f4" exitCode=0 Oct 02 02:04:16 crc kubenswrapper[4885]: I1002 02:04:16.698460 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b6458faf-0340-43da-b4cf-4fa3b6384909","Type":"ContainerDied","Data":"e7966a9c7055b70c3575805bff8da5acbb90942af542595e888c8c2db22587f4"} Oct 02 02:04:17 crc kubenswrapper[4885]: I1002 02:04:17.984574 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7c49b4cffb-brdz9" Oct 02 02:04:17 crc kubenswrapper[4885]: I1002 02:04:17.984874 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-7c49b4cffb-brdz9" Oct 02 02:04:18 crc kubenswrapper[4885]: I1002 02:04:18.083649 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-6d7c47744-vjcbs" Oct 02 02:04:18 crc kubenswrapper[4885]: I1002 02:04:18.083743 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-6d7c47744-vjcbs" Oct 02 02:04:18 crc kubenswrapper[4885]: I1002 02:04:18.746113 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-79659d7df7-6dkj8" Oct 02 02:04:19 crc kubenswrapper[4885]: I1002 02:04:19.246479 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-8558cb5fd9-8cr6r" Oct 02 02:04:19 crc kubenswrapper[4885]: I1002 02:04:19.672408 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-8b5c85b87-x6dgq" Oct 02 02:04:19 crc kubenswrapper[4885]: I1002 02:04:19.742724 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-76fcf4b695-wpgst"] Oct 02 02:04:19 crc kubenswrapper[4885]: I1002 02:04:19.742945 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-76fcf4b695-wpgst" podUID="5adbde10-fd09-4e06-bad3-7f5d7bdf5b41" containerName="dnsmasq-dns" containerID="cri-o://aff3d2a87f29168594522e9b4f20c0cbbd55852066fe579d0b5a475463828e49" gracePeriod=10 Oct 02 02:04:20 crc kubenswrapper[4885]: I1002 02:04:20.763047 4885 generic.go:334] "Generic (PLEG): container finished" podID="1d4a98a2-1f96-47a2-bacc-a2dd42bc4119" containerID="c6c205ead834e44b7376e00c80be9ca0db099be3ec3ef0e4d6aa2a17921241ce" exitCode=0 Oct 02 02:04:20 crc kubenswrapper[4885]: I1002 02:04:20.764740 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-dr95w" event={"ID":"1d4a98a2-1f96-47a2-bacc-a2dd42bc4119","Type":"ContainerDied","Data":"c6c205ead834e44b7376e00c80be9ca0db099be3ec3ef0e4d6aa2a17921241ce"} Oct 02 02:04:20 crc kubenswrapper[4885]: I1002 02:04:20.767670 4885 generic.go:334] "Generic (PLEG): container finished" podID="5adbde10-fd09-4e06-bad3-7f5d7bdf5b41" containerID="aff3d2a87f29168594522e9b4f20c0cbbd55852066fe579d0b5a475463828e49" exitCode=0 Oct 02 02:04:20 crc kubenswrapper[4885]: I1002 02:04:20.767697 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76fcf4b695-wpgst" event={"ID":"5adbde10-fd09-4e06-bad3-7f5d7bdf5b41","Type":"ContainerDied","Data":"aff3d2a87f29168594522e9b4f20c0cbbd55852066fe579d0b5a475463828e49"} Oct 02 02:04:21 crc kubenswrapper[4885]: I1002 02:04:21.438448 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76fcf4b695-wpgst" Oct 02 02:04:21 crc kubenswrapper[4885]: I1002 02:04:21.507637 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 02 02:04:21 crc kubenswrapper[4885]: I1002 02:04:21.603236 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b6458faf-0340-43da-b4cf-4fa3b6384909-logs\") pod \"b6458faf-0340-43da-b4cf-4fa3b6384909\" (UID: \"b6458faf-0340-43da-b4cf-4fa3b6384909\") " Oct 02 02:04:21 crc kubenswrapper[4885]: I1002 02:04:21.603309 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"b6458faf-0340-43da-b4cf-4fa3b6384909\" (UID: \"b6458faf-0340-43da-b4cf-4fa3b6384909\") " Oct 02 02:04:21 crc kubenswrapper[4885]: I1002 02:04:21.603334 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5adbde10-fd09-4e06-bad3-7f5d7bdf5b41-dns-svc\") pod \"5adbde10-fd09-4e06-bad3-7f5d7bdf5b41\" (UID: \"5adbde10-fd09-4e06-bad3-7f5d7bdf5b41\") " Oct 02 02:04:21 crc kubenswrapper[4885]: I1002 02:04:21.603359 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6458faf-0340-43da-b4cf-4fa3b6384909-combined-ca-bundle\") pod \"b6458faf-0340-43da-b4cf-4fa3b6384909\" (UID: \"b6458faf-0340-43da-b4cf-4fa3b6384909\") " Oct 02 02:04:21 crc kubenswrapper[4885]: I1002 02:04:21.603395 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qxwt6\" (UniqueName: \"kubernetes.io/projected/b6458faf-0340-43da-b4cf-4fa3b6384909-kube-api-access-qxwt6\") pod \"b6458faf-0340-43da-b4cf-4fa3b6384909\" (UID: \"b6458faf-0340-43da-b4cf-4fa3b6384909\") " Oct 02 02:04:21 crc kubenswrapper[4885]: I1002 02:04:21.603461 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5adbde10-fd09-4e06-bad3-7f5d7bdf5b41-dns-swift-storage-0\") pod \"5adbde10-fd09-4e06-bad3-7f5d7bdf5b41\" (UID: \"5adbde10-fd09-4e06-bad3-7f5d7bdf5b41\") " Oct 02 02:04:21 crc kubenswrapper[4885]: I1002 02:04:21.603622 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b6458faf-0340-43da-b4cf-4fa3b6384909-logs" (OuterVolumeSpecName: "logs") pod "b6458faf-0340-43da-b4cf-4fa3b6384909" (UID: "b6458faf-0340-43da-b4cf-4fa3b6384909"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:04:21 crc kubenswrapper[4885]: I1002 02:04:21.603973 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5adbde10-fd09-4e06-bad3-7f5d7bdf5b41-ovsdbserver-sb\") pod \"5adbde10-fd09-4e06-bad3-7f5d7bdf5b41\" (UID: \"5adbde10-fd09-4e06-bad3-7f5d7bdf5b41\") " Oct 02 02:04:21 crc kubenswrapper[4885]: I1002 02:04:21.604005 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5adbde10-fd09-4e06-bad3-7f5d7bdf5b41-ovsdbserver-nb\") pod \"5adbde10-fd09-4e06-bad3-7f5d7bdf5b41\" (UID: \"5adbde10-fd09-4e06-bad3-7f5d7bdf5b41\") " Oct 02 02:04:21 crc kubenswrapper[4885]: I1002 02:04:21.604029 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-58fgd\" (UniqueName: \"kubernetes.io/projected/5adbde10-fd09-4e06-bad3-7f5d7bdf5b41-kube-api-access-58fgd\") pod \"5adbde10-fd09-4e06-bad3-7f5d7bdf5b41\" (UID: \"5adbde10-fd09-4e06-bad3-7f5d7bdf5b41\") " Oct 02 02:04:21 crc kubenswrapper[4885]: I1002 02:04:21.604070 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b6458faf-0340-43da-b4cf-4fa3b6384909-scripts\") pod \"b6458faf-0340-43da-b4cf-4fa3b6384909\" (UID: \"b6458faf-0340-43da-b4cf-4fa3b6384909\") " Oct 02 02:04:21 crc kubenswrapper[4885]: I1002 02:04:21.604109 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b6458faf-0340-43da-b4cf-4fa3b6384909-httpd-run\") pod \"b6458faf-0340-43da-b4cf-4fa3b6384909\" (UID: \"b6458faf-0340-43da-b4cf-4fa3b6384909\") " Oct 02 02:04:21 crc kubenswrapper[4885]: I1002 02:04:21.604127 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6458faf-0340-43da-b4cf-4fa3b6384909-config-data\") pod \"b6458faf-0340-43da-b4cf-4fa3b6384909\" (UID: \"b6458faf-0340-43da-b4cf-4fa3b6384909\") " Oct 02 02:04:21 crc kubenswrapper[4885]: I1002 02:04:21.604158 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5adbde10-fd09-4e06-bad3-7f5d7bdf5b41-config\") pod \"5adbde10-fd09-4e06-bad3-7f5d7bdf5b41\" (UID: \"5adbde10-fd09-4e06-bad3-7f5d7bdf5b41\") " Oct 02 02:04:21 crc kubenswrapper[4885]: I1002 02:04:21.604508 4885 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b6458faf-0340-43da-b4cf-4fa3b6384909-logs\") on node \"crc\" DevicePath \"\"" Oct 02 02:04:21 crc kubenswrapper[4885]: I1002 02:04:21.606730 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b6458faf-0340-43da-b4cf-4fa3b6384909-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "b6458faf-0340-43da-b4cf-4fa3b6384909" (UID: "b6458faf-0340-43da-b4cf-4fa3b6384909"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:04:21 crc kubenswrapper[4885]: I1002 02:04:21.608463 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "glance") pod "b6458faf-0340-43da-b4cf-4fa3b6384909" (UID: "b6458faf-0340-43da-b4cf-4fa3b6384909"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 02 02:04:21 crc kubenswrapper[4885]: I1002 02:04:21.610401 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6458faf-0340-43da-b4cf-4fa3b6384909-scripts" (OuterVolumeSpecName: "scripts") pod "b6458faf-0340-43da-b4cf-4fa3b6384909" (UID: "b6458faf-0340-43da-b4cf-4fa3b6384909"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:04:21 crc kubenswrapper[4885]: I1002 02:04:21.610552 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6458faf-0340-43da-b4cf-4fa3b6384909-kube-api-access-qxwt6" (OuterVolumeSpecName: "kube-api-access-qxwt6") pod "b6458faf-0340-43da-b4cf-4fa3b6384909" (UID: "b6458faf-0340-43da-b4cf-4fa3b6384909"). InnerVolumeSpecName "kube-api-access-qxwt6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:04:21 crc kubenswrapper[4885]: I1002 02:04:21.621692 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5adbde10-fd09-4e06-bad3-7f5d7bdf5b41-kube-api-access-58fgd" (OuterVolumeSpecName: "kube-api-access-58fgd") pod "5adbde10-fd09-4e06-bad3-7f5d7bdf5b41" (UID: "5adbde10-fd09-4e06-bad3-7f5d7bdf5b41"). InnerVolumeSpecName "kube-api-access-58fgd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:04:21 crc kubenswrapper[4885]: I1002 02:04:21.639863 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6458faf-0340-43da-b4cf-4fa3b6384909-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b6458faf-0340-43da-b4cf-4fa3b6384909" (UID: "b6458faf-0340-43da-b4cf-4fa3b6384909"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:04:21 crc kubenswrapper[4885]: I1002 02:04:21.649270 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5adbde10-fd09-4e06-bad3-7f5d7bdf5b41-config" (OuterVolumeSpecName: "config") pod "5adbde10-fd09-4e06-bad3-7f5d7bdf5b41" (UID: "5adbde10-fd09-4e06-bad3-7f5d7bdf5b41"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:04:21 crc kubenswrapper[4885]: I1002 02:04:21.653420 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5adbde10-fd09-4e06-bad3-7f5d7bdf5b41-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "5adbde10-fd09-4e06-bad3-7f5d7bdf5b41" (UID: "5adbde10-fd09-4e06-bad3-7f5d7bdf5b41"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:04:21 crc kubenswrapper[4885]: I1002 02:04:21.659675 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5adbde10-fd09-4e06-bad3-7f5d7bdf5b41-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5adbde10-fd09-4e06-bad3-7f5d7bdf5b41" (UID: "5adbde10-fd09-4e06-bad3-7f5d7bdf5b41"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:04:21 crc kubenswrapper[4885]: I1002 02:04:21.663788 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6458faf-0340-43da-b4cf-4fa3b6384909-config-data" (OuterVolumeSpecName: "config-data") pod "b6458faf-0340-43da-b4cf-4fa3b6384909" (UID: "b6458faf-0340-43da-b4cf-4fa3b6384909"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:04:21 crc kubenswrapper[4885]: I1002 02:04:21.667921 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5adbde10-fd09-4e06-bad3-7f5d7bdf5b41-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "5adbde10-fd09-4e06-bad3-7f5d7bdf5b41" (UID: "5adbde10-fd09-4e06-bad3-7f5d7bdf5b41"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:04:21 crc kubenswrapper[4885]: I1002 02:04:21.668509 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5adbde10-fd09-4e06-bad3-7f5d7bdf5b41-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "5adbde10-fd09-4e06-bad3-7f5d7bdf5b41" (UID: "5adbde10-fd09-4e06-bad3-7f5d7bdf5b41"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:04:21 crc kubenswrapper[4885]: I1002 02:04:21.706333 4885 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Oct 02 02:04:21 crc kubenswrapper[4885]: I1002 02:04:21.706369 4885 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5adbde10-fd09-4e06-bad3-7f5d7bdf5b41-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 02:04:21 crc kubenswrapper[4885]: I1002 02:04:21.706384 4885 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6458faf-0340-43da-b4cf-4fa3b6384909-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:04:21 crc kubenswrapper[4885]: I1002 02:04:21.706397 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qxwt6\" (UniqueName: \"kubernetes.io/projected/b6458faf-0340-43da-b4cf-4fa3b6384909-kube-api-access-qxwt6\") on node \"crc\" DevicePath \"\"" Oct 02 02:04:21 crc kubenswrapper[4885]: I1002 02:04:21.706410 4885 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5adbde10-fd09-4e06-bad3-7f5d7bdf5b41-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 02 02:04:21 crc kubenswrapper[4885]: I1002 02:04:21.706421 4885 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5adbde10-fd09-4e06-bad3-7f5d7bdf5b41-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 02:04:21 crc kubenswrapper[4885]: I1002 02:04:21.706431 4885 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5adbde10-fd09-4e06-bad3-7f5d7bdf5b41-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 02:04:21 crc kubenswrapper[4885]: I1002 02:04:21.706443 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-58fgd\" (UniqueName: \"kubernetes.io/projected/5adbde10-fd09-4e06-bad3-7f5d7bdf5b41-kube-api-access-58fgd\") on node \"crc\" DevicePath \"\"" Oct 02 02:04:21 crc kubenswrapper[4885]: I1002 02:04:21.706455 4885 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b6458faf-0340-43da-b4cf-4fa3b6384909-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 02:04:21 crc kubenswrapper[4885]: I1002 02:04:21.706465 4885 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b6458faf-0340-43da-b4cf-4fa3b6384909-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 02 02:04:21 crc kubenswrapper[4885]: I1002 02:04:21.706476 4885 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6458faf-0340-43da-b4cf-4fa3b6384909-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 02:04:21 crc kubenswrapper[4885]: I1002 02:04:21.706486 4885 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5adbde10-fd09-4e06-bad3-7f5d7bdf5b41-config\") on node \"crc\" DevicePath \"\"" Oct 02 02:04:21 crc kubenswrapper[4885]: I1002 02:04:21.722712 4885 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Oct 02 02:04:21 crc kubenswrapper[4885]: I1002 02:04:21.776102 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76fcf4b695-wpgst" Oct 02 02:04:21 crc kubenswrapper[4885]: I1002 02:04:21.776111 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76fcf4b695-wpgst" event={"ID":"5adbde10-fd09-4e06-bad3-7f5d7bdf5b41","Type":"ContainerDied","Data":"11a4f72ffa16280cdb1784abaa6fcfc17a2f8881f392cb3b30731dde19a72f5a"} Oct 02 02:04:21 crc kubenswrapper[4885]: I1002 02:04:21.776173 4885 scope.go:117] "RemoveContainer" containerID="aff3d2a87f29168594522e9b4f20c0cbbd55852066fe579d0b5a475463828e49" Oct 02 02:04:21 crc kubenswrapper[4885]: I1002 02:04:21.779967 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 02 02:04:21 crc kubenswrapper[4885]: I1002 02:04:21.780613 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b6458faf-0340-43da-b4cf-4fa3b6384909","Type":"ContainerDied","Data":"3cb09135a306c0546056c6e5a8611cd3876b37e41499f2cd78db8dadd2310e00"} Oct 02 02:04:21 crc kubenswrapper[4885]: I1002 02:04:21.807887 4885 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Oct 02 02:04:21 crc kubenswrapper[4885]: I1002 02:04:21.850685 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 02:04:21 crc kubenswrapper[4885]: I1002 02:04:21.858544 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 02:04:21 crc kubenswrapper[4885]: I1002 02:04:21.865119 4885 scope.go:117] "RemoveContainer" containerID="c6a5a15dcf6954c5e3e2d30d5a3cf3359566f0e2eaed48ec5e4bd65bef682e54" Oct 02 02:04:21 crc kubenswrapper[4885]: I1002 02:04:21.868686 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-76fcf4b695-wpgst"] Oct 02 02:04:21 crc kubenswrapper[4885]: I1002 02:04:21.880577 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-76fcf4b695-wpgst"] Oct 02 02:04:21 crc kubenswrapper[4885]: I1002 02:04:21.894009 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 02:04:21 crc kubenswrapper[4885]: E1002 02:04:21.894343 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6458faf-0340-43da-b4cf-4fa3b6384909" containerName="glance-log" Oct 02 02:04:21 crc kubenswrapper[4885]: I1002 02:04:21.894354 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6458faf-0340-43da-b4cf-4fa3b6384909" containerName="glance-log" Oct 02 02:04:21 crc kubenswrapper[4885]: E1002 02:04:21.894371 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5adbde10-fd09-4e06-bad3-7f5d7bdf5b41" containerName="init" Oct 02 02:04:21 crc kubenswrapper[4885]: I1002 02:04:21.894377 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="5adbde10-fd09-4e06-bad3-7f5d7bdf5b41" containerName="init" Oct 02 02:04:21 crc kubenswrapper[4885]: E1002 02:04:21.894388 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5adbde10-fd09-4e06-bad3-7f5d7bdf5b41" containerName="dnsmasq-dns" Oct 02 02:04:21 crc kubenswrapper[4885]: I1002 02:04:21.894394 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="5adbde10-fd09-4e06-bad3-7f5d7bdf5b41" containerName="dnsmasq-dns" Oct 02 02:04:21 crc kubenswrapper[4885]: E1002 02:04:21.894417 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6458faf-0340-43da-b4cf-4fa3b6384909" containerName="glance-httpd" Oct 02 02:04:21 crc kubenswrapper[4885]: I1002 02:04:21.894423 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6458faf-0340-43da-b4cf-4fa3b6384909" containerName="glance-httpd" Oct 02 02:04:21 crc kubenswrapper[4885]: I1002 02:04:21.894592 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="5adbde10-fd09-4e06-bad3-7f5d7bdf5b41" containerName="dnsmasq-dns" Oct 02 02:04:21 crc kubenswrapper[4885]: I1002 02:04:21.894605 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6458faf-0340-43da-b4cf-4fa3b6384909" containerName="glance-httpd" Oct 02 02:04:21 crc kubenswrapper[4885]: I1002 02:04:21.894616 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6458faf-0340-43da-b4cf-4fa3b6384909" containerName="glance-log" Oct 02 02:04:21 crc kubenswrapper[4885]: I1002 02:04:21.895407 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 02:04:21 crc kubenswrapper[4885]: I1002 02:04:21.895485 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 02 02:04:21 crc kubenswrapper[4885]: I1002 02:04:21.899725 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 02 02:04:21 crc kubenswrapper[4885]: I1002 02:04:21.900608 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 02 02:04:21 crc kubenswrapper[4885]: I1002 02:04:21.910226 4885 scope.go:117] "RemoveContainer" containerID="e7966a9c7055b70c3575805bff8da5acbb90942af542595e888c8c2db22587f4" Oct 02 02:04:21 crc kubenswrapper[4885]: I1002 02:04:21.949380 4885 scope.go:117] "RemoveContainer" containerID="b4cfd3f7c1947afe37521744cf9538f5d916bf4e5eebeae8ba4abf7aba4b817f" Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.011292 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ztxwm\" (UniqueName: \"kubernetes.io/projected/1aff4106-2f4b-48ef-93a1-d51285ce6242-kube-api-access-ztxwm\") pod \"glance-default-internal-api-0\" (UID: \"1aff4106-2f4b-48ef-93a1-d51285ce6242\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.011361 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1aff4106-2f4b-48ef-93a1-d51285ce6242-config-data\") pod \"glance-default-internal-api-0\" (UID: \"1aff4106-2f4b-48ef-93a1-d51285ce6242\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.011408 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"1aff4106-2f4b-48ef-93a1-d51285ce6242\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.011457 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1aff4106-2f4b-48ef-93a1-d51285ce6242-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"1aff4106-2f4b-48ef-93a1-d51285ce6242\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.011477 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1aff4106-2f4b-48ef-93a1-d51285ce6242-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"1aff4106-2f4b-48ef-93a1-d51285ce6242\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.011507 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1aff4106-2f4b-48ef-93a1-d51285ce6242-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"1aff4106-2f4b-48ef-93a1-d51285ce6242\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.011567 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1aff4106-2f4b-48ef-93a1-d51285ce6242-scripts\") pod \"glance-default-internal-api-0\" (UID: \"1aff4106-2f4b-48ef-93a1-d51285ce6242\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.011601 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1aff4106-2f4b-48ef-93a1-d51285ce6242-logs\") pod \"glance-default-internal-api-0\" (UID: \"1aff4106-2f4b-48ef-93a1-d51285ce6242\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.061918 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5adbde10-fd09-4e06-bad3-7f5d7bdf5b41" path="/var/lib/kubelet/pods/5adbde10-fd09-4e06-bad3-7f5d7bdf5b41/volumes" Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.063449 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6458faf-0340-43da-b4cf-4fa3b6384909" path="/var/lib/kubelet/pods/b6458faf-0340-43da-b4cf-4fa3b6384909/volumes" Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.113155 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1aff4106-2f4b-48ef-93a1-d51285ce6242-scripts\") pod \"glance-default-internal-api-0\" (UID: \"1aff4106-2f4b-48ef-93a1-d51285ce6242\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.113208 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1aff4106-2f4b-48ef-93a1-d51285ce6242-logs\") pod \"glance-default-internal-api-0\" (UID: \"1aff4106-2f4b-48ef-93a1-d51285ce6242\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.113236 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ztxwm\" (UniqueName: \"kubernetes.io/projected/1aff4106-2f4b-48ef-93a1-d51285ce6242-kube-api-access-ztxwm\") pod \"glance-default-internal-api-0\" (UID: \"1aff4106-2f4b-48ef-93a1-d51285ce6242\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.113250 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1aff4106-2f4b-48ef-93a1-d51285ce6242-config-data\") pod \"glance-default-internal-api-0\" (UID: \"1aff4106-2f4b-48ef-93a1-d51285ce6242\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.113646 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"1aff4106-2f4b-48ef-93a1-d51285ce6242\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.113734 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1aff4106-2f4b-48ef-93a1-d51285ce6242-logs\") pod \"glance-default-internal-api-0\" (UID: \"1aff4106-2f4b-48ef-93a1-d51285ce6242\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.114138 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1aff4106-2f4b-48ef-93a1-d51285ce6242-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"1aff4106-2f4b-48ef-93a1-d51285ce6242\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.114164 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1aff4106-2f4b-48ef-93a1-d51285ce6242-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"1aff4106-2f4b-48ef-93a1-d51285ce6242\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.114192 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1aff4106-2f4b-48ef-93a1-d51285ce6242-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"1aff4106-2f4b-48ef-93a1-d51285ce6242\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.114457 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1aff4106-2f4b-48ef-93a1-d51285ce6242-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"1aff4106-2f4b-48ef-93a1-d51285ce6242\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.114566 4885 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"1aff4106-2f4b-48ef-93a1-d51285ce6242\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/glance-default-internal-api-0" Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.120140 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1aff4106-2f4b-48ef-93a1-d51285ce6242-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"1aff4106-2f4b-48ef-93a1-d51285ce6242\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.130463 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1aff4106-2f4b-48ef-93a1-d51285ce6242-scripts\") pod \"glance-default-internal-api-0\" (UID: \"1aff4106-2f4b-48ef-93a1-d51285ce6242\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.130533 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1aff4106-2f4b-48ef-93a1-d51285ce6242-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"1aff4106-2f4b-48ef-93a1-d51285ce6242\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.134151 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1aff4106-2f4b-48ef-93a1-d51285ce6242-config-data\") pod \"glance-default-internal-api-0\" (UID: \"1aff4106-2f4b-48ef-93a1-d51285ce6242\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.150971 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ztxwm\" (UniqueName: \"kubernetes.io/projected/1aff4106-2f4b-48ef-93a1-d51285ce6242-kube-api-access-ztxwm\") pod \"glance-default-internal-api-0\" (UID: \"1aff4106-2f4b-48ef-93a1-d51285ce6242\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.187347 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"1aff4106-2f4b-48ef-93a1-d51285ce6242\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.222205 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.263759 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-dr95w" Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.374076 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.426721 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d4a98a2-1f96-47a2-bacc-a2dd42bc4119-scripts\") pod \"1d4a98a2-1f96-47a2-bacc-a2dd42bc4119\" (UID: \"1d4a98a2-1f96-47a2-bacc-a2dd42bc4119\") " Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.426791 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1d4a98a2-1f96-47a2-bacc-a2dd42bc4119-fernet-keys\") pod \"1d4a98a2-1f96-47a2-bacc-a2dd42bc4119\" (UID: \"1d4a98a2-1f96-47a2-bacc-a2dd42bc4119\") " Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.426826 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d4a98a2-1f96-47a2-bacc-a2dd42bc4119-config-data\") pod \"1d4a98a2-1f96-47a2-bacc-a2dd42bc4119\" (UID: \"1d4a98a2-1f96-47a2-bacc-a2dd42bc4119\") " Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.426844 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1d4a98a2-1f96-47a2-bacc-a2dd42bc4119-credential-keys\") pod \"1d4a98a2-1f96-47a2-bacc-a2dd42bc4119\" (UID: \"1d4a98a2-1f96-47a2-bacc-a2dd42bc4119\") " Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.426880 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d4a98a2-1f96-47a2-bacc-a2dd42bc4119-combined-ca-bundle\") pod \"1d4a98a2-1f96-47a2-bacc-a2dd42bc4119\" (UID: \"1d4a98a2-1f96-47a2-bacc-a2dd42bc4119\") " Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.426916 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jzrlb\" (UniqueName: \"kubernetes.io/projected/1d4a98a2-1f96-47a2-bacc-a2dd42bc4119-kube-api-access-jzrlb\") pod \"1d4a98a2-1f96-47a2-bacc-a2dd42bc4119\" (UID: \"1d4a98a2-1f96-47a2-bacc-a2dd42bc4119\") " Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.430788 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d4a98a2-1f96-47a2-bacc-a2dd42bc4119-kube-api-access-jzrlb" (OuterVolumeSpecName: "kube-api-access-jzrlb") pod "1d4a98a2-1f96-47a2-bacc-a2dd42bc4119" (UID: "1d4a98a2-1f96-47a2-bacc-a2dd42bc4119"). InnerVolumeSpecName "kube-api-access-jzrlb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.431317 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d4a98a2-1f96-47a2-bacc-a2dd42bc4119-scripts" (OuterVolumeSpecName: "scripts") pod "1d4a98a2-1f96-47a2-bacc-a2dd42bc4119" (UID: "1d4a98a2-1f96-47a2-bacc-a2dd42bc4119"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.433702 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d4a98a2-1f96-47a2-bacc-a2dd42bc4119-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "1d4a98a2-1f96-47a2-bacc-a2dd42bc4119" (UID: "1d4a98a2-1f96-47a2-bacc-a2dd42bc4119"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.433866 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d4a98a2-1f96-47a2-bacc-a2dd42bc4119-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "1d4a98a2-1f96-47a2-bacc-a2dd42bc4119" (UID: "1d4a98a2-1f96-47a2-bacc-a2dd42bc4119"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.458841 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d4a98a2-1f96-47a2-bacc-a2dd42bc4119-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1d4a98a2-1f96-47a2-bacc-a2dd42bc4119" (UID: "1d4a98a2-1f96-47a2-bacc-a2dd42bc4119"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.459777 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d4a98a2-1f96-47a2-bacc-a2dd42bc4119-config-data" (OuterVolumeSpecName: "config-data") pod "1d4a98a2-1f96-47a2-bacc-a2dd42bc4119" (UID: "1d4a98a2-1f96-47a2-bacc-a2dd42bc4119"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.528408 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d091410b-e1bf-423e-929f-18d2a9c65891-combined-ca-bundle\") pod \"d091410b-e1bf-423e-929f-18d2a9c65891\" (UID: \"d091410b-e1bf-423e-929f-18d2a9c65891\") " Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.528516 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s72zw\" (UniqueName: \"kubernetes.io/projected/d091410b-e1bf-423e-929f-18d2a9c65891-kube-api-access-s72zw\") pod \"d091410b-e1bf-423e-929f-18d2a9c65891\" (UID: \"d091410b-e1bf-423e-929f-18d2a9c65891\") " Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.528582 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d091410b-e1bf-423e-929f-18d2a9c65891-config-data\") pod \"d091410b-e1bf-423e-929f-18d2a9c65891\" (UID: \"d091410b-e1bf-423e-929f-18d2a9c65891\") " Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.528618 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d091410b-e1bf-423e-929f-18d2a9c65891-logs\") pod \"d091410b-e1bf-423e-929f-18d2a9c65891\" (UID: \"d091410b-e1bf-423e-929f-18d2a9c65891\") " Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.528660 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d091410b-e1bf-423e-929f-18d2a9c65891-httpd-run\") pod \"d091410b-e1bf-423e-929f-18d2a9c65891\" (UID: \"d091410b-e1bf-423e-929f-18d2a9c65891\") " Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.528798 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d091410b-e1bf-423e-929f-18d2a9c65891-scripts\") pod \"d091410b-e1bf-423e-929f-18d2a9c65891\" (UID: \"d091410b-e1bf-423e-929f-18d2a9c65891\") " Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.528822 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"d091410b-e1bf-423e-929f-18d2a9c65891\" (UID: \"d091410b-e1bf-423e-929f-18d2a9c65891\") " Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.529238 4885 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d4a98a2-1f96-47a2-bacc-a2dd42bc4119-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.529277 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jzrlb\" (UniqueName: \"kubernetes.io/projected/1d4a98a2-1f96-47a2-bacc-a2dd42bc4119-kube-api-access-jzrlb\") on node \"crc\" DevicePath \"\"" Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.529293 4885 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d4a98a2-1f96-47a2-bacc-a2dd42bc4119-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.529305 4885 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1d4a98a2-1f96-47a2-bacc-a2dd42bc4119-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.529316 4885 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d4a98a2-1f96-47a2-bacc-a2dd42bc4119-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.529328 4885 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1d4a98a2-1f96-47a2-bacc-a2dd42bc4119-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.530951 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d091410b-e1bf-423e-929f-18d2a9c65891-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "d091410b-e1bf-423e-929f-18d2a9c65891" (UID: "d091410b-e1bf-423e-929f-18d2a9c65891"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.531147 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d091410b-e1bf-423e-929f-18d2a9c65891-logs" (OuterVolumeSpecName: "logs") pod "d091410b-e1bf-423e-929f-18d2a9c65891" (UID: "d091410b-e1bf-423e-929f-18d2a9c65891"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.551494 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance") pod "d091410b-e1bf-423e-929f-18d2a9c65891" (UID: "d091410b-e1bf-423e-929f-18d2a9c65891"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.551509 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d091410b-e1bf-423e-929f-18d2a9c65891-scripts" (OuterVolumeSpecName: "scripts") pod "d091410b-e1bf-423e-929f-18d2a9c65891" (UID: "d091410b-e1bf-423e-929f-18d2a9c65891"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.554761 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d091410b-e1bf-423e-929f-18d2a9c65891-kube-api-access-s72zw" (OuterVolumeSpecName: "kube-api-access-s72zw") pod "d091410b-e1bf-423e-929f-18d2a9c65891" (UID: "d091410b-e1bf-423e-929f-18d2a9c65891"). InnerVolumeSpecName "kube-api-access-s72zw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.630906 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d091410b-e1bf-423e-929f-18d2a9c65891-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d091410b-e1bf-423e-929f-18d2a9c65891" (UID: "d091410b-e1bf-423e-929f-18d2a9c65891"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.631138 4885 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d091410b-e1bf-423e-929f-18d2a9c65891-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.631812 4885 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d091410b-e1bf-423e-929f-18d2a9c65891-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.631899 4885 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.631912 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s72zw\" (UniqueName: \"kubernetes.io/projected/d091410b-e1bf-423e-929f-18d2a9c65891-kube-api-access-s72zw\") on node \"crc\" DevicePath \"\"" Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.631921 4885 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d091410b-e1bf-423e-929f-18d2a9c65891-logs\") on node \"crc\" DevicePath \"\"" Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.638861 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d091410b-e1bf-423e-929f-18d2a9c65891-config-data" (OuterVolumeSpecName: "config-data") pod "d091410b-e1bf-423e-929f-18d2a9c65891" (UID: "d091410b-e1bf-423e-929f-18d2a9c65891"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.715148 4885 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.733392 4885 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.733419 4885 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d091410b-e1bf-423e-929f-18d2a9c65891-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.733430 4885 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d091410b-e1bf-423e-929f-18d2a9c65891-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.828510 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9226a04e-0195-42a1-b651-2f5379b1e2a9","Type":"ContainerStarted","Data":"cc2a1cdc982003a22e30118b6756d23620f619fe3029c82308eb0d2d8ccbcfa2"} Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.839905 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-dr95w" event={"ID":"1d4a98a2-1f96-47a2-bacc-a2dd42bc4119","Type":"ContainerDied","Data":"15942f3d3388cc1972ff0c9b83d7728fba10a2f5975a6c727868882c69b37f18"} Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.839963 4885 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="15942f3d3388cc1972ff0c9b83d7728fba10a2f5975a6c727868882c69b37f18" Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.840045 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-dr95w" Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.855532 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.857026 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d091410b-e1bf-423e-929f-18d2a9c65891","Type":"ContainerDied","Data":"585f5e96d87df6f4e4087afa3df8bf0be66da056f6a5252ea24694ddc51ebdc4"} Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.857063 4885 scope.go:117] "RemoveContainer" containerID="70c7aba074612eaff2e55f0f7d4ebed65c6d37d359fa30ef18084ec9df97e1aa" Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.857189 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.897823 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-vj2tf" event={"ID":"8fa66af9-7b4b-480c-8f96-bda850e2eef6","Type":"ContainerStarted","Data":"8d9aa7b7c3aacfe216138f05551b49ff0afe4a9f79e28ef88f3baf68e620b0fb"} Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.913483 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-59f7bf489d-7tfk4"] Oct 02 02:04:22 crc kubenswrapper[4885]: E1002 02:04:22.913800 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d091410b-e1bf-423e-929f-18d2a9c65891" containerName="glance-httpd" Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.913816 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="d091410b-e1bf-423e-929f-18d2a9c65891" containerName="glance-httpd" Oct 02 02:04:22 crc kubenswrapper[4885]: E1002 02:04:22.913844 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d4a98a2-1f96-47a2-bacc-a2dd42bc4119" containerName="keystone-bootstrap" Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.913851 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d4a98a2-1f96-47a2-bacc-a2dd42bc4119" containerName="keystone-bootstrap" Oct 02 02:04:22 crc kubenswrapper[4885]: E1002 02:04:22.913861 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d091410b-e1bf-423e-929f-18d2a9c65891" containerName="glance-log" Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.913867 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="d091410b-e1bf-423e-929f-18d2a9c65891" containerName="glance-log" Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.914026 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="d091410b-e1bf-423e-929f-18d2a9c65891" containerName="glance-httpd" Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.914045 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="d091410b-e1bf-423e-929f-18d2a9c65891" containerName="glance-log" Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.914055 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d4a98a2-1f96-47a2-bacc-a2dd42bc4119" containerName="keystone-bootstrap" Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.914652 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-59f7bf489d-7tfk4" Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.919611 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.919826 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-h4g4g" Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.919927 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.920039 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.920139 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.920239 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.941783 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-59f7bf489d-7tfk4"] Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.950844 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-vj2tf" podStartSLOduration=2.480097748 podStartE2EDuration="44.950823584s" podCreationTimestamp="2025-10-02 02:03:38 +0000 UTC" firstStartedPulling="2025-10-02 02:03:39.995833476 +0000 UTC m=+1008.807580875" lastFinishedPulling="2025-10-02 02:04:22.466559312 +0000 UTC m=+1051.278306711" observedRunningTime="2025-10-02 02:04:22.911811962 +0000 UTC m=+1051.723559361" watchObservedRunningTime="2025-10-02 02:04:22.950823584 +0000 UTC m=+1051.762570983" Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.957911 4885 scope.go:117] "RemoveContainer" containerID="775dde44ca3ce75c6bc52c6fbbe3f6e7cdbf8f9c303eb816d831b82b8e22f86c" Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.958052 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.973733 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.988952 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.990602 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.995046 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 02 02:04:22 crc kubenswrapper[4885]: I1002 02:04:22.995339 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 02 02:04:23 crc kubenswrapper[4885]: I1002 02:04:23.021359 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 02:04:23 crc kubenswrapper[4885]: I1002 02:04:23.039196 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5f5ca9c7-0cfd-4c95-a160-4d909f8f3d00-scripts\") pod \"keystone-59f7bf489d-7tfk4\" (UID: \"5f5ca9c7-0cfd-4c95-a160-4d909f8f3d00\") " pod="openstack/keystone-59f7bf489d-7tfk4" Oct 02 02:04:23 crc kubenswrapper[4885]: I1002 02:04:23.039245 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5f5ca9c7-0cfd-4c95-a160-4d909f8f3d00-public-tls-certs\") pod \"keystone-59f7bf489d-7tfk4\" (UID: \"5f5ca9c7-0cfd-4c95-a160-4d909f8f3d00\") " pod="openstack/keystone-59f7bf489d-7tfk4" Oct 02 02:04:23 crc kubenswrapper[4885]: I1002 02:04:23.039295 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pxckw\" (UniqueName: \"kubernetes.io/projected/5f5ca9c7-0cfd-4c95-a160-4d909f8f3d00-kube-api-access-pxckw\") pod \"keystone-59f7bf489d-7tfk4\" (UID: \"5f5ca9c7-0cfd-4c95-a160-4d909f8f3d00\") " pod="openstack/keystone-59f7bf489d-7tfk4" Oct 02 02:04:23 crc kubenswrapper[4885]: I1002 02:04:23.039377 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5f5ca9c7-0cfd-4c95-a160-4d909f8f3d00-credential-keys\") pod \"keystone-59f7bf489d-7tfk4\" (UID: \"5f5ca9c7-0cfd-4c95-a160-4d909f8f3d00\") " pod="openstack/keystone-59f7bf489d-7tfk4" Oct 02 02:04:23 crc kubenswrapper[4885]: I1002 02:04:23.039401 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5f5ca9c7-0cfd-4c95-a160-4d909f8f3d00-fernet-keys\") pod \"keystone-59f7bf489d-7tfk4\" (UID: \"5f5ca9c7-0cfd-4c95-a160-4d909f8f3d00\") " pod="openstack/keystone-59f7bf489d-7tfk4" Oct 02 02:04:23 crc kubenswrapper[4885]: I1002 02:04:23.039416 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5f5ca9c7-0cfd-4c95-a160-4d909f8f3d00-internal-tls-certs\") pod \"keystone-59f7bf489d-7tfk4\" (UID: \"5f5ca9c7-0cfd-4c95-a160-4d909f8f3d00\") " pod="openstack/keystone-59f7bf489d-7tfk4" Oct 02 02:04:23 crc kubenswrapper[4885]: I1002 02:04:23.039436 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f5ca9c7-0cfd-4c95-a160-4d909f8f3d00-combined-ca-bundle\") pod \"keystone-59f7bf489d-7tfk4\" (UID: \"5f5ca9c7-0cfd-4c95-a160-4d909f8f3d00\") " pod="openstack/keystone-59f7bf489d-7tfk4" Oct 02 02:04:23 crc kubenswrapper[4885]: I1002 02:04:23.055170 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f5ca9c7-0cfd-4c95-a160-4d909f8f3d00-config-data\") pod \"keystone-59f7bf489d-7tfk4\" (UID: \"5f5ca9c7-0cfd-4c95-a160-4d909f8f3d00\") " pod="openstack/keystone-59f7bf489d-7tfk4" Oct 02 02:04:23 crc kubenswrapper[4885]: I1002 02:04:23.157039 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5f5ca9c7-0cfd-4c95-a160-4d909f8f3d00-fernet-keys\") pod \"keystone-59f7bf489d-7tfk4\" (UID: \"5f5ca9c7-0cfd-4c95-a160-4d909f8f3d00\") " pod="openstack/keystone-59f7bf489d-7tfk4" Oct 02 02:04:23 crc kubenswrapper[4885]: I1002 02:04:23.157078 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5f5ca9c7-0cfd-4c95-a160-4d909f8f3d00-internal-tls-certs\") pod \"keystone-59f7bf489d-7tfk4\" (UID: \"5f5ca9c7-0cfd-4c95-a160-4d909f8f3d00\") " pod="openstack/keystone-59f7bf489d-7tfk4" Oct 02 02:04:23 crc kubenswrapper[4885]: I1002 02:04:23.157099 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f5ca9c7-0cfd-4c95-a160-4d909f8f3d00-combined-ca-bundle\") pod \"keystone-59f7bf489d-7tfk4\" (UID: \"5f5ca9c7-0cfd-4c95-a160-4d909f8f3d00\") " pod="openstack/keystone-59f7bf489d-7tfk4" Oct 02 02:04:23 crc kubenswrapper[4885]: I1002 02:04:23.157120 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/345ef025-b8fb-4955-baec-9af9be17792e-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"345ef025-b8fb-4955-baec-9af9be17792e\") " pod="openstack/glance-default-external-api-0" Oct 02 02:04:23 crc kubenswrapper[4885]: I1002 02:04:23.157720 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f5ca9c7-0cfd-4c95-a160-4d909f8f3d00-config-data\") pod \"keystone-59f7bf489d-7tfk4\" (UID: \"5f5ca9c7-0cfd-4c95-a160-4d909f8f3d00\") " pod="openstack/keystone-59f7bf489d-7tfk4" Oct 02 02:04:23 crc kubenswrapper[4885]: I1002 02:04:23.157742 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/345ef025-b8fb-4955-baec-9af9be17792e-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"345ef025-b8fb-4955-baec-9af9be17792e\") " pod="openstack/glance-default-external-api-0" Oct 02 02:04:23 crc kubenswrapper[4885]: I1002 02:04:23.157796 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/345ef025-b8fb-4955-baec-9af9be17792e-scripts\") pod \"glance-default-external-api-0\" (UID: \"345ef025-b8fb-4955-baec-9af9be17792e\") " pod="openstack/glance-default-external-api-0" Oct 02 02:04:23 crc kubenswrapper[4885]: I1002 02:04:23.157825 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5f5ca9c7-0cfd-4c95-a160-4d909f8f3d00-scripts\") pod \"keystone-59f7bf489d-7tfk4\" (UID: \"5f5ca9c7-0cfd-4c95-a160-4d909f8f3d00\") " pod="openstack/keystone-59f7bf489d-7tfk4" Oct 02 02:04:23 crc kubenswrapper[4885]: I1002 02:04:23.157846 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xkjt2\" (UniqueName: \"kubernetes.io/projected/345ef025-b8fb-4955-baec-9af9be17792e-kube-api-access-xkjt2\") pod \"glance-default-external-api-0\" (UID: \"345ef025-b8fb-4955-baec-9af9be17792e\") " pod="openstack/glance-default-external-api-0" Oct 02 02:04:23 crc kubenswrapper[4885]: I1002 02:04:23.157878 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5f5ca9c7-0cfd-4c95-a160-4d909f8f3d00-public-tls-certs\") pod \"keystone-59f7bf489d-7tfk4\" (UID: \"5f5ca9c7-0cfd-4c95-a160-4d909f8f3d00\") " pod="openstack/keystone-59f7bf489d-7tfk4" Oct 02 02:04:23 crc kubenswrapper[4885]: I1002 02:04:23.157903 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/345ef025-b8fb-4955-baec-9af9be17792e-config-data\") pod \"glance-default-external-api-0\" (UID: \"345ef025-b8fb-4955-baec-9af9be17792e\") " pod="openstack/glance-default-external-api-0" Oct 02 02:04:23 crc kubenswrapper[4885]: I1002 02:04:23.157932 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pxckw\" (UniqueName: \"kubernetes.io/projected/5f5ca9c7-0cfd-4c95-a160-4d909f8f3d00-kube-api-access-pxckw\") pod \"keystone-59f7bf489d-7tfk4\" (UID: \"5f5ca9c7-0cfd-4c95-a160-4d909f8f3d00\") " pod="openstack/keystone-59f7bf489d-7tfk4" Oct 02 02:04:23 crc kubenswrapper[4885]: I1002 02:04:23.157997 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/345ef025-b8fb-4955-baec-9af9be17792e-logs\") pod \"glance-default-external-api-0\" (UID: \"345ef025-b8fb-4955-baec-9af9be17792e\") " pod="openstack/glance-default-external-api-0" Oct 02 02:04:23 crc kubenswrapper[4885]: I1002 02:04:23.158017 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"345ef025-b8fb-4955-baec-9af9be17792e\") " pod="openstack/glance-default-external-api-0" Oct 02 02:04:23 crc kubenswrapper[4885]: I1002 02:04:23.158032 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/345ef025-b8fb-4955-baec-9af9be17792e-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"345ef025-b8fb-4955-baec-9af9be17792e\") " pod="openstack/glance-default-external-api-0" Oct 02 02:04:23 crc kubenswrapper[4885]: I1002 02:04:23.158078 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5f5ca9c7-0cfd-4c95-a160-4d909f8f3d00-credential-keys\") pod \"keystone-59f7bf489d-7tfk4\" (UID: \"5f5ca9c7-0cfd-4c95-a160-4d909f8f3d00\") " pod="openstack/keystone-59f7bf489d-7tfk4" Oct 02 02:04:23 crc kubenswrapper[4885]: I1002 02:04:23.176016 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f5ca9c7-0cfd-4c95-a160-4d909f8f3d00-combined-ca-bundle\") pod \"keystone-59f7bf489d-7tfk4\" (UID: \"5f5ca9c7-0cfd-4c95-a160-4d909f8f3d00\") " pod="openstack/keystone-59f7bf489d-7tfk4" Oct 02 02:04:23 crc kubenswrapper[4885]: I1002 02:04:23.181082 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5f5ca9c7-0cfd-4c95-a160-4d909f8f3d00-scripts\") pod \"keystone-59f7bf489d-7tfk4\" (UID: \"5f5ca9c7-0cfd-4c95-a160-4d909f8f3d00\") " pod="openstack/keystone-59f7bf489d-7tfk4" Oct 02 02:04:23 crc kubenswrapper[4885]: I1002 02:04:23.181791 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5f5ca9c7-0cfd-4c95-a160-4d909f8f3d00-public-tls-certs\") pod \"keystone-59f7bf489d-7tfk4\" (UID: \"5f5ca9c7-0cfd-4c95-a160-4d909f8f3d00\") " pod="openstack/keystone-59f7bf489d-7tfk4" Oct 02 02:04:23 crc kubenswrapper[4885]: I1002 02:04:23.182028 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5f5ca9c7-0cfd-4c95-a160-4d909f8f3d00-fernet-keys\") pod \"keystone-59f7bf489d-7tfk4\" (UID: \"5f5ca9c7-0cfd-4c95-a160-4d909f8f3d00\") " pod="openstack/keystone-59f7bf489d-7tfk4" Oct 02 02:04:23 crc kubenswrapper[4885]: I1002 02:04:23.182479 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5f5ca9c7-0cfd-4c95-a160-4d909f8f3d00-credential-keys\") pod \"keystone-59f7bf489d-7tfk4\" (UID: \"5f5ca9c7-0cfd-4c95-a160-4d909f8f3d00\") " pod="openstack/keystone-59f7bf489d-7tfk4" Oct 02 02:04:23 crc kubenswrapper[4885]: I1002 02:04:23.183310 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5f5ca9c7-0cfd-4c95-a160-4d909f8f3d00-internal-tls-certs\") pod \"keystone-59f7bf489d-7tfk4\" (UID: \"5f5ca9c7-0cfd-4c95-a160-4d909f8f3d00\") " pod="openstack/keystone-59f7bf489d-7tfk4" Oct 02 02:04:23 crc kubenswrapper[4885]: I1002 02:04:23.185755 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f5ca9c7-0cfd-4c95-a160-4d909f8f3d00-config-data\") pod \"keystone-59f7bf489d-7tfk4\" (UID: \"5f5ca9c7-0cfd-4c95-a160-4d909f8f3d00\") " pod="openstack/keystone-59f7bf489d-7tfk4" Oct 02 02:04:23 crc kubenswrapper[4885]: I1002 02:04:23.198731 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pxckw\" (UniqueName: \"kubernetes.io/projected/5f5ca9c7-0cfd-4c95-a160-4d909f8f3d00-kube-api-access-pxckw\") pod \"keystone-59f7bf489d-7tfk4\" (UID: \"5f5ca9c7-0cfd-4c95-a160-4d909f8f3d00\") " pod="openstack/keystone-59f7bf489d-7tfk4" Oct 02 02:04:23 crc kubenswrapper[4885]: I1002 02:04:23.235723 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-59f7bf489d-7tfk4" Oct 02 02:04:23 crc kubenswrapper[4885]: I1002 02:04:23.260678 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/345ef025-b8fb-4955-baec-9af9be17792e-scripts\") pod \"glance-default-external-api-0\" (UID: \"345ef025-b8fb-4955-baec-9af9be17792e\") " pod="openstack/glance-default-external-api-0" Oct 02 02:04:23 crc kubenswrapper[4885]: I1002 02:04:23.260736 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xkjt2\" (UniqueName: \"kubernetes.io/projected/345ef025-b8fb-4955-baec-9af9be17792e-kube-api-access-xkjt2\") pod \"glance-default-external-api-0\" (UID: \"345ef025-b8fb-4955-baec-9af9be17792e\") " pod="openstack/glance-default-external-api-0" Oct 02 02:04:23 crc kubenswrapper[4885]: I1002 02:04:23.260781 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/345ef025-b8fb-4955-baec-9af9be17792e-config-data\") pod \"glance-default-external-api-0\" (UID: \"345ef025-b8fb-4955-baec-9af9be17792e\") " pod="openstack/glance-default-external-api-0" Oct 02 02:04:23 crc kubenswrapper[4885]: I1002 02:04:23.260839 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/345ef025-b8fb-4955-baec-9af9be17792e-logs\") pod \"glance-default-external-api-0\" (UID: \"345ef025-b8fb-4955-baec-9af9be17792e\") " pod="openstack/glance-default-external-api-0" Oct 02 02:04:23 crc kubenswrapper[4885]: I1002 02:04:23.260862 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"345ef025-b8fb-4955-baec-9af9be17792e\") " pod="openstack/glance-default-external-api-0" Oct 02 02:04:23 crc kubenswrapper[4885]: I1002 02:04:23.260876 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/345ef025-b8fb-4955-baec-9af9be17792e-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"345ef025-b8fb-4955-baec-9af9be17792e\") " pod="openstack/glance-default-external-api-0" Oct 02 02:04:23 crc kubenswrapper[4885]: I1002 02:04:23.260920 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/345ef025-b8fb-4955-baec-9af9be17792e-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"345ef025-b8fb-4955-baec-9af9be17792e\") " pod="openstack/glance-default-external-api-0" Oct 02 02:04:23 crc kubenswrapper[4885]: I1002 02:04:23.260945 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/345ef025-b8fb-4955-baec-9af9be17792e-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"345ef025-b8fb-4955-baec-9af9be17792e\") " pod="openstack/glance-default-external-api-0" Oct 02 02:04:23 crc kubenswrapper[4885]: I1002 02:04:23.261425 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/345ef025-b8fb-4955-baec-9af9be17792e-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"345ef025-b8fb-4955-baec-9af9be17792e\") " pod="openstack/glance-default-external-api-0" Oct 02 02:04:23 crc kubenswrapper[4885]: I1002 02:04:23.262580 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/345ef025-b8fb-4955-baec-9af9be17792e-logs\") pod \"glance-default-external-api-0\" (UID: \"345ef025-b8fb-4955-baec-9af9be17792e\") " pod="openstack/glance-default-external-api-0" Oct 02 02:04:23 crc kubenswrapper[4885]: I1002 02:04:23.270683 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/345ef025-b8fb-4955-baec-9af9be17792e-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"345ef025-b8fb-4955-baec-9af9be17792e\") " pod="openstack/glance-default-external-api-0" Oct 02 02:04:23 crc kubenswrapper[4885]: I1002 02:04:23.272153 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/345ef025-b8fb-4955-baec-9af9be17792e-scripts\") pod \"glance-default-external-api-0\" (UID: \"345ef025-b8fb-4955-baec-9af9be17792e\") " pod="openstack/glance-default-external-api-0" Oct 02 02:04:23 crc kubenswrapper[4885]: I1002 02:04:23.272800 4885 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"345ef025-b8fb-4955-baec-9af9be17792e\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/glance-default-external-api-0" Oct 02 02:04:23 crc kubenswrapper[4885]: I1002 02:04:23.274143 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/345ef025-b8fb-4955-baec-9af9be17792e-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"345ef025-b8fb-4955-baec-9af9be17792e\") " pod="openstack/glance-default-external-api-0" Oct 02 02:04:23 crc kubenswrapper[4885]: I1002 02:04:23.287038 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xkjt2\" (UniqueName: \"kubernetes.io/projected/345ef025-b8fb-4955-baec-9af9be17792e-kube-api-access-xkjt2\") pod \"glance-default-external-api-0\" (UID: \"345ef025-b8fb-4955-baec-9af9be17792e\") " pod="openstack/glance-default-external-api-0" Oct 02 02:04:23 crc kubenswrapper[4885]: I1002 02:04:23.288093 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/345ef025-b8fb-4955-baec-9af9be17792e-config-data\") pod \"glance-default-external-api-0\" (UID: \"345ef025-b8fb-4955-baec-9af9be17792e\") " pod="openstack/glance-default-external-api-0" Oct 02 02:04:23 crc kubenswrapper[4885]: I1002 02:04:23.322430 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"345ef025-b8fb-4955-baec-9af9be17792e\") " pod="openstack/glance-default-external-api-0" Oct 02 02:04:23 crc kubenswrapper[4885]: I1002 02:04:23.395966 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 02 02:04:23 crc kubenswrapper[4885]: I1002 02:04:23.754439 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-59f7bf489d-7tfk4"] Oct 02 02:04:23 crc kubenswrapper[4885]: I1002 02:04:23.954700 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1aff4106-2f4b-48ef-93a1-d51285ce6242","Type":"ContainerStarted","Data":"71ffbb4c2cbbe9022cc1a62d8f0fc6cd2ac0fed0d96f405f4ec7589f182ddbbd"} Oct 02 02:04:23 crc kubenswrapper[4885]: I1002 02:04:23.954958 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1aff4106-2f4b-48ef-93a1-d51285ce6242","Type":"ContainerStarted","Data":"19a5f22bd884950be4e1415ac73ee8a37e2dfbc16378d95a7b8d8e6d45ed33e9"} Oct 02 02:04:23 crc kubenswrapper[4885]: I1002 02:04:23.957408 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-59f7bf489d-7tfk4" event={"ID":"5f5ca9c7-0cfd-4c95-a160-4d909f8f3d00","Type":"ContainerStarted","Data":"e1348a7610dccfeaa57d8b02376f05a74c930939155e4532ee28fc0558f42dfa"} Oct 02 02:04:24 crc kubenswrapper[4885]: I1002 02:04:24.058999 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d091410b-e1bf-423e-929f-18d2a9c65891" path="/var/lib/kubelet/pods/d091410b-e1bf-423e-929f-18d2a9c65891/volumes" Oct 02 02:04:24 crc kubenswrapper[4885]: I1002 02:04:24.079642 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 02:04:24 crc kubenswrapper[4885]: W1002 02:04:24.101646 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod345ef025_b8fb_4955_baec_9af9be17792e.slice/crio-99a5bdbb40f5c2e4ebc7ec16830e9ee8b017533039be1ad5e43e4d261465daf7 WatchSource:0}: Error finding container 99a5bdbb40f5c2e4ebc7ec16830e9ee8b017533039be1ad5e43e4d261465daf7: Status 404 returned error can't find the container with id 99a5bdbb40f5c2e4ebc7ec16830e9ee8b017533039be1ad5e43e4d261465daf7 Oct 02 02:04:24 crc kubenswrapper[4885]: I1002 02:04:24.997469 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-59f7bf489d-7tfk4" event={"ID":"5f5ca9c7-0cfd-4c95-a160-4d909f8f3d00","Type":"ContainerStarted","Data":"adb28d1d71999f8a8cf4f8b68f5ad9a168551cebff70ed2d0569c5cdfeef7a53"} Oct 02 02:04:24 crc kubenswrapper[4885]: I1002 02:04:24.998881 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-59f7bf489d-7tfk4" Oct 02 02:04:25 crc kubenswrapper[4885]: I1002 02:04:25.021026 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-tjcz4" event={"ID":"9c66bb6a-5a3c-4223-aba6-c5d9d0ea83ba","Type":"ContainerStarted","Data":"521d56adf2bb767b794788e9e4a1227efa64a91af1a488857a83713bc3a46de9"} Oct 02 02:04:25 crc kubenswrapper[4885]: I1002 02:04:25.021489 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-59f7bf489d-7tfk4" podStartSLOduration=3.021473668 podStartE2EDuration="3.021473668s" podCreationTimestamp="2025-10-02 02:04:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:04:25.016857337 +0000 UTC m=+1053.828604746" watchObservedRunningTime="2025-10-02 02:04:25.021473668 +0000 UTC m=+1053.833221067" Oct 02 02:04:25 crc kubenswrapper[4885]: I1002 02:04:25.033136 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1aff4106-2f4b-48ef-93a1-d51285ce6242","Type":"ContainerStarted","Data":"a9cf43c19d008a62e4225357cfd4c6ca237f24f5da97101bd39435507af05ae8"} Oct 02 02:04:25 crc kubenswrapper[4885]: I1002 02:04:25.041921 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-gp9st" event={"ID":"0feba5bf-20c5-4578-9cfc-f37d6511c2d6","Type":"ContainerStarted","Data":"0e6ab596da7147218797e106083b8cfd31507cec30d5f1645bcc1a5f69320aac"} Oct 02 02:04:25 crc kubenswrapper[4885]: I1002 02:04:25.058359 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"345ef025-b8fb-4955-baec-9af9be17792e","Type":"ContainerStarted","Data":"0a4f01b0e93fc3d8b30b56a1c50f9a5b6be0c884595ab22522466917a5ed9451"} Oct 02 02:04:25 crc kubenswrapper[4885]: I1002 02:04:25.058398 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"345ef025-b8fb-4955-baec-9af9be17792e","Type":"ContainerStarted","Data":"99a5bdbb40f5c2e4ebc7ec16830e9ee8b017533039be1ad5e43e4d261465daf7"} Oct 02 02:04:25 crc kubenswrapper[4885]: I1002 02:04:25.060751 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-tjcz4" podStartSLOduration=3.749181357 podStartE2EDuration="47.060732357s" podCreationTimestamp="2025-10-02 02:03:38 +0000 UTC" firstStartedPulling="2025-10-02 02:03:39.704629454 +0000 UTC m=+1008.516376843" lastFinishedPulling="2025-10-02 02:04:23.016180444 +0000 UTC m=+1051.827927843" observedRunningTime="2025-10-02 02:04:25.042620378 +0000 UTC m=+1053.854367777" watchObservedRunningTime="2025-10-02 02:04:25.060732357 +0000 UTC m=+1053.872479756" Oct 02 02:04:25 crc kubenswrapper[4885]: I1002 02:04:25.068423 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-gp9st" podStartSLOduration=3.318180909 podStartE2EDuration="47.06840709s" podCreationTimestamp="2025-10-02 02:03:38 +0000 UTC" firstStartedPulling="2025-10-02 02:03:39.871254872 +0000 UTC m=+1008.683002271" lastFinishedPulling="2025-10-02 02:04:23.621481053 +0000 UTC m=+1052.433228452" observedRunningTime="2025-10-02 02:04:25.058585993 +0000 UTC m=+1053.870333382" watchObservedRunningTime="2025-10-02 02:04:25.06840709 +0000 UTC m=+1053.880154489" Oct 02 02:04:25 crc kubenswrapper[4885]: I1002 02:04:25.076598 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.076582888 podStartE2EDuration="4.076582888s" podCreationTimestamp="2025-10-02 02:04:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:04:25.072804283 +0000 UTC m=+1053.884551682" watchObservedRunningTime="2025-10-02 02:04:25.076582888 +0000 UTC m=+1053.888330287" Oct 02 02:04:26 crc kubenswrapper[4885]: I1002 02:04:26.072241 4885 generic.go:334] "Generic (PLEG): container finished" podID="c036a628-a241-45ea-bd2a-5d6d50804f61" containerID="1a1f0197c0399a742a3103b0db978338c9e66ae73dd5516c960bfa34cfe0a60a" exitCode=0 Oct 02 02:04:26 crc kubenswrapper[4885]: I1002 02:04:26.072461 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-xf24l" event={"ID":"c036a628-a241-45ea-bd2a-5d6d50804f61","Type":"ContainerDied","Data":"1a1f0197c0399a742a3103b0db978338c9e66ae73dd5516c960bfa34cfe0a60a"} Oct 02 02:04:27 crc kubenswrapper[4885]: I1002 02:04:27.083791 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"345ef025-b8fb-4955-baec-9af9be17792e","Type":"ContainerStarted","Data":"c2dd3394b4b25538f87dffc1aa40df5a3104efa6e9908de25784a09ee89b7f7b"} Oct 02 02:04:27 crc kubenswrapper[4885]: I1002 02:04:27.131449 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=5.131425493 podStartE2EDuration="5.131425493s" podCreationTimestamp="2025-10-02 02:04:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:04:27.117770119 +0000 UTC m=+1055.929517538" watchObservedRunningTime="2025-10-02 02:04:27.131425493 +0000 UTC m=+1055.943172912" Oct 02 02:04:27 crc kubenswrapper[4885]: I1002 02:04:27.986050 4885 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7c49b4cffb-brdz9" podUID="b0fddcb3-3649-48f4-aa8f-85dd1c8137cf" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.146:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.146:8443: connect: connection refused" Oct 02 02:04:28 crc kubenswrapper[4885]: I1002 02:04:28.085899 4885 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-6d7c47744-vjcbs" podUID="607de89d-b68e-49e8-beb4-7664a37c6105" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.147:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.147:8443: connect: connection refused" Oct 02 02:04:31 crc kubenswrapper[4885]: I1002 02:04:31.991454 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-xf24l" Oct 02 02:04:32 crc kubenswrapper[4885]: I1002 02:04:32.084854 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/c036a628-a241-45ea-bd2a-5d6d50804f61-config\") pod \"c036a628-a241-45ea-bd2a-5d6d50804f61\" (UID: \"c036a628-a241-45ea-bd2a-5d6d50804f61\") " Oct 02 02:04:32 crc kubenswrapper[4885]: I1002 02:04:32.084978 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c036a628-a241-45ea-bd2a-5d6d50804f61-combined-ca-bundle\") pod \"c036a628-a241-45ea-bd2a-5d6d50804f61\" (UID: \"c036a628-a241-45ea-bd2a-5d6d50804f61\") " Oct 02 02:04:32 crc kubenswrapper[4885]: I1002 02:04:32.085046 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4jkb4\" (UniqueName: \"kubernetes.io/projected/c036a628-a241-45ea-bd2a-5d6d50804f61-kube-api-access-4jkb4\") pod \"c036a628-a241-45ea-bd2a-5d6d50804f61\" (UID: \"c036a628-a241-45ea-bd2a-5d6d50804f61\") " Oct 02 02:04:32 crc kubenswrapper[4885]: I1002 02:04:32.090876 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c036a628-a241-45ea-bd2a-5d6d50804f61-kube-api-access-4jkb4" (OuterVolumeSpecName: "kube-api-access-4jkb4") pod "c036a628-a241-45ea-bd2a-5d6d50804f61" (UID: "c036a628-a241-45ea-bd2a-5d6d50804f61"). InnerVolumeSpecName "kube-api-access-4jkb4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:04:32 crc kubenswrapper[4885]: I1002 02:04:32.109545 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c036a628-a241-45ea-bd2a-5d6d50804f61-config" (OuterVolumeSpecName: "config") pod "c036a628-a241-45ea-bd2a-5d6d50804f61" (UID: "c036a628-a241-45ea-bd2a-5d6d50804f61"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:04:32 crc kubenswrapper[4885]: I1002 02:04:32.111782 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c036a628-a241-45ea-bd2a-5d6d50804f61-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c036a628-a241-45ea-bd2a-5d6d50804f61" (UID: "c036a628-a241-45ea-bd2a-5d6d50804f61"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:04:32 crc kubenswrapper[4885]: I1002 02:04:32.130145 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-xf24l" event={"ID":"c036a628-a241-45ea-bd2a-5d6d50804f61","Type":"ContainerDied","Data":"a90523b46bfc0032de65212bd4351879dd04017043fc5b44bd795c65402bec04"} Oct 02 02:04:32 crc kubenswrapper[4885]: I1002 02:04:32.130179 4885 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a90523b46bfc0032de65212bd4351879dd04017043fc5b44bd795c65402bec04" Oct 02 02:04:32 crc kubenswrapper[4885]: I1002 02:04:32.130185 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-xf24l" Oct 02 02:04:32 crc kubenswrapper[4885]: I1002 02:04:32.191776 4885 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c036a628-a241-45ea-bd2a-5d6d50804f61-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:04:32 crc kubenswrapper[4885]: I1002 02:04:32.191803 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4jkb4\" (UniqueName: \"kubernetes.io/projected/c036a628-a241-45ea-bd2a-5d6d50804f61-kube-api-access-4jkb4\") on node \"crc\" DevicePath \"\"" Oct 02 02:04:32 crc kubenswrapper[4885]: I1002 02:04:32.191815 4885 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/c036a628-a241-45ea-bd2a-5d6d50804f61-config\") on node \"crc\" DevicePath \"\"" Oct 02 02:04:32 crc kubenswrapper[4885]: I1002 02:04:32.223110 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 02 02:04:32 crc kubenswrapper[4885]: I1002 02:04:32.223143 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 02 02:04:32 crc kubenswrapper[4885]: I1002 02:04:32.252792 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 02 02:04:32 crc kubenswrapper[4885]: I1002 02:04:32.295558 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 02 02:04:33 crc kubenswrapper[4885]: I1002 02:04:33.140504 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 02 02:04:33 crc kubenswrapper[4885]: I1002 02:04:33.140935 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 02 02:04:33 crc kubenswrapper[4885]: I1002 02:04:33.305049 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-h9tlf"] Oct 02 02:04:33 crc kubenswrapper[4885]: E1002 02:04:33.305538 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c036a628-a241-45ea-bd2a-5d6d50804f61" containerName="neutron-db-sync" Oct 02 02:04:33 crc kubenswrapper[4885]: I1002 02:04:33.305556 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="c036a628-a241-45ea-bd2a-5d6d50804f61" containerName="neutron-db-sync" Oct 02 02:04:33 crc kubenswrapper[4885]: I1002 02:04:33.305723 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="c036a628-a241-45ea-bd2a-5d6d50804f61" containerName="neutron-db-sync" Oct 02 02:04:33 crc kubenswrapper[4885]: I1002 02:04:33.306632 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84b966f6c9-h9tlf" Oct 02 02:04:33 crc kubenswrapper[4885]: I1002 02:04:33.319302 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-h9tlf"] Oct 02 02:04:33 crc kubenswrapper[4885]: I1002 02:04:33.394418 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-9c5d5c444-vkjrr"] Oct 02 02:04:33 crc kubenswrapper[4885]: I1002 02:04:33.399828 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 02 02:04:33 crc kubenswrapper[4885]: I1002 02:04:33.399865 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 02 02:04:33 crc kubenswrapper[4885]: I1002 02:04:33.400338 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-9c5d5c444-vkjrr" Oct 02 02:04:33 crc kubenswrapper[4885]: I1002 02:04:33.409502 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 02 02:04:33 crc kubenswrapper[4885]: I1002 02:04:33.411344 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Oct 02 02:04:33 crc kubenswrapper[4885]: I1002 02:04:33.411589 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 02 02:04:33 crc kubenswrapper[4885]: I1002 02:04:33.411741 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-ft6t6" Oct 02 02:04:33 crc kubenswrapper[4885]: I1002 02:04:33.417870 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/00531889-6c5d-444b-bfeb-4bb628d7ef94-dns-swift-storage-0\") pod \"dnsmasq-dns-84b966f6c9-h9tlf\" (UID: \"00531889-6c5d-444b-bfeb-4bb628d7ef94\") " pod="openstack/dnsmasq-dns-84b966f6c9-h9tlf" Oct 02 02:04:33 crc kubenswrapper[4885]: I1002 02:04:33.417909 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v5nt5\" (UniqueName: \"kubernetes.io/projected/00531889-6c5d-444b-bfeb-4bb628d7ef94-kube-api-access-v5nt5\") pod \"dnsmasq-dns-84b966f6c9-h9tlf\" (UID: \"00531889-6c5d-444b-bfeb-4bb628d7ef94\") " pod="openstack/dnsmasq-dns-84b966f6c9-h9tlf" Oct 02 02:04:33 crc kubenswrapper[4885]: I1002 02:04:33.417991 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00531889-6c5d-444b-bfeb-4bb628d7ef94-config\") pod \"dnsmasq-dns-84b966f6c9-h9tlf\" (UID: \"00531889-6c5d-444b-bfeb-4bb628d7ef94\") " pod="openstack/dnsmasq-dns-84b966f6c9-h9tlf" Oct 02 02:04:33 crc kubenswrapper[4885]: I1002 02:04:33.418013 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/00531889-6c5d-444b-bfeb-4bb628d7ef94-ovsdbserver-nb\") pod \"dnsmasq-dns-84b966f6c9-h9tlf\" (UID: \"00531889-6c5d-444b-bfeb-4bb628d7ef94\") " pod="openstack/dnsmasq-dns-84b966f6c9-h9tlf" Oct 02 02:04:33 crc kubenswrapper[4885]: I1002 02:04:33.418053 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/00531889-6c5d-444b-bfeb-4bb628d7ef94-ovsdbserver-sb\") pod \"dnsmasq-dns-84b966f6c9-h9tlf\" (UID: \"00531889-6c5d-444b-bfeb-4bb628d7ef94\") " pod="openstack/dnsmasq-dns-84b966f6c9-h9tlf" Oct 02 02:04:33 crc kubenswrapper[4885]: I1002 02:04:33.418094 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/00531889-6c5d-444b-bfeb-4bb628d7ef94-dns-svc\") pod \"dnsmasq-dns-84b966f6c9-h9tlf\" (UID: \"00531889-6c5d-444b-bfeb-4bb628d7ef94\") " pod="openstack/dnsmasq-dns-84b966f6c9-h9tlf" Oct 02 02:04:33 crc kubenswrapper[4885]: I1002 02:04:33.433177 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-9c5d5c444-vkjrr"] Oct 02 02:04:33 crc kubenswrapper[4885]: I1002 02:04:33.455377 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 02 02:04:33 crc kubenswrapper[4885]: I1002 02:04:33.457862 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 02 02:04:33 crc kubenswrapper[4885]: I1002 02:04:33.519144 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/887b9c10-e0ae-41b2-bfc3-a329f6013d98-config\") pod \"neutron-9c5d5c444-vkjrr\" (UID: \"887b9c10-e0ae-41b2-bfc3-a329f6013d98\") " pod="openstack/neutron-9c5d5c444-vkjrr" Oct 02 02:04:33 crc kubenswrapper[4885]: I1002 02:04:33.519186 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/887b9c10-e0ae-41b2-bfc3-a329f6013d98-httpd-config\") pod \"neutron-9c5d5c444-vkjrr\" (UID: \"887b9c10-e0ae-41b2-bfc3-a329f6013d98\") " pod="openstack/neutron-9c5d5c444-vkjrr" Oct 02 02:04:33 crc kubenswrapper[4885]: I1002 02:04:33.519239 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00531889-6c5d-444b-bfeb-4bb628d7ef94-config\") pod \"dnsmasq-dns-84b966f6c9-h9tlf\" (UID: \"00531889-6c5d-444b-bfeb-4bb628d7ef94\") " pod="openstack/dnsmasq-dns-84b966f6c9-h9tlf" Oct 02 02:04:33 crc kubenswrapper[4885]: I1002 02:04:33.519274 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/00531889-6c5d-444b-bfeb-4bb628d7ef94-ovsdbserver-nb\") pod \"dnsmasq-dns-84b966f6c9-h9tlf\" (UID: \"00531889-6c5d-444b-bfeb-4bb628d7ef94\") " pod="openstack/dnsmasq-dns-84b966f6c9-h9tlf" Oct 02 02:04:33 crc kubenswrapper[4885]: I1002 02:04:33.519325 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/00531889-6c5d-444b-bfeb-4bb628d7ef94-ovsdbserver-sb\") pod \"dnsmasq-dns-84b966f6c9-h9tlf\" (UID: \"00531889-6c5d-444b-bfeb-4bb628d7ef94\") " pod="openstack/dnsmasq-dns-84b966f6c9-h9tlf" Oct 02 02:04:33 crc kubenswrapper[4885]: I1002 02:04:33.519345 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/887b9c10-e0ae-41b2-bfc3-a329f6013d98-ovndb-tls-certs\") pod \"neutron-9c5d5c444-vkjrr\" (UID: \"887b9c10-e0ae-41b2-bfc3-a329f6013d98\") " pod="openstack/neutron-9c5d5c444-vkjrr" Oct 02 02:04:33 crc kubenswrapper[4885]: I1002 02:04:33.519398 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/00531889-6c5d-444b-bfeb-4bb628d7ef94-dns-svc\") pod \"dnsmasq-dns-84b966f6c9-h9tlf\" (UID: \"00531889-6c5d-444b-bfeb-4bb628d7ef94\") " pod="openstack/dnsmasq-dns-84b966f6c9-h9tlf" Oct 02 02:04:33 crc kubenswrapper[4885]: I1002 02:04:33.519433 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sh2v7\" (UniqueName: \"kubernetes.io/projected/887b9c10-e0ae-41b2-bfc3-a329f6013d98-kube-api-access-sh2v7\") pod \"neutron-9c5d5c444-vkjrr\" (UID: \"887b9c10-e0ae-41b2-bfc3-a329f6013d98\") " pod="openstack/neutron-9c5d5c444-vkjrr" Oct 02 02:04:33 crc kubenswrapper[4885]: I1002 02:04:33.519455 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/00531889-6c5d-444b-bfeb-4bb628d7ef94-dns-swift-storage-0\") pod \"dnsmasq-dns-84b966f6c9-h9tlf\" (UID: \"00531889-6c5d-444b-bfeb-4bb628d7ef94\") " pod="openstack/dnsmasq-dns-84b966f6c9-h9tlf" Oct 02 02:04:33 crc kubenswrapper[4885]: I1002 02:04:33.519472 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v5nt5\" (UniqueName: \"kubernetes.io/projected/00531889-6c5d-444b-bfeb-4bb628d7ef94-kube-api-access-v5nt5\") pod \"dnsmasq-dns-84b966f6c9-h9tlf\" (UID: \"00531889-6c5d-444b-bfeb-4bb628d7ef94\") " pod="openstack/dnsmasq-dns-84b966f6c9-h9tlf" Oct 02 02:04:33 crc kubenswrapper[4885]: I1002 02:04:33.519494 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/887b9c10-e0ae-41b2-bfc3-a329f6013d98-combined-ca-bundle\") pod \"neutron-9c5d5c444-vkjrr\" (UID: \"887b9c10-e0ae-41b2-bfc3-a329f6013d98\") " pod="openstack/neutron-9c5d5c444-vkjrr" Oct 02 02:04:33 crc kubenswrapper[4885]: I1002 02:04:33.520495 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00531889-6c5d-444b-bfeb-4bb628d7ef94-config\") pod \"dnsmasq-dns-84b966f6c9-h9tlf\" (UID: \"00531889-6c5d-444b-bfeb-4bb628d7ef94\") " pod="openstack/dnsmasq-dns-84b966f6c9-h9tlf" Oct 02 02:04:33 crc kubenswrapper[4885]: I1002 02:04:33.523985 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/00531889-6c5d-444b-bfeb-4bb628d7ef94-dns-svc\") pod \"dnsmasq-dns-84b966f6c9-h9tlf\" (UID: \"00531889-6c5d-444b-bfeb-4bb628d7ef94\") " pod="openstack/dnsmasq-dns-84b966f6c9-h9tlf" Oct 02 02:04:33 crc kubenswrapper[4885]: I1002 02:04:33.524207 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/00531889-6c5d-444b-bfeb-4bb628d7ef94-dns-swift-storage-0\") pod \"dnsmasq-dns-84b966f6c9-h9tlf\" (UID: \"00531889-6c5d-444b-bfeb-4bb628d7ef94\") " pod="openstack/dnsmasq-dns-84b966f6c9-h9tlf" Oct 02 02:04:33 crc kubenswrapper[4885]: I1002 02:04:33.524707 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/00531889-6c5d-444b-bfeb-4bb628d7ef94-ovsdbserver-sb\") pod \"dnsmasq-dns-84b966f6c9-h9tlf\" (UID: \"00531889-6c5d-444b-bfeb-4bb628d7ef94\") " pod="openstack/dnsmasq-dns-84b966f6c9-h9tlf" Oct 02 02:04:33 crc kubenswrapper[4885]: I1002 02:04:33.524966 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/00531889-6c5d-444b-bfeb-4bb628d7ef94-ovsdbserver-nb\") pod \"dnsmasq-dns-84b966f6c9-h9tlf\" (UID: \"00531889-6c5d-444b-bfeb-4bb628d7ef94\") " pod="openstack/dnsmasq-dns-84b966f6c9-h9tlf" Oct 02 02:04:33 crc kubenswrapper[4885]: I1002 02:04:33.537959 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v5nt5\" (UniqueName: \"kubernetes.io/projected/00531889-6c5d-444b-bfeb-4bb628d7ef94-kube-api-access-v5nt5\") pod \"dnsmasq-dns-84b966f6c9-h9tlf\" (UID: \"00531889-6c5d-444b-bfeb-4bb628d7ef94\") " pod="openstack/dnsmasq-dns-84b966f6c9-h9tlf" Oct 02 02:04:33 crc kubenswrapper[4885]: I1002 02:04:33.620658 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/887b9c10-e0ae-41b2-bfc3-a329f6013d98-ovndb-tls-certs\") pod \"neutron-9c5d5c444-vkjrr\" (UID: \"887b9c10-e0ae-41b2-bfc3-a329f6013d98\") " pod="openstack/neutron-9c5d5c444-vkjrr" Oct 02 02:04:33 crc kubenswrapper[4885]: I1002 02:04:33.620752 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sh2v7\" (UniqueName: \"kubernetes.io/projected/887b9c10-e0ae-41b2-bfc3-a329f6013d98-kube-api-access-sh2v7\") pod \"neutron-9c5d5c444-vkjrr\" (UID: \"887b9c10-e0ae-41b2-bfc3-a329f6013d98\") " pod="openstack/neutron-9c5d5c444-vkjrr" Oct 02 02:04:33 crc kubenswrapper[4885]: I1002 02:04:33.620787 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/887b9c10-e0ae-41b2-bfc3-a329f6013d98-combined-ca-bundle\") pod \"neutron-9c5d5c444-vkjrr\" (UID: \"887b9c10-e0ae-41b2-bfc3-a329f6013d98\") " pod="openstack/neutron-9c5d5c444-vkjrr" Oct 02 02:04:33 crc kubenswrapper[4885]: I1002 02:04:33.620835 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/887b9c10-e0ae-41b2-bfc3-a329f6013d98-config\") pod \"neutron-9c5d5c444-vkjrr\" (UID: \"887b9c10-e0ae-41b2-bfc3-a329f6013d98\") " pod="openstack/neutron-9c5d5c444-vkjrr" Oct 02 02:04:33 crc kubenswrapper[4885]: I1002 02:04:33.620851 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/887b9c10-e0ae-41b2-bfc3-a329f6013d98-httpd-config\") pod \"neutron-9c5d5c444-vkjrr\" (UID: \"887b9c10-e0ae-41b2-bfc3-a329f6013d98\") " pod="openstack/neutron-9c5d5c444-vkjrr" Oct 02 02:04:33 crc kubenswrapper[4885]: I1002 02:04:33.624719 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/887b9c10-e0ae-41b2-bfc3-a329f6013d98-httpd-config\") pod \"neutron-9c5d5c444-vkjrr\" (UID: \"887b9c10-e0ae-41b2-bfc3-a329f6013d98\") " pod="openstack/neutron-9c5d5c444-vkjrr" Oct 02 02:04:33 crc kubenswrapper[4885]: I1002 02:04:33.626316 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/887b9c10-e0ae-41b2-bfc3-a329f6013d98-config\") pod \"neutron-9c5d5c444-vkjrr\" (UID: \"887b9c10-e0ae-41b2-bfc3-a329f6013d98\") " pod="openstack/neutron-9c5d5c444-vkjrr" Oct 02 02:04:33 crc kubenswrapper[4885]: I1002 02:04:33.627250 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/887b9c10-e0ae-41b2-bfc3-a329f6013d98-combined-ca-bundle\") pod \"neutron-9c5d5c444-vkjrr\" (UID: \"887b9c10-e0ae-41b2-bfc3-a329f6013d98\") " pod="openstack/neutron-9c5d5c444-vkjrr" Oct 02 02:04:33 crc kubenswrapper[4885]: I1002 02:04:33.629882 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/887b9c10-e0ae-41b2-bfc3-a329f6013d98-ovndb-tls-certs\") pod \"neutron-9c5d5c444-vkjrr\" (UID: \"887b9c10-e0ae-41b2-bfc3-a329f6013d98\") " pod="openstack/neutron-9c5d5c444-vkjrr" Oct 02 02:04:33 crc kubenswrapper[4885]: I1002 02:04:33.632425 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84b966f6c9-h9tlf" Oct 02 02:04:33 crc kubenswrapper[4885]: I1002 02:04:33.649364 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sh2v7\" (UniqueName: \"kubernetes.io/projected/887b9c10-e0ae-41b2-bfc3-a329f6013d98-kube-api-access-sh2v7\") pod \"neutron-9c5d5c444-vkjrr\" (UID: \"887b9c10-e0ae-41b2-bfc3-a329f6013d98\") " pod="openstack/neutron-9c5d5c444-vkjrr" Oct 02 02:04:33 crc kubenswrapper[4885]: I1002 02:04:33.732298 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-9c5d5c444-vkjrr" Oct 02 02:04:34 crc kubenswrapper[4885]: E1002 02:04:34.023729 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ceilometer-0" podUID="9226a04e-0195-42a1-b651-2f5379b1e2a9" Oct 02 02:04:34 crc kubenswrapper[4885]: I1002 02:04:34.153099 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9226a04e-0195-42a1-b651-2f5379b1e2a9","Type":"ContainerStarted","Data":"f0fbd443af35d475dd0d4bf7f4e6dcb68f738868a512f56d411ef9996a3aadca"} Oct 02 02:04:34 crc kubenswrapper[4885]: I1002 02:04:34.153381 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9226a04e-0195-42a1-b651-2f5379b1e2a9" containerName="ceilometer-notification-agent" containerID="cri-o://bc5a69f8f4d26b9f7c36647a448c3943fedcadcf35e24b1c46b0c1cb51a86fef" gracePeriod=30 Oct 02 02:04:34 crc kubenswrapper[4885]: I1002 02:04:34.153595 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9226a04e-0195-42a1-b651-2f5379b1e2a9" containerName="sg-core" containerID="cri-o://cc2a1cdc982003a22e30118b6756d23620f619fe3029c82308eb0d2d8ccbcfa2" gracePeriod=30 Oct 02 02:04:34 crc kubenswrapper[4885]: I1002 02:04:34.153655 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9226a04e-0195-42a1-b651-2f5379b1e2a9" containerName="proxy-httpd" containerID="cri-o://f0fbd443af35d475dd0d4bf7f4e6dcb68f738868a512f56d411ef9996a3aadca" gracePeriod=30 Oct 02 02:04:34 crc kubenswrapper[4885]: I1002 02:04:34.153816 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 02 02:04:34 crc kubenswrapper[4885]: I1002 02:04:34.153841 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 02 02:04:34 crc kubenswrapper[4885]: I1002 02:04:34.231498 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-h9tlf"] Oct 02 02:04:34 crc kubenswrapper[4885]: W1002 02:04:34.232270 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod00531889_6c5d_444b_bfeb_4bb628d7ef94.slice/crio-0e0d258454fe494f02c5999bf34c7c24e261eaab00218f001ef3b7cc432d49c3 WatchSource:0}: Error finding container 0e0d258454fe494f02c5999bf34c7c24e261eaab00218f001ef3b7cc432d49c3: Status 404 returned error can't find the container with id 0e0d258454fe494f02c5999bf34c7c24e261eaab00218f001ef3b7cc432d49c3 Oct 02 02:04:34 crc kubenswrapper[4885]: I1002 02:04:34.394311 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-9c5d5c444-vkjrr"] Oct 02 02:04:35 crc kubenswrapper[4885]: I1002 02:04:35.163879 4885 generic.go:334] "Generic (PLEG): container finished" podID="9226a04e-0195-42a1-b651-2f5379b1e2a9" containerID="f0fbd443af35d475dd0d4bf7f4e6dcb68f738868a512f56d411ef9996a3aadca" exitCode=0 Oct 02 02:04:35 crc kubenswrapper[4885]: I1002 02:04:35.164170 4885 generic.go:334] "Generic (PLEG): container finished" podID="9226a04e-0195-42a1-b651-2f5379b1e2a9" containerID="cc2a1cdc982003a22e30118b6756d23620f619fe3029c82308eb0d2d8ccbcfa2" exitCode=2 Oct 02 02:04:35 crc kubenswrapper[4885]: I1002 02:04:35.163936 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9226a04e-0195-42a1-b651-2f5379b1e2a9","Type":"ContainerDied","Data":"f0fbd443af35d475dd0d4bf7f4e6dcb68f738868a512f56d411ef9996a3aadca"} Oct 02 02:04:35 crc kubenswrapper[4885]: I1002 02:04:35.164234 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9226a04e-0195-42a1-b651-2f5379b1e2a9","Type":"ContainerDied","Data":"cc2a1cdc982003a22e30118b6756d23620f619fe3029c82308eb0d2d8ccbcfa2"} Oct 02 02:04:35 crc kubenswrapper[4885]: I1002 02:04:35.165899 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-9c5d5c444-vkjrr" event={"ID":"887b9c10-e0ae-41b2-bfc3-a329f6013d98","Type":"ContainerStarted","Data":"772e2dc86bdea03785060468cdc4451b409b9c8bea7dad586dca7ef5e11995a4"} Oct 02 02:04:35 crc kubenswrapper[4885]: I1002 02:04:35.165930 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-9c5d5c444-vkjrr" event={"ID":"887b9c10-e0ae-41b2-bfc3-a329f6013d98","Type":"ContainerStarted","Data":"01ca921473539e0109a018c1729a703dbf2f939cc7f6e4997e2af653d9fc5def"} Oct 02 02:04:35 crc kubenswrapper[4885]: I1002 02:04:35.165944 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-9c5d5c444-vkjrr" event={"ID":"887b9c10-e0ae-41b2-bfc3-a329f6013d98","Type":"ContainerStarted","Data":"0b5865dc809e8e154588c9a9e4cdf2257e724e2d55495dd83fa32e279c2e4fc8"} Oct 02 02:04:35 crc kubenswrapper[4885]: I1002 02:04:35.166053 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-9c5d5c444-vkjrr" Oct 02 02:04:35 crc kubenswrapper[4885]: I1002 02:04:35.167369 4885 generic.go:334] "Generic (PLEG): container finished" podID="00531889-6c5d-444b-bfeb-4bb628d7ef94" containerID="8dc74af4b51491aba9029712d26b660cffe33813b769beea80ccb9b654f48b56" exitCode=0 Oct 02 02:04:35 crc kubenswrapper[4885]: I1002 02:04:35.167432 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b966f6c9-h9tlf" event={"ID":"00531889-6c5d-444b-bfeb-4bb628d7ef94","Type":"ContainerDied","Data":"8dc74af4b51491aba9029712d26b660cffe33813b769beea80ccb9b654f48b56"} Oct 02 02:04:35 crc kubenswrapper[4885]: I1002 02:04:35.167466 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b966f6c9-h9tlf" event={"ID":"00531889-6c5d-444b-bfeb-4bb628d7ef94","Type":"ContainerStarted","Data":"0e0d258454fe494f02c5999bf34c7c24e261eaab00218f001ef3b7cc432d49c3"} Oct 02 02:04:35 crc kubenswrapper[4885]: I1002 02:04:35.193970 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-9c5d5c444-vkjrr" podStartSLOduration=2.193954611 podStartE2EDuration="2.193954611s" podCreationTimestamp="2025-10-02 02:04:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:04:35.189858657 +0000 UTC m=+1064.001606056" watchObservedRunningTime="2025-10-02 02:04:35.193954611 +0000 UTC m=+1064.005702010" Oct 02 02:04:35 crc kubenswrapper[4885]: I1002 02:04:35.324428 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 02 02:04:35 crc kubenswrapper[4885]: I1002 02:04:35.324763 4885 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 02 02:04:35 crc kubenswrapper[4885]: I1002 02:04:35.399784 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 02 02:04:35 crc kubenswrapper[4885]: I1002 02:04:35.442914 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-684bd7f7b9-p7lvf"] Oct 02 02:04:35 crc kubenswrapper[4885]: I1002 02:04:35.444795 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-684bd7f7b9-p7lvf" Oct 02 02:04:35 crc kubenswrapper[4885]: I1002 02:04:35.447503 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Oct 02 02:04:35 crc kubenswrapper[4885]: I1002 02:04:35.448584 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Oct 02 02:04:35 crc kubenswrapper[4885]: I1002 02:04:35.477512 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-684bd7f7b9-p7lvf"] Oct 02 02:04:35 crc kubenswrapper[4885]: I1002 02:04:35.562148 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ddb14564-1e3d-43d7-9c9d-2665b53124b6-ovndb-tls-certs\") pod \"neutron-684bd7f7b9-p7lvf\" (UID: \"ddb14564-1e3d-43d7-9c9d-2665b53124b6\") " pod="openstack/neutron-684bd7f7b9-p7lvf" Oct 02 02:04:35 crc kubenswrapper[4885]: I1002 02:04:35.562187 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ddb14564-1e3d-43d7-9c9d-2665b53124b6-internal-tls-certs\") pod \"neutron-684bd7f7b9-p7lvf\" (UID: \"ddb14564-1e3d-43d7-9c9d-2665b53124b6\") " pod="openstack/neutron-684bd7f7b9-p7lvf" Oct 02 02:04:35 crc kubenswrapper[4885]: I1002 02:04:35.562208 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4828b\" (UniqueName: \"kubernetes.io/projected/ddb14564-1e3d-43d7-9c9d-2665b53124b6-kube-api-access-4828b\") pod \"neutron-684bd7f7b9-p7lvf\" (UID: \"ddb14564-1e3d-43d7-9c9d-2665b53124b6\") " pod="openstack/neutron-684bd7f7b9-p7lvf" Oct 02 02:04:35 crc kubenswrapper[4885]: I1002 02:04:35.562270 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/ddb14564-1e3d-43d7-9c9d-2665b53124b6-httpd-config\") pod \"neutron-684bd7f7b9-p7lvf\" (UID: \"ddb14564-1e3d-43d7-9c9d-2665b53124b6\") " pod="openstack/neutron-684bd7f7b9-p7lvf" Oct 02 02:04:35 crc kubenswrapper[4885]: I1002 02:04:35.562307 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddb14564-1e3d-43d7-9c9d-2665b53124b6-combined-ca-bundle\") pod \"neutron-684bd7f7b9-p7lvf\" (UID: \"ddb14564-1e3d-43d7-9c9d-2665b53124b6\") " pod="openstack/neutron-684bd7f7b9-p7lvf" Oct 02 02:04:35 crc kubenswrapper[4885]: I1002 02:04:35.562321 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/ddb14564-1e3d-43d7-9c9d-2665b53124b6-config\") pod \"neutron-684bd7f7b9-p7lvf\" (UID: \"ddb14564-1e3d-43d7-9c9d-2665b53124b6\") " pod="openstack/neutron-684bd7f7b9-p7lvf" Oct 02 02:04:35 crc kubenswrapper[4885]: I1002 02:04:35.562339 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ddb14564-1e3d-43d7-9c9d-2665b53124b6-public-tls-certs\") pod \"neutron-684bd7f7b9-p7lvf\" (UID: \"ddb14564-1e3d-43d7-9c9d-2665b53124b6\") " pod="openstack/neutron-684bd7f7b9-p7lvf" Oct 02 02:04:35 crc kubenswrapper[4885]: I1002 02:04:35.664439 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ddb14564-1e3d-43d7-9c9d-2665b53124b6-public-tls-certs\") pod \"neutron-684bd7f7b9-p7lvf\" (UID: \"ddb14564-1e3d-43d7-9c9d-2665b53124b6\") " pod="openstack/neutron-684bd7f7b9-p7lvf" Oct 02 02:04:35 crc kubenswrapper[4885]: I1002 02:04:35.664559 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ddb14564-1e3d-43d7-9c9d-2665b53124b6-ovndb-tls-certs\") pod \"neutron-684bd7f7b9-p7lvf\" (UID: \"ddb14564-1e3d-43d7-9c9d-2665b53124b6\") " pod="openstack/neutron-684bd7f7b9-p7lvf" Oct 02 02:04:35 crc kubenswrapper[4885]: I1002 02:04:35.664586 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ddb14564-1e3d-43d7-9c9d-2665b53124b6-internal-tls-certs\") pod \"neutron-684bd7f7b9-p7lvf\" (UID: \"ddb14564-1e3d-43d7-9c9d-2665b53124b6\") " pod="openstack/neutron-684bd7f7b9-p7lvf" Oct 02 02:04:35 crc kubenswrapper[4885]: I1002 02:04:35.664607 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4828b\" (UniqueName: \"kubernetes.io/projected/ddb14564-1e3d-43d7-9c9d-2665b53124b6-kube-api-access-4828b\") pod \"neutron-684bd7f7b9-p7lvf\" (UID: \"ddb14564-1e3d-43d7-9c9d-2665b53124b6\") " pod="openstack/neutron-684bd7f7b9-p7lvf" Oct 02 02:04:35 crc kubenswrapper[4885]: I1002 02:04:35.664659 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/ddb14564-1e3d-43d7-9c9d-2665b53124b6-httpd-config\") pod \"neutron-684bd7f7b9-p7lvf\" (UID: \"ddb14564-1e3d-43d7-9c9d-2665b53124b6\") " pod="openstack/neutron-684bd7f7b9-p7lvf" Oct 02 02:04:35 crc kubenswrapper[4885]: I1002 02:04:35.664701 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddb14564-1e3d-43d7-9c9d-2665b53124b6-combined-ca-bundle\") pod \"neutron-684bd7f7b9-p7lvf\" (UID: \"ddb14564-1e3d-43d7-9c9d-2665b53124b6\") " pod="openstack/neutron-684bd7f7b9-p7lvf" Oct 02 02:04:35 crc kubenswrapper[4885]: I1002 02:04:35.664718 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/ddb14564-1e3d-43d7-9c9d-2665b53124b6-config\") pod \"neutron-684bd7f7b9-p7lvf\" (UID: \"ddb14564-1e3d-43d7-9c9d-2665b53124b6\") " pod="openstack/neutron-684bd7f7b9-p7lvf" Oct 02 02:04:35 crc kubenswrapper[4885]: I1002 02:04:35.679196 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ddb14564-1e3d-43d7-9c9d-2665b53124b6-ovndb-tls-certs\") pod \"neutron-684bd7f7b9-p7lvf\" (UID: \"ddb14564-1e3d-43d7-9c9d-2665b53124b6\") " pod="openstack/neutron-684bd7f7b9-p7lvf" Oct 02 02:04:35 crc kubenswrapper[4885]: I1002 02:04:35.679984 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ddb14564-1e3d-43d7-9c9d-2665b53124b6-public-tls-certs\") pod \"neutron-684bd7f7b9-p7lvf\" (UID: \"ddb14564-1e3d-43d7-9c9d-2665b53124b6\") " pod="openstack/neutron-684bd7f7b9-p7lvf" Oct 02 02:04:35 crc kubenswrapper[4885]: I1002 02:04:35.681435 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ddb14564-1e3d-43d7-9c9d-2665b53124b6-internal-tls-certs\") pod \"neutron-684bd7f7b9-p7lvf\" (UID: \"ddb14564-1e3d-43d7-9c9d-2665b53124b6\") " pod="openstack/neutron-684bd7f7b9-p7lvf" Oct 02 02:04:35 crc kubenswrapper[4885]: I1002 02:04:35.681445 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/ddb14564-1e3d-43d7-9c9d-2665b53124b6-httpd-config\") pod \"neutron-684bd7f7b9-p7lvf\" (UID: \"ddb14564-1e3d-43d7-9c9d-2665b53124b6\") " pod="openstack/neutron-684bd7f7b9-p7lvf" Oct 02 02:04:35 crc kubenswrapper[4885]: I1002 02:04:35.682824 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddb14564-1e3d-43d7-9c9d-2665b53124b6-combined-ca-bundle\") pod \"neutron-684bd7f7b9-p7lvf\" (UID: \"ddb14564-1e3d-43d7-9c9d-2665b53124b6\") " pod="openstack/neutron-684bd7f7b9-p7lvf" Oct 02 02:04:35 crc kubenswrapper[4885]: I1002 02:04:35.684152 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4828b\" (UniqueName: \"kubernetes.io/projected/ddb14564-1e3d-43d7-9c9d-2665b53124b6-kube-api-access-4828b\") pod \"neutron-684bd7f7b9-p7lvf\" (UID: \"ddb14564-1e3d-43d7-9c9d-2665b53124b6\") " pod="openstack/neutron-684bd7f7b9-p7lvf" Oct 02 02:04:35 crc kubenswrapper[4885]: I1002 02:04:35.686155 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/ddb14564-1e3d-43d7-9c9d-2665b53124b6-config\") pod \"neutron-684bd7f7b9-p7lvf\" (UID: \"ddb14564-1e3d-43d7-9c9d-2665b53124b6\") " pod="openstack/neutron-684bd7f7b9-p7lvf" Oct 02 02:04:35 crc kubenswrapper[4885]: I1002 02:04:35.763776 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-684bd7f7b9-p7lvf" Oct 02 02:04:36 crc kubenswrapper[4885]: I1002 02:04:36.176683 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b966f6c9-h9tlf" event={"ID":"00531889-6c5d-444b-bfeb-4bb628d7ef94","Type":"ContainerStarted","Data":"712c4978547cdf1d6a36eaa4a251de9c427d874458af321b65b0c84c2922da84"} Oct 02 02:04:36 crc kubenswrapper[4885]: I1002 02:04:36.177090 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-84b966f6c9-h9tlf" Oct 02 02:04:36 crc kubenswrapper[4885]: I1002 02:04:36.201734 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 02 02:04:36 crc kubenswrapper[4885]: I1002 02:04:36.201843 4885 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 02 02:04:36 crc kubenswrapper[4885]: I1002 02:04:36.228535 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-84b966f6c9-h9tlf" podStartSLOduration=3.228520786 podStartE2EDuration="3.228520786s" podCreationTimestamp="2025-10-02 02:04:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:04:36.202413174 +0000 UTC m=+1065.014160563" watchObservedRunningTime="2025-10-02 02:04:36.228520786 +0000 UTC m=+1065.040268185" Oct 02 02:04:36 crc kubenswrapper[4885]: I1002 02:04:36.273726 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-684bd7f7b9-p7lvf"] Oct 02 02:04:36 crc kubenswrapper[4885]: W1002 02:04:36.279075 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podddb14564_1e3d_43d7_9c9d_2665b53124b6.slice/crio-ec692b5eb126dfa106a222efa3dfa7a1736eb9dfe4fe76436ad638eccc80df2a WatchSource:0}: Error finding container ec692b5eb126dfa106a222efa3dfa7a1736eb9dfe4fe76436ad638eccc80df2a: Status 404 returned error can't find the container with id ec692b5eb126dfa106a222efa3dfa7a1736eb9dfe4fe76436ad638eccc80df2a Oct 02 02:04:36 crc kubenswrapper[4885]: I1002 02:04:36.319682 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 02 02:04:37 crc kubenswrapper[4885]: I1002 02:04:37.188650 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-684bd7f7b9-p7lvf" event={"ID":"ddb14564-1e3d-43d7-9c9d-2665b53124b6","Type":"ContainerStarted","Data":"198da4d5f74ff28c1aad53fb0aa0b5bee0c14ad9b31407d2e9b6bcab2f93a43a"} Oct 02 02:04:37 crc kubenswrapper[4885]: I1002 02:04:37.189889 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-684bd7f7b9-p7lvf" event={"ID":"ddb14564-1e3d-43d7-9c9d-2665b53124b6","Type":"ContainerStarted","Data":"ec692b5eb126dfa106a222efa3dfa7a1736eb9dfe4fe76436ad638eccc80df2a"} Oct 02 02:04:38 crc kubenswrapper[4885]: I1002 02:04:38.200920 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-684bd7f7b9-p7lvf" event={"ID":"ddb14564-1e3d-43d7-9c9d-2665b53124b6","Type":"ContainerStarted","Data":"a80d9d3bd87a53ced09beda498c88b7f3827e80a6e9938f90113db6488505b48"} Oct 02 02:04:39 crc kubenswrapper[4885]: I1002 02:04:39.106763 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 02:04:39 crc kubenswrapper[4885]: I1002 02:04:39.150864 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9226a04e-0195-42a1-b651-2f5379b1e2a9-scripts\") pod \"9226a04e-0195-42a1-b651-2f5379b1e2a9\" (UID: \"9226a04e-0195-42a1-b651-2f5379b1e2a9\") " Oct 02 02:04:39 crc kubenswrapper[4885]: I1002 02:04:39.150972 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9226a04e-0195-42a1-b651-2f5379b1e2a9-config-data\") pod \"9226a04e-0195-42a1-b651-2f5379b1e2a9\" (UID: \"9226a04e-0195-42a1-b651-2f5379b1e2a9\") " Oct 02 02:04:39 crc kubenswrapper[4885]: I1002 02:04:39.151839 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9226a04e-0195-42a1-b651-2f5379b1e2a9-combined-ca-bundle\") pod \"9226a04e-0195-42a1-b651-2f5379b1e2a9\" (UID: \"9226a04e-0195-42a1-b651-2f5379b1e2a9\") " Oct 02 02:04:39 crc kubenswrapper[4885]: I1002 02:04:39.151878 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9226a04e-0195-42a1-b651-2f5379b1e2a9-log-httpd\") pod \"9226a04e-0195-42a1-b651-2f5379b1e2a9\" (UID: \"9226a04e-0195-42a1-b651-2f5379b1e2a9\") " Oct 02 02:04:39 crc kubenswrapper[4885]: I1002 02:04:39.151923 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9226a04e-0195-42a1-b651-2f5379b1e2a9-run-httpd\") pod \"9226a04e-0195-42a1-b651-2f5379b1e2a9\" (UID: \"9226a04e-0195-42a1-b651-2f5379b1e2a9\") " Oct 02 02:04:39 crc kubenswrapper[4885]: I1002 02:04:39.151971 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9226a04e-0195-42a1-b651-2f5379b1e2a9-sg-core-conf-yaml\") pod \"9226a04e-0195-42a1-b651-2f5379b1e2a9\" (UID: \"9226a04e-0195-42a1-b651-2f5379b1e2a9\") " Oct 02 02:04:39 crc kubenswrapper[4885]: I1002 02:04:39.152056 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z2mzr\" (UniqueName: \"kubernetes.io/projected/9226a04e-0195-42a1-b651-2f5379b1e2a9-kube-api-access-z2mzr\") pod \"9226a04e-0195-42a1-b651-2f5379b1e2a9\" (UID: \"9226a04e-0195-42a1-b651-2f5379b1e2a9\") " Oct 02 02:04:39 crc kubenswrapper[4885]: I1002 02:04:39.152449 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9226a04e-0195-42a1-b651-2f5379b1e2a9-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "9226a04e-0195-42a1-b651-2f5379b1e2a9" (UID: "9226a04e-0195-42a1-b651-2f5379b1e2a9"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:04:39 crc kubenswrapper[4885]: I1002 02:04:39.152536 4885 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9226a04e-0195-42a1-b651-2f5379b1e2a9-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 02:04:39 crc kubenswrapper[4885]: I1002 02:04:39.153101 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9226a04e-0195-42a1-b651-2f5379b1e2a9-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "9226a04e-0195-42a1-b651-2f5379b1e2a9" (UID: "9226a04e-0195-42a1-b651-2f5379b1e2a9"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:04:39 crc kubenswrapper[4885]: I1002 02:04:39.156436 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9226a04e-0195-42a1-b651-2f5379b1e2a9-kube-api-access-z2mzr" (OuterVolumeSpecName: "kube-api-access-z2mzr") pod "9226a04e-0195-42a1-b651-2f5379b1e2a9" (UID: "9226a04e-0195-42a1-b651-2f5379b1e2a9"). InnerVolumeSpecName "kube-api-access-z2mzr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:04:39 crc kubenswrapper[4885]: I1002 02:04:39.157106 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9226a04e-0195-42a1-b651-2f5379b1e2a9-scripts" (OuterVolumeSpecName: "scripts") pod "9226a04e-0195-42a1-b651-2f5379b1e2a9" (UID: "9226a04e-0195-42a1-b651-2f5379b1e2a9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:04:39 crc kubenswrapper[4885]: I1002 02:04:39.181576 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9226a04e-0195-42a1-b651-2f5379b1e2a9-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "9226a04e-0195-42a1-b651-2f5379b1e2a9" (UID: "9226a04e-0195-42a1-b651-2f5379b1e2a9"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:04:39 crc kubenswrapper[4885]: I1002 02:04:39.211921 4885 generic.go:334] "Generic (PLEG): container finished" podID="9226a04e-0195-42a1-b651-2f5379b1e2a9" containerID="bc5a69f8f4d26b9f7c36647a448c3943fedcadcf35e24b1c46b0c1cb51a86fef" exitCode=0 Oct 02 02:04:39 crc kubenswrapper[4885]: I1002 02:04:39.211960 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9226a04e-0195-42a1-b651-2f5379b1e2a9","Type":"ContainerDied","Data":"bc5a69f8f4d26b9f7c36647a448c3943fedcadcf35e24b1c46b0c1cb51a86fef"} Oct 02 02:04:39 crc kubenswrapper[4885]: I1002 02:04:39.212065 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 02:04:39 crc kubenswrapper[4885]: I1002 02:04:39.212119 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9226a04e-0195-42a1-b651-2f5379b1e2a9","Type":"ContainerDied","Data":"471e6c4e91ce2890981d6a2361c21e1b7a21510bbff4072c0385cb579d915bbf"} Oct 02 02:04:39 crc kubenswrapper[4885]: I1002 02:04:39.212157 4885 scope.go:117] "RemoveContainer" containerID="f0fbd443af35d475dd0d4bf7f4e6dcb68f738868a512f56d411ef9996a3aadca" Oct 02 02:04:39 crc kubenswrapper[4885]: I1002 02:04:39.212190 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-684bd7f7b9-p7lvf" Oct 02 02:04:39 crc kubenswrapper[4885]: I1002 02:04:39.216665 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9226a04e-0195-42a1-b651-2f5379b1e2a9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9226a04e-0195-42a1-b651-2f5379b1e2a9" (UID: "9226a04e-0195-42a1-b651-2f5379b1e2a9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:04:39 crc kubenswrapper[4885]: I1002 02:04:39.246277 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-684bd7f7b9-p7lvf" podStartSLOduration=4.246243643 podStartE2EDuration="4.246243643s" podCreationTimestamp="2025-10-02 02:04:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:04:39.24151591 +0000 UTC m=+1068.053263329" watchObservedRunningTime="2025-10-02 02:04:39.246243643 +0000 UTC m=+1068.057991032" Oct 02 02:04:39 crc kubenswrapper[4885]: I1002 02:04:39.251591 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9226a04e-0195-42a1-b651-2f5379b1e2a9-config-data" (OuterVolumeSpecName: "config-data") pod "9226a04e-0195-42a1-b651-2f5379b1e2a9" (UID: "9226a04e-0195-42a1-b651-2f5379b1e2a9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:04:39 crc kubenswrapper[4885]: I1002 02:04:39.254341 4885 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9226a04e-0195-42a1-b651-2f5379b1e2a9-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 02 02:04:39 crc kubenswrapper[4885]: I1002 02:04:39.254385 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z2mzr\" (UniqueName: \"kubernetes.io/projected/9226a04e-0195-42a1-b651-2f5379b1e2a9-kube-api-access-z2mzr\") on node \"crc\" DevicePath \"\"" Oct 02 02:04:39 crc kubenswrapper[4885]: I1002 02:04:39.254491 4885 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9226a04e-0195-42a1-b651-2f5379b1e2a9-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 02:04:39 crc kubenswrapper[4885]: I1002 02:04:39.255103 4885 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9226a04e-0195-42a1-b651-2f5379b1e2a9-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 02:04:39 crc kubenswrapper[4885]: I1002 02:04:39.255133 4885 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9226a04e-0195-42a1-b651-2f5379b1e2a9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:04:39 crc kubenswrapper[4885]: I1002 02:04:39.255146 4885 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9226a04e-0195-42a1-b651-2f5379b1e2a9-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 02:04:39 crc kubenswrapper[4885]: I1002 02:04:39.339430 4885 scope.go:117] "RemoveContainer" containerID="cc2a1cdc982003a22e30118b6756d23620f619fe3029c82308eb0d2d8ccbcfa2" Oct 02 02:04:39 crc kubenswrapper[4885]: I1002 02:04:39.365336 4885 scope.go:117] "RemoveContainer" containerID="bc5a69f8f4d26b9f7c36647a448c3943fedcadcf35e24b1c46b0c1cb51a86fef" Oct 02 02:04:39 crc kubenswrapper[4885]: I1002 02:04:39.385244 4885 scope.go:117] "RemoveContainer" containerID="f0fbd443af35d475dd0d4bf7f4e6dcb68f738868a512f56d411ef9996a3aadca" Oct 02 02:04:39 crc kubenswrapper[4885]: E1002 02:04:39.385861 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f0fbd443af35d475dd0d4bf7f4e6dcb68f738868a512f56d411ef9996a3aadca\": container with ID starting with f0fbd443af35d475dd0d4bf7f4e6dcb68f738868a512f56d411ef9996a3aadca not found: ID does not exist" containerID="f0fbd443af35d475dd0d4bf7f4e6dcb68f738868a512f56d411ef9996a3aadca" Oct 02 02:04:39 crc kubenswrapper[4885]: I1002 02:04:39.385926 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f0fbd443af35d475dd0d4bf7f4e6dcb68f738868a512f56d411ef9996a3aadca"} err="failed to get container status \"f0fbd443af35d475dd0d4bf7f4e6dcb68f738868a512f56d411ef9996a3aadca\": rpc error: code = NotFound desc = could not find container \"f0fbd443af35d475dd0d4bf7f4e6dcb68f738868a512f56d411ef9996a3aadca\": container with ID starting with f0fbd443af35d475dd0d4bf7f4e6dcb68f738868a512f56d411ef9996a3aadca not found: ID does not exist" Oct 02 02:04:39 crc kubenswrapper[4885]: I1002 02:04:39.385960 4885 scope.go:117] "RemoveContainer" containerID="cc2a1cdc982003a22e30118b6756d23620f619fe3029c82308eb0d2d8ccbcfa2" Oct 02 02:04:39 crc kubenswrapper[4885]: E1002 02:04:39.386403 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cc2a1cdc982003a22e30118b6756d23620f619fe3029c82308eb0d2d8ccbcfa2\": container with ID starting with cc2a1cdc982003a22e30118b6756d23620f619fe3029c82308eb0d2d8ccbcfa2 not found: ID does not exist" containerID="cc2a1cdc982003a22e30118b6756d23620f619fe3029c82308eb0d2d8ccbcfa2" Oct 02 02:04:39 crc kubenswrapper[4885]: I1002 02:04:39.386434 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc2a1cdc982003a22e30118b6756d23620f619fe3029c82308eb0d2d8ccbcfa2"} err="failed to get container status \"cc2a1cdc982003a22e30118b6756d23620f619fe3029c82308eb0d2d8ccbcfa2\": rpc error: code = NotFound desc = could not find container \"cc2a1cdc982003a22e30118b6756d23620f619fe3029c82308eb0d2d8ccbcfa2\": container with ID starting with cc2a1cdc982003a22e30118b6756d23620f619fe3029c82308eb0d2d8ccbcfa2 not found: ID does not exist" Oct 02 02:04:39 crc kubenswrapper[4885]: I1002 02:04:39.386452 4885 scope.go:117] "RemoveContainer" containerID="bc5a69f8f4d26b9f7c36647a448c3943fedcadcf35e24b1c46b0c1cb51a86fef" Oct 02 02:04:39 crc kubenswrapper[4885]: E1002 02:04:39.386682 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bc5a69f8f4d26b9f7c36647a448c3943fedcadcf35e24b1c46b0c1cb51a86fef\": container with ID starting with bc5a69f8f4d26b9f7c36647a448c3943fedcadcf35e24b1c46b0c1cb51a86fef not found: ID does not exist" containerID="bc5a69f8f4d26b9f7c36647a448c3943fedcadcf35e24b1c46b0c1cb51a86fef" Oct 02 02:04:39 crc kubenswrapper[4885]: I1002 02:04:39.386715 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bc5a69f8f4d26b9f7c36647a448c3943fedcadcf35e24b1c46b0c1cb51a86fef"} err="failed to get container status \"bc5a69f8f4d26b9f7c36647a448c3943fedcadcf35e24b1c46b0c1cb51a86fef\": rpc error: code = NotFound desc = could not find container \"bc5a69f8f4d26b9f7c36647a448c3943fedcadcf35e24b1c46b0c1cb51a86fef\": container with ID starting with bc5a69f8f4d26b9f7c36647a448c3943fedcadcf35e24b1c46b0c1cb51a86fef not found: ID does not exist" Oct 02 02:04:39 crc kubenswrapper[4885]: I1002 02:04:39.795092 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 02:04:39 crc kubenswrapper[4885]: I1002 02:04:39.804115 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 02 02:04:39 crc kubenswrapper[4885]: I1002 02:04:39.828645 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 02 02:04:39 crc kubenswrapper[4885]: E1002 02:04:39.829078 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9226a04e-0195-42a1-b651-2f5379b1e2a9" containerName="proxy-httpd" Oct 02 02:04:39 crc kubenswrapper[4885]: I1002 02:04:39.829133 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="9226a04e-0195-42a1-b651-2f5379b1e2a9" containerName="proxy-httpd" Oct 02 02:04:39 crc kubenswrapper[4885]: E1002 02:04:39.829150 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9226a04e-0195-42a1-b651-2f5379b1e2a9" containerName="ceilometer-notification-agent" Oct 02 02:04:39 crc kubenswrapper[4885]: I1002 02:04:39.829156 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="9226a04e-0195-42a1-b651-2f5379b1e2a9" containerName="ceilometer-notification-agent" Oct 02 02:04:39 crc kubenswrapper[4885]: E1002 02:04:39.829174 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9226a04e-0195-42a1-b651-2f5379b1e2a9" containerName="sg-core" Oct 02 02:04:39 crc kubenswrapper[4885]: I1002 02:04:39.829180 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="9226a04e-0195-42a1-b651-2f5379b1e2a9" containerName="sg-core" Oct 02 02:04:39 crc kubenswrapper[4885]: I1002 02:04:39.829382 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="9226a04e-0195-42a1-b651-2f5379b1e2a9" containerName="proxy-httpd" Oct 02 02:04:39 crc kubenswrapper[4885]: I1002 02:04:39.829396 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="9226a04e-0195-42a1-b651-2f5379b1e2a9" containerName="ceilometer-notification-agent" Oct 02 02:04:39 crc kubenswrapper[4885]: I1002 02:04:39.829411 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="9226a04e-0195-42a1-b651-2f5379b1e2a9" containerName="sg-core" Oct 02 02:04:39 crc kubenswrapper[4885]: I1002 02:04:39.831674 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 02:04:39 crc kubenswrapper[4885]: I1002 02:04:39.834313 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 02 02:04:39 crc kubenswrapper[4885]: I1002 02:04:39.834676 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 02 02:04:39 crc kubenswrapper[4885]: I1002 02:04:39.840474 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 02:04:39 crc kubenswrapper[4885]: I1002 02:04:39.864851 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1811543-6bba-4856-b8e1-814c7592723b-config-data\") pod \"ceilometer-0\" (UID: \"d1811543-6bba-4856-b8e1-814c7592723b\") " pod="openstack/ceilometer-0" Oct 02 02:04:39 crc kubenswrapper[4885]: I1002 02:04:39.864911 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d1811543-6bba-4856-b8e1-814c7592723b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d1811543-6bba-4856-b8e1-814c7592723b\") " pod="openstack/ceilometer-0" Oct 02 02:04:39 crc kubenswrapper[4885]: I1002 02:04:39.865000 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nj4k4\" (UniqueName: \"kubernetes.io/projected/d1811543-6bba-4856-b8e1-814c7592723b-kube-api-access-nj4k4\") pod \"ceilometer-0\" (UID: \"d1811543-6bba-4856-b8e1-814c7592723b\") " pod="openstack/ceilometer-0" Oct 02 02:04:39 crc kubenswrapper[4885]: I1002 02:04:39.865142 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1811543-6bba-4856-b8e1-814c7592723b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d1811543-6bba-4856-b8e1-814c7592723b\") " pod="openstack/ceilometer-0" Oct 02 02:04:39 crc kubenswrapper[4885]: I1002 02:04:39.865213 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d1811543-6bba-4856-b8e1-814c7592723b-log-httpd\") pod \"ceilometer-0\" (UID: \"d1811543-6bba-4856-b8e1-814c7592723b\") " pod="openstack/ceilometer-0" Oct 02 02:04:39 crc kubenswrapper[4885]: I1002 02:04:39.865304 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d1811543-6bba-4856-b8e1-814c7592723b-run-httpd\") pod \"ceilometer-0\" (UID: \"d1811543-6bba-4856-b8e1-814c7592723b\") " pod="openstack/ceilometer-0" Oct 02 02:04:39 crc kubenswrapper[4885]: I1002 02:04:39.865559 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d1811543-6bba-4856-b8e1-814c7592723b-scripts\") pod \"ceilometer-0\" (UID: \"d1811543-6bba-4856-b8e1-814c7592723b\") " pod="openstack/ceilometer-0" Oct 02 02:04:39 crc kubenswrapper[4885]: I1002 02:04:39.869900 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-79659d7df7-6dkj8" Oct 02 02:04:39 crc kubenswrapper[4885]: I1002 02:04:39.948416 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-6d7c47744-vjcbs" Oct 02 02:04:39 crc kubenswrapper[4885]: I1002 02:04:39.976740 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/35248aca-4932-408d-a758-766cce4ae517-config-data\") pod \"35248aca-4932-408d-a758-766cce4ae517\" (UID: \"35248aca-4932-408d-a758-766cce4ae517\") " Oct 02 02:04:39 crc kubenswrapper[4885]: I1002 02:04:39.976784 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/35248aca-4932-408d-a758-766cce4ae517-horizon-secret-key\") pod \"35248aca-4932-408d-a758-766cce4ae517\" (UID: \"35248aca-4932-408d-a758-766cce4ae517\") " Oct 02 02:04:39 crc kubenswrapper[4885]: I1002 02:04:39.976837 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/35248aca-4932-408d-a758-766cce4ae517-scripts\") pod \"35248aca-4932-408d-a758-766cce4ae517\" (UID: \"35248aca-4932-408d-a758-766cce4ae517\") " Oct 02 02:04:39 crc kubenswrapper[4885]: I1002 02:04:39.976893 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tm2rl\" (UniqueName: \"kubernetes.io/projected/35248aca-4932-408d-a758-766cce4ae517-kube-api-access-tm2rl\") pod \"35248aca-4932-408d-a758-766cce4ae517\" (UID: \"35248aca-4932-408d-a758-766cce4ae517\") " Oct 02 02:04:39 crc kubenswrapper[4885]: I1002 02:04:39.977081 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/35248aca-4932-408d-a758-766cce4ae517-logs\") pod \"35248aca-4932-408d-a758-766cce4ae517\" (UID: \"35248aca-4932-408d-a758-766cce4ae517\") " Oct 02 02:04:39 crc kubenswrapper[4885]: I1002 02:04:39.977925 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d1811543-6bba-4856-b8e1-814c7592723b-scripts\") pod \"ceilometer-0\" (UID: \"d1811543-6bba-4856-b8e1-814c7592723b\") " pod="openstack/ceilometer-0" Oct 02 02:04:39 crc kubenswrapper[4885]: I1002 02:04:39.977976 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1811543-6bba-4856-b8e1-814c7592723b-config-data\") pod \"ceilometer-0\" (UID: \"d1811543-6bba-4856-b8e1-814c7592723b\") " pod="openstack/ceilometer-0" Oct 02 02:04:39 crc kubenswrapper[4885]: I1002 02:04:39.978003 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d1811543-6bba-4856-b8e1-814c7592723b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d1811543-6bba-4856-b8e1-814c7592723b\") " pod="openstack/ceilometer-0" Oct 02 02:04:39 crc kubenswrapper[4885]: I1002 02:04:39.978026 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nj4k4\" (UniqueName: \"kubernetes.io/projected/d1811543-6bba-4856-b8e1-814c7592723b-kube-api-access-nj4k4\") pod \"ceilometer-0\" (UID: \"d1811543-6bba-4856-b8e1-814c7592723b\") " pod="openstack/ceilometer-0" Oct 02 02:04:39 crc kubenswrapper[4885]: I1002 02:04:39.978115 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1811543-6bba-4856-b8e1-814c7592723b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d1811543-6bba-4856-b8e1-814c7592723b\") " pod="openstack/ceilometer-0" Oct 02 02:04:39 crc kubenswrapper[4885]: I1002 02:04:39.978147 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d1811543-6bba-4856-b8e1-814c7592723b-log-httpd\") pod \"ceilometer-0\" (UID: \"d1811543-6bba-4856-b8e1-814c7592723b\") " pod="openstack/ceilometer-0" Oct 02 02:04:39 crc kubenswrapper[4885]: I1002 02:04:39.978168 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d1811543-6bba-4856-b8e1-814c7592723b-run-httpd\") pod \"ceilometer-0\" (UID: \"d1811543-6bba-4856-b8e1-814c7592723b\") " pod="openstack/ceilometer-0" Oct 02 02:04:39 crc kubenswrapper[4885]: I1002 02:04:39.978596 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d1811543-6bba-4856-b8e1-814c7592723b-run-httpd\") pod \"ceilometer-0\" (UID: \"d1811543-6bba-4856-b8e1-814c7592723b\") " pod="openstack/ceilometer-0" Oct 02 02:04:39 crc kubenswrapper[4885]: I1002 02:04:39.979541 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d1811543-6bba-4856-b8e1-814c7592723b-log-httpd\") pod \"ceilometer-0\" (UID: \"d1811543-6bba-4856-b8e1-814c7592723b\") " pod="openstack/ceilometer-0" Oct 02 02:04:39 crc kubenswrapper[4885]: I1002 02:04:39.980056 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/35248aca-4932-408d-a758-766cce4ae517-logs" (OuterVolumeSpecName: "logs") pod "35248aca-4932-408d-a758-766cce4ae517" (UID: "35248aca-4932-408d-a758-766cce4ae517"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:04:39 crc kubenswrapper[4885]: I1002 02:04:39.983608 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35248aca-4932-408d-a758-766cce4ae517-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "35248aca-4932-408d-a758-766cce4ae517" (UID: "35248aca-4932-408d-a758-766cce4ae517"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:04:39 crc kubenswrapper[4885]: I1002 02:04:39.984181 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35248aca-4932-408d-a758-766cce4ae517-kube-api-access-tm2rl" (OuterVolumeSpecName: "kube-api-access-tm2rl") pod "35248aca-4932-408d-a758-766cce4ae517" (UID: "35248aca-4932-408d-a758-766cce4ae517"). InnerVolumeSpecName "kube-api-access-tm2rl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:04:39 crc kubenswrapper[4885]: I1002 02:04:39.984802 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1811543-6bba-4856-b8e1-814c7592723b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d1811543-6bba-4856-b8e1-814c7592723b\") " pod="openstack/ceilometer-0" Oct 02 02:04:39 crc kubenswrapper[4885]: I1002 02:04:39.986236 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d1811543-6bba-4856-b8e1-814c7592723b-scripts\") pod \"ceilometer-0\" (UID: \"d1811543-6bba-4856-b8e1-814c7592723b\") " pod="openstack/ceilometer-0" Oct 02 02:04:39 crc kubenswrapper[4885]: I1002 02:04:39.986998 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1811543-6bba-4856-b8e1-814c7592723b-config-data\") pod \"ceilometer-0\" (UID: \"d1811543-6bba-4856-b8e1-814c7592723b\") " pod="openstack/ceilometer-0" Oct 02 02:04:39 crc kubenswrapper[4885]: I1002 02:04:39.999540 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d1811543-6bba-4856-b8e1-814c7592723b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d1811543-6bba-4856-b8e1-814c7592723b\") " pod="openstack/ceilometer-0" Oct 02 02:04:40 crc kubenswrapper[4885]: I1002 02:04:40.001248 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nj4k4\" (UniqueName: \"kubernetes.io/projected/d1811543-6bba-4856-b8e1-814c7592723b-kube-api-access-nj4k4\") pod \"ceilometer-0\" (UID: \"d1811543-6bba-4856-b8e1-814c7592723b\") " pod="openstack/ceilometer-0" Oct 02 02:04:40 crc kubenswrapper[4885]: I1002 02:04:40.027804 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/35248aca-4932-408d-a758-766cce4ae517-scripts" (OuterVolumeSpecName: "scripts") pod "35248aca-4932-408d-a758-766cce4ae517" (UID: "35248aca-4932-408d-a758-766cce4ae517"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:04:40 crc kubenswrapper[4885]: I1002 02:04:40.029640 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/35248aca-4932-408d-a758-766cce4ae517-config-data" (OuterVolumeSpecName: "config-data") pod "35248aca-4932-408d-a758-766cce4ae517" (UID: "35248aca-4932-408d-a758-766cce4ae517"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:04:40 crc kubenswrapper[4885]: I1002 02:04:40.059901 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9226a04e-0195-42a1-b651-2f5379b1e2a9" path="/var/lib/kubelet/pods/9226a04e-0195-42a1-b651-2f5379b1e2a9/volumes" Oct 02 02:04:40 crc kubenswrapper[4885]: I1002 02:04:40.081001 4885 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/35248aca-4932-408d-a758-766cce4ae517-logs\") on node \"crc\" DevicePath \"\"" Oct 02 02:04:40 crc kubenswrapper[4885]: I1002 02:04:40.081025 4885 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/35248aca-4932-408d-a758-766cce4ae517-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 02:04:40 crc kubenswrapper[4885]: I1002 02:04:40.081035 4885 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/35248aca-4932-408d-a758-766cce4ae517-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 02 02:04:40 crc kubenswrapper[4885]: I1002 02:04:40.081044 4885 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/35248aca-4932-408d-a758-766cce4ae517-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 02:04:40 crc kubenswrapper[4885]: I1002 02:04:40.081052 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tm2rl\" (UniqueName: \"kubernetes.io/projected/35248aca-4932-408d-a758-766cce4ae517-kube-api-access-tm2rl\") on node \"crc\" DevicePath \"\"" Oct 02 02:04:40 crc kubenswrapper[4885]: I1002 02:04:40.095718 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-8558cb5fd9-8cr6r" Oct 02 02:04:40 crc kubenswrapper[4885]: I1002 02:04:40.156072 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 02:04:40 crc kubenswrapper[4885]: I1002 02:04:40.181938 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8ad39937-75ce-468c-90df-b3abbaefe498-logs\") pod \"8ad39937-75ce-468c-90df-b3abbaefe498\" (UID: \"8ad39937-75ce-468c-90df-b3abbaefe498\") " Oct 02 02:04:40 crc kubenswrapper[4885]: I1002 02:04:40.182117 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8ad39937-75ce-468c-90df-b3abbaefe498-scripts\") pod \"8ad39937-75ce-468c-90df-b3abbaefe498\" (UID: \"8ad39937-75ce-468c-90df-b3abbaefe498\") " Oct 02 02:04:40 crc kubenswrapper[4885]: I1002 02:04:40.182200 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2mfc\" (UniqueName: \"kubernetes.io/projected/8ad39937-75ce-468c-90df-b3abbaefe498-kube-api-access-x2mfc\") pod \"8ad39937-75ce-468c-90df-b3abbaefe498\" (UID: \"8ad39937-75ce-468c-90df-b3abbaefe498\") " Oct 02 02:04:40 crc kubenswrapper[4885]: I1002 02:04:40.182289 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8ad39937-75ce-468c-90df-b3abbaefe498-logs" (OuterVolumeSpecName: "logs") pod "8ad39937-75ce-468c-90df-b3abbaefe498" (UID: "8ad39937-75ce-468c-90df-b3abbaefe498"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:04:40 crc kubenswrapper[4885]: I1002 02:04:40.182380 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8ad39937-75ce-468c-90df-b3abbaefe498-config-data\") pod \"8ad39937-75ce-468c-90df-b3abbaefe498\" (UID: \"8ad39937-75ce-468c-90df-b3abbaefe498\") " Oct 02 02:04:40 crc kubenswrapper[4885]: I1002 02:04:40.182487 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8ad39937-75ce-468c-90df-b3abbaefe498-horizon-secret-key\") pod \"8ad39937-75ce-468c-90df-b3abbaefe498\" (UID: \"8ad39937-75ce-468c-90df-b3abbaefe498\") " Oct 02 02:04:40 crc kubenswrapper[4885]: I1002 02:04:40.182913 4885 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8ad39937-75ce-468c-90df-b3abbaefe498-logs\") on node \"crc\" DevicePath \"\"" Oct 02 02:04:40 crc kubenswrapper[4885]: I1002 02:04:40.183444 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5686549c7f-j9kgc" Oct 02 02:04:40 crc kubenswrapper[4885]: I1002 02:04:40.187599 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ad39937-75ce-468c-90df-b3abbaefe498-kube-api-access-x2mfc" (OuterVolumeSpecName: "kube-api-access-x2mfc") pod "8ad39937-75ce-468c-90df-b3abbaefe498" (UID: "8ad39937-75ce-468c-90df-b3abbaefe498"). InnerVolumeSpecName "kube-api-access-x2mfc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:04:40 crc kubenswrapper[4885]: I1002 02:04:40.187824 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ad39937-75ce-468c-90df-b3abbaefe498-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "8ad39937-75ce-468c-90df-b3abbaefe498" (UID: "8ad39937-75ce-468c-90df-b3abbaefe498"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:04:40 crc kubenswrapper[4885]: I1002 02:04:40.207289 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8ad39937-75ce-468c-90df-b3abbaefe498-config-data" (OuterVolumeSpecName: "config-data") pod "8ad39937-75ce-468c-90df-b3abbaefe498" (UID: "8ad39937-75ce-468c-90df-b3abbaefe498"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:04:40 crc kubenswrapper[4885]: I1002 02:04:40.216594 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8ad39937-75ce-468c-90df-b3abbaefe498-scripts" (OuterVolumeSpecName: "scripts") pod "8ad39937-75ce-468c-90df-b3abbaefe498" (UID: "8ad39937-75ce-468c-90df-b3abbaefe498"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:04:40 crc kubenswrapper[4885]: I1002 02:04:40.225699 4885 generic.go:334] "Generic (PLEG): container finished" podID="35248aca-4932-408d-a758-766cce4ae517" containerID="ef22dce64ab597d1b805bb0b4263ecaa74301822abb44a47719aa7375f39e18f" exitCode=137 Oct 02 02:04:40 crc kubenswrapper[4885]: I1002 02:04:40.225728 4885 generic.go:334] "Generic (PLEG): container finished" podID="35248aca-4932-408d-a758-766cce4ae517" containerID="8d94317efe50870f3aa4725f3ee16894d0d9d196ed65560002f5e1578770a242" exitCode=137 Oct 02 02:04:40 crc kubenswrapper[4885]: I1002 02:04:40.225811 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-79659d7df7-6dkj8" Oct 02 02:04:40 crc kubenswrapper[4885]: I1002 02:04:40.226677 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-79659d7df7-6dkj8" event={"ID":"35248aca-4932-408d-a758-766cce4ae517","Type":"ContainerDied","Data":"ef22dce64ab597d1b805bb0b4263ecaa74301822abb44a47719aa7375f39e18f"} Oct 02 02:04:40 crc kubenswrapper[4885]: I1002 02:04:40.226704 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-79659d7df7-6dkj8" event={"ID":"35248aca-4932-408d-a758-766cce4ae517","Type":"ContainerDied","Data":"8d94317efe50870f3aa4725f3ee16894d0d9d196ed65560002f5e1578770a242"} Oct 02 02:04:40 crc kubenswrapper[4885]: I1002 02:04:40.226718 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-79659d7df7-6dkj8" event={"ID":"35248aca-4932-408d-a758-766cce4ae517","Type":"ContainerDied","Data":"a508a5af806d151a21550ebba73caf8ae175cddf2a7941e341c4d67e97c3af77"} Oct 02 02:04:40 crc kubenswrapper[4885]: I1002 02:04:40.226735 4885 scope.go:117] "RemoveContainer" containerID="ef22dce64ab597d1b805bb0b4263ecaa74301822abb44a47719aa7375f39e18f" Oct 02 02:04:40 crc kubenswrapper[4885]: I1002 02:04:40.246452 4885 generic.go:334] "Generic (PLEG): container finished" podID="5056f944-c3cf-45e1-99f5-f93132717b93" containerID="41549198f0212f18def5c424af45e5fc7686634cca8e0d8a4bff946650eb6e5e" exitCode=137 Oct 02 02:04:40 crc kubenswrapper[4885]: I1002 02:04:40.246486 4885 generic.go:334] "Generic (PLEG): container finished" podID="5056f944-c3cf-45e1-99f5-f93132717b93" containerID="62c198b211a751154fad0eba8680b9e2bb3fd8d45f10cc75b898a36e966de946" exitCode=137 Oct 02 02:04:40 crc kubenswrapper[4885]: I1002 02:04:40.246528 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5686549c7f-j9kgc" event={"ID":"5056f944-c3cf-45e1-99f5-f93132717b93","Type":"ContainerDied","Data":"41549198f0212f18def5c424af45e5fc7686634cca8e0d8a4bff946650eb6e5e"} Oct 02 02:04:40 crc kubenswrapper[4885]: I1002 02:04:40.246556 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5686549c7f-j9kgc" event={"ID":"5056f944-c3cf-45e1-99f5-f93132717b93","Type":"ContainerDied","Data":"62c198b211a751154fad0eba8680b9e2bb3fd8d45f10cc75b898a36e966de946"} Oct 02 02:04:40 crc kubenswrapper[4885]: I1002 02:04:40.246567 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5686549c7f-j9kgc" event={"ID":"5056f944-c3cf-45e1-99f5-f93132717b93","Type":"ContainerDied","Data":"ec729febcc65ceb7baec51b7003eb6b379eac5a90a92b9c59151858053988768"} Oct 02 02:04:40 crc kubenswrapper[4885]: I1002 02:04:40.246623 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5686549c7f-j9kgc" Oct 02 02:04:40 crc kubenswrapper[4885]: I1002 02:04:40.250410 4885 generic.go:334] "Generic (PLEG): container finished" podID="8ad39937-75ce-468c-90df-b3abbaefe498" containerID="1ebc0d49cd4350c0748387677aa9ada287742667a4fa2985825e62867011aa82" exitCode=137 Oct 02 02:04:40 crc kubenswrapper[4885]: I1002 02:04:40.250449 4885 generic.go:334] "Generic (PLEG): container finished" podID="8ad39937-75ce-468c-90df-b3abbaefe498" containerID="39adfd0e77c1141d8aa33b2b5ea71968edba0ddc693a26b57ab6d5ff366e56a5" exitCode=137 Oct 02 02:04:40 crc kubenswrapper[4885]: I1002 02:04:40.250456 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-8558cb5fd9-8cr6r" event={"ID":"8ad39937-75ce-468c-90df-b3abbaefe498","Type":"ContainerDied","Data":"1ebc0d49cd4350c0748387677aa9ada287742667a4fa2985825e62867011aa82"} Oct 02 02:04:40 crc kubenswrapper[4885]: I1002 02:04:40.250476 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-8558cb5fd9-8cr6r" event={"ID":"8ad39937-75ce-468c-90df-b3abbaefe498","Type":"ContainerDied","Data":"39adfd0e77c1141d8aa33b2b5ea71968edba0ddc693a26b57ab6d5ff366e56a5"} Oct 02 02:04:40 crc kubenswrapper[4885]: I1002 02:04:40.250486 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-8558cb5fd9-8cr6r" event={"ID":"8ad39937-75ce-468c-90df-b3abbaefe498","Type":"ContainerDied","Data":"6f02b41e568cd4c191d3ac36f5fd8d6536883727af2c4417485c66f8afc7b87f"} Oct 02 02:04:40 crc kubenswrapper[4885]: I1002 02:04:40.250431 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-8558cb5fd9-8cr6r" Oct 02 02:04:40 crc kubenswrapper[4885]: I1002 02:04:40.284143 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/5056f944-c3cf-45e1-99f5-f93132717b93-horizon-secret-key\") pod \"5056f944-c3cf-45e1-99f5-f93132717b93\" (UID: \"5056f944-c3cf-45e1-99f5-f93132717b93\") " Oct 02 02:04:40 crc kubenswrapper[4885]: I1002 02:04:40.284206 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gx8x4\" (UniqueName: \"kubernetes.io/projected/5056f944-c3cf-45e1-99f5-f93132717b93-kube-api-access-gx8x4\") pod \"5056f944-c3cf-45e1-99f5-f93132717b93\" (UID: \"5056f944-c3cf-45e1-99f5-f93132717b93\") " Oct 02 02:04:40 crc kubenswrapper[4885]: I1002 02:04:40.284227 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5056f944-c3cf-45e1-99f5-f93132717b93-config-data\") pod \"5056f944-c3cf-45e1-99f5-f93132717b93\" (UID: \"5056f944-c3cf-45e1-99f5-f93132717b93\") " Oct 02 02:04:40 crc kubenswrapper[4885]: I1002 02:04:40.284332 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5056f944-c3cf-45e1-99f5-f93132717b93-logs\") pod \"5056f944-c3cf-45e1-99f5-f93132717b93\" (UID: \"5056f944-c3cf-45e1-99f5-f93132717b93\") " Oct 02 02:04:40 crc kubenswrapper[4885]: I1002 02:04:40.284363 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5056f944-c3cf-45e1-99f5-f93132717b93-scripts\") pod \"5056f944-c3cf-45e1-99f5-f93132717b93\" (UID: \"5056f944-c3cf-45e1-99f5-f93132717b93\") " Oct 02 02:04:40 crc kubenswrapper[4885]: I1002 02:04:40.285751 4885 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8ad39937-75ce-468c-90df-b3abbaefe498-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 02:04:40 crc kubenswrapper[4885]: I1002 02:04:40.285942 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2mfc\" (UniqueName: \"kubernetes.io/projected/8ad39937-75ce-468c-90df-b3abbaefe498-kube-api-access-x2mfc\") on node \"crc\" DevicePath \"\"" Oct 02 02:04:40 crc kubenswrapper[4885]: I1002 02:04:40.285954 4885 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8ad39937-75ce-468c-90df-b3abbaefe498-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 02:04:40 crc kubenswrapper[4885]: I1002 02:04:40.285963 4885 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8ad39937-75ce-468c-90df-b3abbaefe498-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 02 02:04:40 crc kubenswrapper[4885]: I1002 02:04:40.286551 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5056f944-c3cf-45e1-99f5-f93132717b93-logs" (OuterVolumeSpecName: "logs") pod "5056f944-c3cf-45e1-99f5-f93132717b93" (UID: "5056f944-c3cf-45e1-99f5-f93132717b93"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:04:40 crc kubenswrapper[4885]: I1002 02:04:40.292177 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5056f944-c3cf-45e1-99f5-f93132717b93-kube-api-access-gx8x4" (OuterVolumeSpecName: "kube-api-access-gx8x4") pod "5056f944-c3cf-45e1-99f5-f93132717b93" (UID: "5056f944-c3cf-45e1-99f5-f93132717b93"). InnerVolumeSpecName "kube-api-access-gx8x4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:04:40 crc kubenswrapper[4885]: I1002 02:04:40.292599 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5056f944-c3cf-45e1-99f5-f93132717b93-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "5056f944-c3cf-45e1-99f5-f93132717b93" (UID: "5056f944-c3cf-45e1-99f5-f93132717b93"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:04:40 crc kubenswrapper[4885]: I1002 02:04:40.309367 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5056f944-c3cf-45e1-99f5-f93132717b93-config-data" (OuterVolumeSpecName: "config-data") pod "5056f944-c3cf-45e1-99f5-f93132717b93" (UID: "5056f944-c3cf-45e1-99f5-f93132717b93"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:04:40 crc kubenswrapper[4885]: I1002 02:04:40.312127 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5056f944-c3cf-45e1-99f5-f93132717b93-scripts" (OuterVolumeSpecName: "scripts") pod "5056f944-c3cf-45e1-99f5-f93132717b93" (UID: "5056f944-c3cf-45e1-99f5-f93132717b93"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:04:40 crc kubenswrapper[4885]: I1002 02:04:40.391023 4885 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5056f944-c3cf-45e1-99f5-f93132717b93-logs\") on node \"crc\" DevicePath \"\"" Oct 02 02:04:40 crc kubenswrapper[4885]: I1002 02:04:40.391050 4885 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5056f944-c3cf-45e1-99f5-f93132717b93-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 02:04:40 crc kubenswrapper[4885]: I1002 02:04:40.391060 4885 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/5056f944-c3cf-45e1-99f5-f93132717b93-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 02 02:04:40 crc kubenswrapper[4885]: I1002 02:04:40.391069 4885 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5056f944-c3cf-45e1-99f5-f93132717b93-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 02:04:40 crc kubenswrapper[4885]: I1002 02:04:40.391077 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gx8x4\" (UniqueName: \"kubernetes.io/projected/5056f944-c3cf-45e1-99f5-f93132717b93-kube-api-access-gx8x4\") on node \"crc\" DevicePath \"\"" Oct 02 02:04:40 crc kubenswrapper[4885]: I1002 02:04:40.397456 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-79659d7df7-6dkj8"] Oct 02 02:04:40 crc kubenswrapper[4885]: I1002 02:04:40.405230 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-79659d7df7-6dkj8"] Oct 02 02:04:40 crc kubenswrapper[4885]: I1002 02:04:40.413236 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-8558cb5fd9-8cr6r"] Oct 02 02:04:40 crc kubenswrapper[4885]: I1002 02:04:40.418307 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-8558cb5fd9-8cr6r"] Oct 02 02:04:40 crc kubenswrapper[4885]: I1002 02:04:40.478913 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-7c49b4cffb-brdz9" Oct 02 02:04:40 crc kubenswrapper[4885]: I1002 02:04:40.567834 4885 scope.go:117] "RemoveContainer" containerID="8d94317efe50870f3aa4725f3ee16894d0d9d196ed65560002f5e1578770a242" Oct 02 02:04:40 crc kubenswrapper[4885]: I1002 02:04:40.603438 4885 scope.go:117] "RemoveContainer" containerID="ef22dce64ab597d1b805bb0b4263ecaa74301822abb44a47719aa7375f39e18f" Oct 02 02:04:40 crc kubenswrapper[4885]: E1002 02:04:40.604078 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ef22dce64ab597d1b805bb0b4263ecaa74301822abb44a47719aa7375f39e18f\": container with ID starting with ef22dce64ab597d1b805bb0b4263ecaa74301822abb44a47719aa7375f39e18f not found: ID does not exist" containerID="ef22dce64ab597d1b805bb0b4263ecaa74301822abb44a47719aa7375f39e18f" Oct 02 02:04:40 crc kubenswrapper[4885]: I1002 02:04:40.604128 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef22dce64ab597d1b805bb0b4263ecaa74301822abb44a47719aa7375f39e18f"} err="failed to get container status \"ef22dce64ab597d1b805bb0b4263ecaa74301822abb44a47719aa7375f39e18f\": rpc error: code = NotFound desc = could not find container \"ef22dce64ab597d1b805bb0b4263ecaa74301822abb44a47719aa7375f39e18f\": container with ID starting with ef22dce64ab597d1b805bb0b4263ecaa74301822abb44a47719aa7375f39e18f not found: ID does not exist" Oct 02 02:04:40 crc kubenswrapper[4885]: I1002 02:04:40.604165 4885 scope.go:117] "RemoveContainer" containerID="8d94317efe50870f3aa4725f3ee16894d0d9d196ed65560002f5e1578770a242" Oct 02 02:04:40 crc kubenswrapper[4885]: E1002 02:04:40.604676 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8d94317efe50870f3aa4725f3ee16894d0d9d196ed65560002f5e1578770a242\": container with ID starting with 8d94317efe50870f3aa4725f3ee16894d0d9d196ed65560002f5e1578770a242 not found: ID does not exist" containerID="8d94317efe50870f3aa4725f3ee16894d0d9d196ed65560002f5e1578770a242" Oct 02 02:04:40 crc kubenswrapper[4885]: I1002 02:04:40.604720 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d94317efe50870f3aa4725f3ee16894d0d9d196ed65560002f5e1578770a242"} err="failed to get container status \"8d94317efe50870f3aa4725f3ee16894d0d9d196ed65560002f5e1578770a242\": rpc error: code = NotFound desc = could not find container \"8d94317efe50870f3aa4725f3ee16894d0d9d196ed65560002f5e1578770a242\": container with ID starting with 8d94317efe50870f3aa4725f3ee16894d0d9d196ed65560002f5e1578770a242 not found: ID does not exist" Oct 02 02:04:40 crc kubenswrapper[4885]: I1002 02:04:40.604748 4885 scope.go:117] "RemoveContainer" containerID="ef22dce64ab597d1b805bb0b4263ecaa74301822abb44a47719aa7375f39e18f" Oct 02 02:04:40 crc kubenswrapper[4885]: I1002 02:04:40.605110 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef22dce64ab597d1b805bb0b4263ecaa74301822abb44a47719aa7375f39e18f"} err="failed to get container status \"ef22dce64ab597d1b805bb0b4263ecaa74301822abb44a47719aa7375f39e18f\": rpc error: code = NotFound desc = could not find container \"ef22dce64ab597d1b805bb0b4263ecaa74301822abb44a47719aa7375f39e18f\": container with ID starting with ef22dce64ab597d1b805bb0b4263ecaa74301822abb44a47719aa7375f39e18f not found: ID does not exist" Oct 02 02:04:40 crc kubenswrapper[4885]: I1002 02:04:40.605134 4885 scope.go:117] "RemoveContainer" containerID="8d94317efe50870f3aa4725f3ee16894d0d9d196ed65560002f5e1578770a242" Oct 02 02:04:40 crc kubenswrapper[4885]: I1002 02:04:40.605384 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d94317efe50870f3aa4725f3ee16894d0d9d196ed65560002f5e1578770a242"} err="failed to get container status \"8d94317efe50870f3aa4725f3ee16894d0d9d196ed65560002f5e1578770a242\": rpc error: code = NotFound desc = could not find container \"8d94317efe50870f3aa4725f3ee16894d0d9d196ed65560002f5e1578770a242\": container with ID starting with 8d94317efe50870f3aa4725f3ee16894d0d9d196ed65560002f5e1578770a242 not found: ID does not exist" Oct 02 02:04:40 crc kubenswrapper[4885]: I1002 02:04:40.605411 4885 scope.go:117] "RemoveContainer" containerID="41549198f0212f18def5c424af45e5fc7686634cca8e0d8a4bff946650eb6e5e" Oct 02 02:04:40 crc kubenswrapper[4885]: I1002 02:04:40.608644 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5686549c7f-j9kgc"] Oct 02 02:04:40 crc kubenswrapper[4885]: I1002 02:04:40.619391 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-5686549c7f-j9kgc"] Oct 02 02:04:40 crc kubenswrapper[4885]: I1002 02:04:40.653468 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 02:04:40 crc kubenswrapper[4885]: I1002 02:04:40.834635 4885 scope.go:117] "RemoveContainer" containerID="62c198b211a751154fad0eba8680b9e2bb3fd8d45f10cc75b898a36e966de946" Oct 02 02:04:40 crc kubenswrapper[4885]: W1002 02:04:40.840841 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd1811543_6bba_4856_b8e1_814c7592723b.slice/crio-30faa88c919983f7626a53ca812e2ca526eae85993762541b5a6fd6c39e4db3f WatchSource:0}: Error finding container 30faa88c919983f7626a53ca812e2ca526eae85993762541b5a6fd6c39e4db3f: Status 404 returned error can't find the container with id 30faa88c919983f7626a53ca812e2ca526eae85993762541b5a6fd6c39e4db3f Oct 02 02:04:40 crc kubenswrapper[4885]: I1002 02:04:40.907453 4885 scope.go:117] "RemoveContainer" containerID="41549198f0212f18def5c424af45e5fc7686634cca8e0d8a4bff946650eb6e5e" Oct 02 02:04:40 crc kubenswrapper[4885]: E1002 02:04:40.908012 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"41549198f0212f18def5c424af45e5fc7686634cca8e0d8a4bff946650eb6e5e\": container with ID starting with 41549198f0212f18def5c424af45e5fc7686634cca8e0d8a4bff946650eb6e5e not found: ID does not exist" containerID="41549198f0212f18def5c424af45e5fc7686634cca8e0d8a4bff946650eb6e5e" Oct 02 02:04:40 crc kubenswrapper[4885]: I1002 02:04:40.908084 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"41549198f0212f18def5c424af45e5fc7686634cca8e0d8a4bff946650eb6e5e"} err="failed to get container status \"41549198f0212f18def5c424af45e5fc7686634cca8e0d8a4bff946650eb6e5e\": rpc error: code = NotFound desc = could not find container \"41549198f0212f18def5c424af45e5fc7686634cca8e0d8a4bff946650eb6e5e\": container with ID starting with 41549198f0212f18def5c424af45e5fc7686634cca8e0d8a4bff946650eb6e5e not found: ID does not exist" Oct 02 02:04:40 crc kubenswrapper[4885]: I1002 02:04:40.908129 4885 scope.go:117] "RemoveContainer" containerID="62c198b211a751154fad0eba8680b9e2bb3fd8d45f10cc75b898a36e966de946" Oct 02 02:04:40 crc kubenswrapper[4885]: E1002 02:04:40.908615 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"62c198b211a751154fad0eba8680b9e2bb3fd8d45f10cc75b898a36e966de946\": container with ID starting with 62c198b211a751154fad0eba8680b9e2bb3fd8d45f10cc75b898a36e966de946 not found: ID does not exist" containerID="62c198b211a751154fad0eba8680b9e2bb3fd8d45f10cc75b898a36e966de946" Oct 02 02:04:40 crc kubenswrapper[4885]: I1002 02:04:40.908662 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"62c198b211a751154fad0eba8680b9e2bb3fd8d45f10cc75b898a36e966de946"} err="failed to get container status \"62c198b211a751154fad0eba8680b9e2bb3fd8d45f10cc75b898a36e966de946\": rpc error: code = NotFound desc = could not find container \"62c198b211a751154fad0eba8680b9e2bb3fd8d45f10cc75b898a36e966de946\": container with ID starting with 62c198b211a751154fad0eba8680b9e2bb3fd8d45f10cc75b898a36e966de946 not found: ID does not exist" Oct 02 02:04:40 crc kubenswrapper[4885]: I1002 02:04:40.908696 4885 scope.go:117] "RemoveContainer" containerID="41549198f0212f18def5c424af45e5fc7686634cca8e0d8a4bff946650eb6e5e" Oct 02 02:04:40 crc kubenswrapper[4885]: I1002 02:04:40.909164 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"41549198f0212f18def5c424af45e5fc7686634cca8e0d8a4bff946650eb6e5e"} err="failed to get container status \"41549198f0212f18def5c424af45e5fc7686634cca8e0d8a4bff946650eb6e5e\": rpc error: code = NotFound desc = could not find container \"41549198f0212f18def5c424af45e5fc7686634cca8e0d8a4bff946650eb6e5e\": container with ID starting with 41549198f0212f18def5c424af45e5fc7686634cca8e0d8a4bff946650eb6e5e not found: ID does not exist" Oct 02 02:04:40 crc kubenswrapper[4885]: I1002 02:04:40.909202 4885 scope.go:117] "RemoveContainer" containerID="62c198b211a751154fad0eba8680b9e2bb3fd8d45f10cc75b898a36e966de946" Oct 02 02:04:40 crc kubenswrapper[4885]: I1002 02:04:40.909630 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"62c198b211a751154fad0eba8680b9e2bb3fd8d45f10cc75b898a36e966de946"} err="failed to get container status \"62c198b211a751154fad0eba8680b9e2bb3fd8d45f10cc75b898a36e966de946\": rpc error: code = NotFound desc = could not find container \"62c198b211a751154fad0eba8680b9e2bb3fd8d45f10cc75b898a36e966de946\": container with ID starting with 62c198b211a751154fad0eba8680b9e2bb3fd8d45f10cc75b898a36e966de946 not found: ID does not exist" Oct 02 02:04:40 crc kubenswrapper[4885]: I1002 02:04:40.909694 4885 scope.go:117] "RemoveContainer" containerID="1ebc0d49cd4350c0748387677aa9ada287742667a4fa2985825e62867011aa82" Oct 02 02:04:41 crc kubenswrapper[4885]: I1002 02:04:41.215686 4885 scope.go:117] "RemoveContainer" containerID="39adfd0e77c1141d8aa33b2b5ea71968edba0ddc693a26b57ab6d5ff366e56a5" Oct 02 02:04:41 crc kubenswrapper[4885]: I1002 02:04:41.241240 4885 scope.go:117] "RemoveContainer" containerID="1ebc0d49cd4350c0748387677aa9ada287742667a4fa2985825e62867011aa82" Oct 02 02:04:41 crc kubenswrapper[4885]: E1002 02:04:41.242166 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ebc0d49cd4350c0748387677aa9ada287742667a4fa2985825e62867011aa82\": container with ID starting with 1ebc0d49cd4350c0748387677aa9ada287742667a4fa2985825e62867011aa82 not found: ID does not exist" containerID="1ebc0d49cd4350c0748387677aa9ada287742667a4fa2985825e62867011aa82" Oct 02 02:04:41 crc kubenswrapper[4885]: I1002 02:04:41.242222 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ebc0d49cd4350c0748387677aa9ada287742667a4fa2985825e62867011aa82"} err="failed to get container status \"1ebc0d49cd4350c0748387677aa9ada287742667a4fa2985825e62867011aa82\": rpc error: code = NotFound desc = could not find container \"1ebc0d49cd4350c0748387677aa9ada287742667a4fa2985825e62867011aa82\": container with ID starting with 1ebc0d49cd4350c0748387677aa9ada287742667a4fa2985825e62867011aa82 not found: ID does not exist" Oct 02 02:04:41 crc kubenswrapper[4885]: I1002 02:04:41.242330 4885 scope.go:117] "RemoveContainer" containerID="39adfd0e77c1141d8aa33b2b5ea71968edba0ddc693a26b57ab6d5ff366e56a5" Oct 02 02:04:41 crc kubenswrapper[4885]: E1002 02:04:41.242786 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"39adfd0e77c1141d8aa33b2b5ea71968edba0ddc693a26b57ab6d5ff366e56a5\": container with ID starting with 39adfd0e77c1141d8aa33b2b5ea71968edba0ddc693a26b57ab6d5ff366e56a5 not found: ID does not exist" containerID="39adfd0e77c1141d8aa33b2b5ea71968edba0ddc693a26b57ab6d5ff366e56a5" Oct 02 02:04:41 crc kubenswrapper[4885]: I1002 02:04:41.242842 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39adfd0e77c1141d8aa33b2b5ea71968edba0ddc693a26b57ab6d5ff366e56a5"} err="failed to get container status \"39adfd0e77c1141d8aa33b2b5ea71968edba0ddc693a26b57ab6d5ff366e56a5\": rpc error: code = NotFound desc = could not find container \"39adfd0e77c1141d8aa33b2b5ea71968edba0ddc693a26b57ab6d5ff366e56a5\": container with ID starting with 39adfd0e77c1141d8aa33b2b5ea71968edba0ddc693a26b57ab6d5ff366e56a5 not found: ID does not exist" Oct 02 02:04:41 crc kubenswrapper[4885]: I1002 02:04:41.242880 4885 scope.go:117] "RemoveContainer" containerID="1ebc0d49cd4350c0748387677aa9ada287742667a4fa2985825e62867011aa82" Oct 02 02:04:41 crc kubenswrapper[4885]: I1002 02:04:41.243447 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ebc0d49cd4350c0748387677aa9ada287742667a4fa2985825e62867011aa82"} err="failed to get container status \"1ebc0d49cd4350c0748387677aa9ada287742667a4fa2985825e62867011aa82\": rpc error: code = NotFound desc = could not find container \"1ebc0d49cd4350c0748387677aa9ada287742667a4fa2985825e62867011aa82\": container with ID starting with 1ebc0d49cd4350c0748387677aa9ada287742667a4fa2985825e62867011aa82 not found: ID does not exist" Oct 02 02:04:41 crc kubenswrapper[4885]: I1002 02:04:41.243479 4885 scope.go:117] "RemoveContainer" containerID="39adfd0e77c1141d8aa33b2b5ea71968edba0ddc693a26b57ab6d5ff366e56a5" Oct 02 02:04:41 crc kubenswrapper[4885]: I1002 02:04:41.243836 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39adfd0e77c1141d8aa33b2b5ea71968edba0ddc693a26b57ab6d5ff366e56a5"} err="failed to get container status \"39adfd0e77c1141d8aa33b2b5ea71968edba0ddc693a26b57ab6d5ff366e56a5\": rpc error: code = NotFound desc = could not find container \"39adfd0e77c1141d8aa33b2b5ea71968edba0ddc693a26b57ab6d5ff366e56a5\": container with ID starting with 39adfd0e77c1141d8aa33b2b5ea71968edba0ddc693a26b57ab6d5ff366e56a5 not found: ID does not exist" Oct 02 02:04:41 crc kubenswrapper[4885]: I1002 02:04:41.266832 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d1811543-6bba-4856-b8e1-814c7592723b","Type":"ContainerStarted","Data":"30faa88c919983f7626a53ca812e2ca526eae85993762541b5a6fd6c39e4db3f"} Oct 02 02:04:41 crc kubenswrapper[4885]: I1002 02:04:41.578042 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-6d7c47744-vjcbs" Oct 02 02:04:41 crc kubenswrapper[4885]: I1002 02:04:41.657187 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7c49b4cffb-brdz9"] Oct 02 02:04:41 crc kubenswrapper[4885]: I1002 02:04:41.657501 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7c49b4cffb-brdz9" podUID="b0fddcb3-3649-48f4-aa8f-85dd1c8137cf" containerName="horizon-log" containerID="cri-o://9a7c1423da72336ecfa2cfe108bb2a287a4349ab5629c7bfb13486150d01b56a" gracePeriod=30 Oct 02 02:04:41 crc kubenswrapper[4885]: I1002 02:04:41.657902 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7c49b4cffb-brdz9" podUID="b0fddcb3-3649-48f4-aa8f-85dd1c8137cf" containerName="horizon" containerID="cri-o://87f582c0838b65442115c622d40cc23454484a73c597b4939299160d6e09873a" gracePeriod=30 Oct 02 02:04:41 crc kubenswrapper[4885]: I1002 02:04:41.667395 4885 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-7c49b4cffb-brdz9" podUID="b0fddcb3-3649-48f4-aa8f-85dd1c8137cf" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.146:8443/dashboard/auth/login/?next=/dashboard/\": EOF" Oct 02 02:04:41 crc kubenswrapper[4885]: I1002 02:04:41.672758 4885 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-7c49b4cffb-brdz9" podUID="b0fddcb3-3649-48f4-aa8f-85dd1c8137cf" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.146:8443/dashboard/auth/login/?next=/dashboard/\": EOF" Oct 02 02:04:42 crc kubenswrapper[4885]: I1002 02:04:42.059207 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="35248aca-4932-408d-a758-766cce4ae517" path="/var/lib/kubelet/pods/35248aca-4932-408d-a758-766cce4ae517/volumes" Oct 02 02:04:42 crc kubenswrapper[4885]: I1002 02:04:42.060654 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5056f944-c3cf-45e1-99f5-f93132717b93" path="/var/lib/kubelet/pods/5056f944-c3cf-45e1-99f5-f93132717b93/volumes" Oct 02 02:04:42 crc kubenswrapper[4885]: I1002 02:04:42.061621 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8ad39937-75ce-468c-90df-b3abbaefe498" path="/var/lib/kubelet/pods/8ad39937-75ce-468c-90df-b3abbaefe498/volumes" Oct 02 02:04:42 crc kubenswrapper[4885]: I1002 02:04:42.288702 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d1811543-6bba-4856-b8e1-814c7592723b","Type":"ContainerStarted","Data":"698a9529d687dfa8a21be0b32df8bc1601b7907df0cc82a1db5354711128fd20"} Oct 02 02:04:43 crc kubenswrapper[4885]: I1002 02:04:43.634513 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-84b966f6c9-h9tlf" Oct 02 02:04:43 crc kubenswrapper[4885]: I1002 02:04:43.709772 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-x6dgq"] Oct 02 02:04:43 crc kubenswrapper[4885]: I1002 02:04:43.710048 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-8b5c85b87-x6dgq" podUID="d055d017-5640-4a3d-b684-6ce2755eab74" containerName="dnsmasq-dns" containerID="cri-o://64f15f37e85e8ff6d8c80cc01cf3fed9b5f658a2dccbcf79ffaccdbed931c8d5" gracePeriod=10 Oct 02 02:04:44 crc kubenswrapper[4885]: I1002 02:04:44.174084 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-x6dgq" Oct 02 02:04:44 crc kubenswrapper[4885]: I1002 02:04:44.281144 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d055d017-5640-4a3d-b684-6ce2755eab74-config\") pod \"d055d017-5640-4a3d-b684-6ce2755eab74\" (UID: \"d055d017-5640-4a3d-b684-6ce2755eab74\") " Oct 02 02:04:44 crc kubenswrapper[4885]: I1002 02:04:44.281219 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d055d017-5640-4a3d-b684-6ce2755eab74-dns-swift-storage-0\") pod \"d055d017-5640-4a3d-b684-6ce2755eab74\" (UID: \"d055d017-5640-4a3d-b684-6ce2755eab74\") " Oct 02 02:04:44 crc kubenswrapper[4885]: I1002 02:04:44.281276 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d055d017-5640-4a3d-b684-6ce2755eab74-dns-svc\") pod \"d055d017-5640-4a3d-b684-6ce2755eab74\" (UID: \"d055d017-5640-4a3d-b684-6ce2755eab74\") " Oct 02 02:04:44 crc kubenswrapper[4885]: I1002 02:04:44.281401 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jz9t7\" (UniqueName: \"kubernetes.io/projected/d055d017-5640-4a3d-b684-6ce2755eab74-kube-api-access-jz9t7\") pod \"d055d017-5640-4a3d-b684-6ce2755eab74\" (UID: \"d055d017-5640-4a3d-b684-6ce2755eab74\") " Oct 02 02:04:44 crc kubenswrapper[4885]: I1002 02:04:44.281422 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d055d017-5640-4a3d-b684-6ce2755eab74-ovsdbserver-sb\") pod \"d055d017-5640-4a3d-b684-6ce2755eab74\" (UID: \"d055d017-5640-4a3d-b684-6ce2755eab74\") " Oct 02 02:04:44 crc kubenswrapper[4885]: I1002 02:04:44.281512 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d055d017-5640-4a3d-b684-6ce2755eab74-ovsdbserver-nb\") pod \"d055d017-5640-4a3d-b684-6ce2755eab74\" (UID: \"d055d017-5640-4a3d-b684-6ce2755eab74\") " Oct 02 02:04:44 crc kubenswrapper[4885]: I1002 02:04:44.287479 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d055d017-5640-4a3d-b684-6ce2755eab74-kube-api-access-jz9t7" (OuterVolumeSpecName: "kube-api-access-jz9t7") pod "d055d017-5640-4a3d-b684-6ce2755eab74" (UID: "d055d017-5640-4a3d-b684-6ce2755eab74"). InnerVolumeSpecName "kube-api-access-jz9t7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:04:44 crc kubenswrapper[4885]: I1002 02:04:44.315227 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d1811543-6bba-4856-b8e1-814c7592723b","Type":"ContainerStarted","Data":"0e30240269c0fe1f2ef2a216874adfdd1ec919db92613966d00fbd96c5b1ed42"} Oct 02 02:04:44 crc kubenswrapper[4885]: I1002 02:04:44.318446 4885 generic.go:334] "Generic (PLEG): container finished" podID="d055d017-5640-4a3d-b684-6ce2755eab74" containerID="64f15f37e85e8ff6d8c80cc01cf3fed9b5f658a2dccbcf79ffaccdbed931c8d5" exitCode=0 Oct 02 02:04:44 crc kubenswrapper[4885]: I1002 02:04:44.318487 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-x6dgq" event={"ID":"d055d017-5640-4a3d-b684-6ce2755eab74","Type":"ContainerDied","Data":"64f15f37e85e8ff6d8c80cc01cf3fed9b5f658a2dccbcf79ffaccdbed931c8d5"} Oct 02 02:04:44 crc kubenswrapper[4885]: I1002 02:04:44.318518 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-x6dgq" event={"ID":"d055d017-5640-4a3d-b684-6ce2755eab74","Type":"ContainerDied","Data":"7863f6254cf6fea29e7ccfd47a19709da3e70fdd1b07e3f178f579f164208669"} Oct 02 02:04:44 crc kubenswrapper[4885]: I1002 02:04:44.318545 4885 scope.go:117] "RemoveContainer" containerID="64f15f37e85e8ff6d8c80cc01cf3fed9b5f658a2dccbcf79ffaccdbed931c8d5" Oct 02 02:04:44 crc kubenswrapper[4885]: I1002 02:04:44.318712 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-x6dgq" Oct 02 02:04:44 crc kubenswrapper[4885]: I1002 02:04:44.337668 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d055d017-5640-4a3d-b684-6ce2755eab74-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d055d017-5640-4a3d-b684-6ce2755eab74" (UID: "d055d017-5640-4a3d-b684-6ce2755eab74"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:04:44 crc kubenswrapper[4885]: I1002 02:04:44.344499 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d055d017-5640-4a3d-b684-6ce2755eab74-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d055d017-5640-4a3d-b684-6ce2755eab74" (UID: "d055d017-5640-4a3d-b684-6ce2755eab74"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:04:44 crc kubenswrapper[4885]: I1002 02:04:44.346208 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d055d017-5640-4a3d-b684-6ce2755eab74-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "d055d017-5640-4a3d-b684-6ce2755eab74" (UID: "d055d017-5640-4a3d-b684-6ce2755eab74"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:04:44 crc kubenswrapper[4885]: I1002 02:04:44.360625 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d055d017-5640-4a3d-b684-6ce2755eab74-config" (OuterVolumeSpecName: "config") pod "d055d017-5640-4a3d-b684-6ce2755eab74" (UID: "d055d017-5640-4a3d-b684-6ce2755eab74"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:04:44 crc kubenswrapper[4885]: I1002 02:04:44.364447 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d055d017-5640-4a3d-b684-6ce2755eab74-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d055d017-5640-4a3d-b684-6ce2755eab74" (UID: "d055d017-5640-4a3d-b684-6ce2755eab74"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:04:44 crc kubenswrapper[4885]: I1002 02:04:44.383646 4885 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d055d017-5640-4a3d-b684-6ce2755eab74-config\") on node \"crc\" DevicePath \"\"" Oct 02 02:04:44 crc kubenswrapper[4885]: I1002 02:04:44.383680 4885 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d055d017-5640-4a3d-b684-6ce2755eab74-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 02 02:04:44 crc kubenswrapper[4885]: I1002 02:04:44.383692 4885 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d055d017-5640-4a3d-b684-6ce2755eab74-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 02:04:44 crc kubenswrapper[4885]: I1002 02:04:44.383700 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jz9t7\" (UniqueName: \"kubernetes.io/projected/d055d017-5640-4a3d-b684-6ce2755eab74-kube-api-access-jz9t7\") on node \"crc\" DevicePath \"\"" Oct 02 02:04:44 crc kubenswrapper[4885]: I1002 02:04:44.383709 4885 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d055d017-5640-4a3d-b684-6ce2755eab74-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 02:04:44 crc kubenswrapper[4885]: I1002 02:04:44.383717 4885 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d055d017-5640-4a3d-b684-6ce2755eab74-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 02:04:44 crc kubenswrapper[4885]: I1002 02:04:44.430838 4885 scope.go:117] "RemoveContainer" containerID="f10b5d12a6ffdc5055f34cf17444d9b03b28e4b17439d618d94a61c06b28736c" Oct 02 02:04:44 crc kubenswrapper[4885]: I1002 02:04:44.446816 4885 scope.go:117] "RemoveContainer" containerID="64f15f37e85e8ff6d8c80cc01cf3fed9b5f658a2dccbcf79ffaccdbed931c8d5" Oct 02 02:04:44 crc kubenswrapper[4885]: E1002 02:04:44.447330 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"64f15f37e85e8ff6d8c80cc01cf3fed9b5f658a2dccbcf79ffaccdbed931c8d5\": container with ID starting with 64f15f37e85e8ff6d8c80cc01cf3fed9b5f658a2dccbcf79ffaccdbed931c8d5 not found: ID does not exist" containerID="64f15f37e85e8ff6d8c80cc01cf3fed9b5f658a2dccbcf79ffaccdbed931c8d5" Oct 02 02:04:44 crc kubenswrapper[4885]: I1002 02:04:44.447388 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"64f15f37e85e8ff6d8c80cc01cf3fed9b5f658a2dccbcf79ffaccdbed931c8d5"} err="failed to get container status \"64f15f37e85e8ff6d8c80cc01cf3fed9b5f658a2dccbcf79ffaccdbed931c8d5\": rpc error: code = NotFound desc = could not find container \"64f15f37e85e8ff6d8c80cc01cf3fed9b5f658a2dccbcf79ffaccdbed931c8d5\": container with ID starting with 64f15f37e85e8ff6d8c80cc01cf3fed9b5f658a2dccbcf79ffaccdbed931c8d5 not found: ID does not exist" Oct 02 02:04:44 crc kubenswrapper[4885]: I1002 02:04:44.447423 4885 scope.go:117] "RemoveContainer" containerID="f10b5d12a6ffdc5055f34cf17444d9b03b28e4b17439d618d94a61c06b28736c" Oct 02 02:04:44 crc kubenswrapper[4885]: E1002 02:04:44.447827 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f10b5d12a6ffdc5055f34cf17444d9b03b28e4b17439d618d94a61c06b28736c\": container with ID starting with f10b5d12a6ffdc5055f34cf17444d9b03b28e4b17439d618d94a61c06b28736c not found: ID does not exist" containerID="f10b5d12a6ffdc5055f34cf17444d9b03b28e4b17439d618d94a61c06b28736c" Oct 02 02:04:44 crc kubenswrapper[4885]: I1002 02:04:44.447861 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f10b5d12a6ffdc5055f34cf17444d9b03b28e4b17439d618d94a61c06b28736c"} err="failed to get container status \"f10b5d12a6ffdc5055f34cf17444d9b03b28e4b17439d618d94a61c06b28736c\": rpc error: code = NotFound desc = could not find container \"f10b5d12a6ffdc5055f34cf17444d9b03b28e4b17439d618d94a61c06b28736c\": container with ID starting with f10b5d12a6ffdc5055f34cf17444d9b03b28e4b17439d618d94a61c06b28736c not found: ID does not exist" Oct 02 02:04:44 crc kubenswrapper[4885]: I1002 02:04:44.660639 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-x6dgq"] Oct 02 02:04:44 crc kubenswrapper[4885]: I1002 02:04:44.691491 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-x6dgq"] Oct 02 02:04:46 crc kubenswrapper[4885]: I1002 02:04:46.060554 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d055d017-5640-4a3d-b684-6ce2755eab74" path="/var/lib/kubelet/pods/d055d017-5640-4a3d-b684-6ce2755eab74/volumes" Oct 02 02:04:46 crc kubenswrapper[4885]: I1002 02:04:46.343147 4885 generic.go:334] "Generic (PLEG): container finished" podID="b0fddcb3-3649-48f4-aa8f-85dd1c8137cf" containerID="87f582c0838b65442115c622d40cc23454484a73c597b4939299160d6e09873a" exitCode=0 Oct 02 02:04:46 crc kubenswrapper[4885]: I1002 02:04:46.343220 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7c49b4cffb-brdz9" event={"ID":"b0fddcb3-3649-48f4-aa8f-85dd1c8137cf","Type":"ContainerDied","Data":"87f582c0838b65442115c622d40cc23454484a73c597b4939299160d6e09873a"} Oct 02 02:04:47 crc kubenswrapper[4885]: I1002 02:04:47.985917 4885 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-7c49b4cffb-brdz9" podUID="b0fddcb3-3649-48f4-aa8f-85dd1c8137cf" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.146:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.146:8443: connect: connection refused" Oct 02 02:04:49 crc kubenswrapper[4885]: I1002 02:04:49.389416 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d1811543-6bba-4856-b8e1-814c7592723b","Type":"ContainerStarted","Data":"232833a51fcd5052f21baebfd44bdd73a1477e207c7fe5b7f94ecac2710bcbdf"} Oct 02 02:04:53 crc kubenswrapper[4885]: I1002 02:04:53.441926 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d1811543-6bba-4856-b8e1-814c7592723b","Type":"ContainerStarted","Data":"af43c4281d0097fcff46438db9415467ee7c17c8499187886167ff52afb265c4"} Oct 02 02:04:53 crc kubenswrapper[4885]: I1002 02:04:53.442559 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 02 02:04:53 crc kubenswrapper[4885]: I1002 02:04:53.472334 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.465331384 podStartE2EDuration="14.472311056s" podCreationTimestamp="2025-10-02 02:04:39 +0000 UTC" firstStartedPulling="2025-10-02 02:04:40.844106523 +0000 UTC m=+1069.655853922" lastFinishedPulling="2025-10-02 02:04:52.851086195 +0000 UTC m=+1081.662833594" observedRunningTime="2025-10-02 02:04:53.464777038 +0000 UTC m=+1082.276524527" watchObservedRunningTime="2025-10-02 02:04:53.472311056 +0000 UTC m=+1082.284058465" Oct 02 02:04:54 crc kubenswrapper[4885]: I1002 02:04:54.845593 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-59f7bf489d-7tfk4" Oct 02 02:04:57 crc kubenswrapper[4885]: I1002 02:04:57.985200 4885 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-7c49b4cffb-brdz9" podUID="b0fddcb3-3649-48f4-aa8f-85dd1c8137cf" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.146:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.146:8443: connect: connection refused" Oct 02 02:04:59 crc kubenswrapper[4885]: I1002 02:04:59.127160 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 02 02:04:59 crc kubenswrapper[4885]: E1002 02:04:59.127715 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35248aca-4932-408d-a758-766cce4ae517" containerName="horizon" Oct 02 02:04:59 crc kubenswrapper[4885]: I1002 02:04:59.127735 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="35248aca-4932-408d-a758-766cce4ae517" containerName="horizon" Oct 02 02:04:59 crc kubenswrapper[4885]: E1002 02:04:59.127755 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5056f944-c3cf-45e1-99f5-f93132717b93" containerName="horizon" Oct 02 02:04:59 crc kubenswrapper[4885]: I1002 02:04:59.127766 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="5056f944-c3cf-45e1-99f5-f93132717b93" containerName="horizon" Oct 02 02:04:59 crc kubenswrapper[4885]: E1002 02:04:59.127781 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d055d017-5640-4a3d-b684-6ce2755eab74" containerName="init" Oct 02 02:04:59 crc kubenswrapper[4885]: I1002 02:04:59.127788 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="d055d017-5640-4a3d-b684-6ce2755eab74" containerName="init" Oct 02 02:04:59 crc kubenswrapper[4885]: E1002 02:04:59.127809 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d055d017-5640-4a3d-b684-6ce2755eab74" containerName="dnsmasq-dns" Oct 02 02:04:59 crc kubenswrapper[4885]: I1002 02:04:59.127816 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="d055d017-5640-4a3d-b684-6ce2755eab74" containerName="dnsmasq-dns" Oct 02 02:04:59 crc kubenswrapper[4885]: E1002 02:04:59.127824 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ad39937-75ce-468c-90df-b3abbaefe498" containerName="horizon" Oct 02 02:04:59 crc kubenswrapper[4885]: I1002 02:04:59.127829 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ad39937-75ce-468c-90df-b3abbaefe498" containerName="horizon" Oct 02 02:04:59 crc kubenswrapper[4885]: E1002 02:04:59.127847 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ad39937-75ce-468c-90df-b3abbaefe498" containerName="horizon-log" Oct 02 02:04:59 crc kubenswrapper[4885]: I1002 02:04:59.127853 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ad39937-75ce-468c-90df-b3abbaefe498" containerName="horizon-log" Oct 02 02:04:59 crc kubenswrapper[4885]: E1002 02:04:59.127864 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5056f944-c3cf-45e1-99f5-f93132717b93" containerName="horizon-log" Oct 02 02:04:59 crc kubenswrapper[4885]: I1002 02:04:59.127870 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="5056f944-c3cf-45e1-99f5-f93132717b93" containerName="horizon-log" Oct 02 02:04:59 crc kubenswrapper[4885]: E1002 02:04:59.127882 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35248aca-4932-408d-a758-766cce4ae517" containerName="horizon-log" Oct 02 02:04:59 crc kubenswrapper[4885]: I1002 02:04:59.127887 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="35248aca-4932-408d-a758-766cce4ae517" containerName="horizon-log" Oct 02 02:04:59 crc kubenswrapper[4885]: I1002 02:04:59.128056 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="d055d017-5640-4a3d-b684-6ce2755eab74" containerName="dnsmasq-dns" Oct 02 02:04:59 crc kubenswrapper[4885]: I1002 02:04:59.128072 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ad39937-75ce-468c-90df-b3abbaefe498" containerName="horizon-log" Oct 02 02:04:59 crc kubenswrapper[4885]: I1002 02:04:59.128083 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="5056f944-c3cf-45e1-99f5-f93132717b93" containerName="horizon" Oct 02 02:04:59 crc kubenswrapper[4885]: I1002 02:04:59.128096 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ad39937-75ce-468c-90df-b3abbaefe498" containerName="horizon" Oct 02 02:04:59 crc kubenswrapper[4885]: I1002 02:04:59.128112 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="35248aca-4932-408d-a758-766cce4ae517" containerName="horizon-log" Oct 02 02:04:59 crc kubenswrapper[4885]: I1002 02:04:59.128122 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="5056f944-c3cf-45e1-99f5-f93132717b93" containerName="horizon-log" Oct 02 02:04:59 crc kubenswrapper[4885]: I1002 02:04:59.128132 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="35248aca-4932-408d-a758-766cce4ae517" containerName="horizon" Oct 02 02:04:59 crc kubenswrapper[4885]: I1002 02:04:59.128740 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 02 02:04:59 crc kubenswrapper[4885]: I1002 02:04:59.140770 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 02 02:04:59 crc kubenswrapper[4885]: I1002 02:04:59.147286 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Oct 02 02:04:59 crc kubenswrapper[4885]: I1002 02:04:59.147737 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-xqcfb" Oct 02 02:04:59 crc kubenswrapper[4885]: I1002 02:04:59.147831 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Oct 02 02:04:59 crc kubenswrapper[4885]: I1002 02:04:59.212167 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ab251703-d0d6-49cf-bd2a-b74dee9e48f7-openstack-config-secret\") pod \"openstackclient\" (UID: \"ab251703-d0d6-49cf-bd2a-b74dee9e48f7\") " pod="openstack/openstackclient" Oct 02 02:04:59 crc kubenswrapper[4885]: I1002 02:04:59.212592 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab251703-d0d6-49cf-bd2a-b74dee9e48f7-combined-ca-bundle\") pod \"openstackclient\" (UID: \"ab251703-d0d6-49cf-bd2a-b74dee9e48f7\") " pod="openstack/openstackclient" Oct 02 02:04:59 crc kubenswrapper[4885]: I1002 02:04:59.212622 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lqp8p\" (UniqueName: \"kubernetes.io/projected/ab251703-d0d6-49cf-bd2a-b74dee9e48f7-kube-api-access-lqp8p\") pod \"openstackclient\" (UID: \"ab251703-d0d6-49cf-bd2a-b74dee9e48f7\") " pod="openstack/openstackclient" Oct 02 02:04:59 crc kubenswrapper[4885]: I1002 02:04:59.212662 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ab251703-d0d6-49cf-bd2a-b74dee9e48f7-openstack-config\") pod \"openstackclient\" (UID: \"ab251703-d0d6-49cf-bd2a-b74dee9e48f7\") " pod="openstack/openstackclient" Oct 02 02:04:59 crc kubenswrapper[4885]: I1002 02:04:59.314086 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lqp8p\" (UniqueName: \"kubernetes.io/projected/ab251703-d0d6-49cf-bd2a-b74dee9e48f7-kube-api-access-lqp8p\") pod \"openstackclient\" (UID: \"ab251703-d0d6-49cf-bd2a-b74dee9e48f7\") " pod="openstack/openstackclient" Oct 02 02:04:59 crc kubenswrapper[4885]: I1002 02:04:59.314159 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ab251703-d0d6-49cf-bd2a-b74dee9e48f7-openstack-config\") pod \"openstackclient\" (UID: \"ab251703-d0d6-49cf-bd2a-b74dee9e48f7\") " pod="openstack/openstackclient" Oct 02 02:04:59 crc kubenswrapper[4885]: I1002 02:04:59.314221 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ab251703-d0d6-49cf-bd2a-b74dee9e48f7-openstack-config-secret\") pod \"openstackclient\" (UID: \"ab251703-d0d6-49cf-bd2a-b74dee9e48f7\") " pod="openstack/openstackclient" Oct 02 02:04:59 crc kubenswrapper[4885]: I1002 02:04:59.314363 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab251703-d0d6-49cf-bd2a-b74dee9e48f7-combined-ca-bundle\") pod \"openstackclient\" (UID: \"ab251703-d0d6-49cf-bd2a-b74dee9e48f7\") " pod="openstack/openstackclient" Oct 02 02:04:59 crc kubenswrapper[4885]: I1002 02:04:59.315375 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ab251703-d0d6-49cf-bd2a-b74dee9e48f7-openstack-config\") pod \"openstackclient\" (UID: \"ab251703-d0d6-49cf-bd2a-b74dee9e48f7\") " pod="openstack/openstackclient" Oct 02 02:04:59 crc kubenswrapper[4885]: I1002 02:04:59.320767 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab251703-d0d6-49cf-bd2a-b74dee9e48f7-combined-ca-bundle\") pod \"openstackclient\" (UID: \"ab251703-d0d6-49cf-bd2a-b74dee9e48f7\") " pod="openstack/openstackclient" Oct 02 02:04:59 crc kubenswrapper[4885]: I1002 02:04:59.321059 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ab251703-d0d6-49cf-bd2a-b74dee9e48f7-openstack-config-secret\") pod \"openstackclient\" (UID: \"ab251703-d0d6-49cf-bd2a-b74dee9e48f7\") " pod="openstack/openstackclient" Oct 02 02:04:59 crc kubenswrapper[4885]: I1002 02:04:59.344397 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lqp8p\" (UniqueName: \"kubernetes.io/projected/ab251703-d0d6-49cf-bd2a-b74dee9e48f7-kube-api-access-lqp8p\") pod \"openstackclient\" (UID: \"ab251703-d0d6-49cf-bd2a-b74dee9e48f7\") " pod="openstack/openstackclient" Oct 02 02:04:59 crc kubenswrapper[4885]: I1002 02:04:59.479245 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 02 02:05:00 crc kubenswrapper[4885]: I1002 02:05:00.078896 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 02 02:05:00 crc kubenswrapper[4885]: I1002 02:05:00.512950 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"ab251703-d0d6-49cf-bd2a-b74dee9e48f7","Type":"ContainerStarted","Data":"36c117c214ae1eb5642ade77f2a57e0d1cb628e6301ccf51aa5c9075f5b98a6c"} Oct 02 02:05:01 crc kubenswrapper[4885]: I1002 02:05:01.523438 4885 generic.go:334] "Generic (PLEG): container finished" podID="8fa66af9-7b4b-480c-8f96-bda850e2eef6" containerID="8d9aa7b7c3aacfe216138f05551b49ff0afe4a9f79e28ef88f3baf68e620b0fb" exitCode=0 Oct 02 02:05:01 crc kubenswrapper[4885]: I1002 02:05:01.523519 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-vj2tf" event={"ID":"8fa66af9-7b4b-480c-8f96-bda850e2eef6","Type":"ContainerDied","Data":"8d9aa7b7c3aacfe216138f05551b49ff0afe4a9f79e28ef88f3baf68e620b0fb"} Oct 02 02:05:02 crc kubenswrapper[4885]: I1002 02:05:02.909145 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-vj2tf" Oct 02 02:05:03 crc kubenswrapper[4885]: I1002 02:05:03.082475 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8fa66af9-7b4b-480c-8f96-bda850e2eef6-logs\") pod \"8fa66af9-7b4b-480c-8f96-bda850e2eef6\" (UID: \"8fa66af9-7b4b-480c-8f96-bda850e2eef6\") " Oct 02 02:05:03 crc kubenswrapper[4885]: I1002 02:05:03.082540 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8fa66af9-7b4b-480c-8f96-bda850e2eef6-scripts\") pod \"8fa66af9-7b4b-480c-8f96-bda850e2eef6\" (UID: \"8fa66af9-7b4b-480c-8f96-bda850e2eef6\") " Oct 02 02:05:03 crc kubenswrapper[4885]: I1002 02:05:03.082620 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fpmmt\" (UniqueName: \"kubernetes.io/projected/8fa66af9-7b4b-480c-8f96-bda850e2eef6-kube-api-access-fpmmt\") pod \"8fa66af9-7b4b-480c-8f96-bda850e2eef6\" (UID: \"8fa66af9-7b4b-480c-8f96-bda850e2eef6\") " Oct 02 02:05:03 crc kubenswrapper[4885]: I1002 02:05:03.082665 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8fa66af9-7b4b-480c-8f96-bda850e2eef6-combined-ca-bundle\") pod \"8fa66af9-7b4b-480c-8f96-bda850e2eef6\" (UID: \"8fa66af9-7b4b-480c-8f96-bda850e2eef6\") " Oct 02 02:05:03 crc kubenswrapper[4885]: I1002 02:05:03.082713 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8fa66af9-7b4b-480c-8f96-bda850e2eef6-config-data\") pod \"8fa66af9-7b4b-480c-8f96-bda850e2eef6\" (UID: \"8fa66af9-7b4b-480c-8f96-bda850e2eef6\") " Oct 02 02:05:03 crc kubenswrapper[4885]: I1002 02:05:03.083647 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8fa66af9-7b4b-480c-8f96-bda850e2eef6-logs" (OuterVolumeSpecName: "logs") pod "8fa66af9-7b4b-480c-8f96-bda850e2eef6" (UID: "8fa66af9-7b4b-480c-8f96-bda850e2eef6"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:05:03 crc kubenswrapper[4885]: I1002 02:05:03.090970 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8fa66af9-7b4b-480c-8f96-bda850e2eef6-kube-api-access-fpmmt" (OuterVolumeSpecName: "kube-api-access-fpmmt") pod "8fa66af9-7b4b-480c-8f96-bda850e2eef6" (UID: "8fa66af9-7b4b-480c-8f96-bda850e2eef6"). InnerVolumeSpecName "kube-api-access-fpmmt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:05:03 crc kubenswrapper[4885]: I1002 02:05:03.091168 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8fa66af9-7b4b-480c-8f96-bda850e2eef6-scripts" (OuterVolumeSpecName: "scripts") pod "8fa66af9-7b4b-480c-8f96-bda850e2eef6" (UID: "8fa66af9-7b4b-480c-8f96-bda850e2eef6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:05:03 crc kubenswrapper[4885]: I1002 02:05:03.123520 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8fa66af9-7b4b-480c-8f96-bda850e2eef6-config-data" (OuterVolumeSpecName: "config-data") pod "8fa66af9-7b4b-480c-8f96-bda850e2eef6" (UID: "8fa66af9-7b4b-480c-8f96-bda850e2eef6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:05:03 crc kubenswrapper[4885]: I1002 02:05:03.130030 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8fa66af9-7b4b-480c-8f96-bda850e2eef6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8fa66af9-7b4b-480c-8f96-bda850e2eef6" (UID: "8fa66af9-7b4b-480c-8f96-bda850e2eef6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:05:03 crc kubenswrapper[4885]: I1002 02:05:03.184990 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fpmmt\" (UniqueName: \"kubernetes.io/projected/8fa66af9-7b4b-480c-8f96-bda850e2eef6-kube-api-access-fpmmt\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:03 crc kubenswrapper[4885]: I1002 02:05:03.185034 4885 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8fa66af9-7b4b-480c-8f96-bda850e2eef6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:03 crc kubenswrapper[4885]: I1002 02:05:03.185048 4885 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8fa66af9-7b4b-480c-8f96-bda850e2eef6-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:03 crc kubenswrapper[4885]: I1002 02:05:03.185066 4885 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8fa66af9-7b4b-480c-8f96-bda850e2eef6-logs\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:03 crc kubenswrapper[4885]: I1002 02:05:03.185075 4885 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8fa66af9-7b4b-480c-8f96-bda850e2eef6-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:03 crc kubenswrapper[4885]: I1002 02:05:03.546639 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-vj2tf" event={"ID":"8fa66af9-7b4b-480c-8f96-bda850e2eef6","Type":"ContainerDied","Data":"55414b8c8eb0471f8443ae7434426584d5315281a20ef9fd8540ffc929bb4f50"} Oct 02 02:05:03 crc kubenswrapper[4885]: I1002 02:05:03.546685 4885 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="55414b8c8eb0471f8443ae7434426584d5315281a20ef9fd8540ffc929bb4f50" Oct 02 02:05:03 crc kubenswrapper[4885]: I1002 02:05:03.546697 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-vj2tf" Oct 02 02:05:03 crc kubenswrapper[4885]: I1002 02:05:03.619720 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-5d65d69454-8bzbn"] Oct 02 02:05:03 crc kubenswrapper[4885]: E1002 02:05:03.620110 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fa66af9-7b4b-480c-8f96-bda850e2eef6" containerName="placement-db-sync" Oct 02 02:05:03 crc kubenswrapper[4885]: I1002 02:05:03.620126 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fa66af9-7b4b-480c-8f96-bda850e2eef6" containerName="placement-db-sync" Oct 02 02:05:03 crc kubenswrapper[4885]: I1002 02:05:03.620379 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="8fa66af9-7b4b-480c-8f96-bda850e2eef6" containerName="placement-db-sync" Oct 02 02:05:03 crc kubenswrapper[4885]: I1002 02:05:03.621211 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-5d65d69454-8bzbn" Oct 02 02:05:03 crc kubenswrapper[4885]: I1002 02:05:03.625308 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 02 02:05:03 crc kubenswrapper[4885]: I1002 02:05:03.628436 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-k6xjq" Oct 02 02:05:03 crc kubenswrapper[4885]: I1002 02:05:03.628829 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Oct 02 02:05:03 crc kubenswrapper[4885]: I1002 02:05:03.629070 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Oct 02 02:05:03 crc kubenswrapper[4885]: I1002 02:05:03.629333 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 02 02:05:03 crc kubenswrapper[4885]: I1002 02:05:03.648245 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-5d65d69454-8bzbn"] Oct 02 02:05:03 crc kubenswrapper[4885]: I1002 02:05:03.695187 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9bcc69ce-fb1a-4265-a6de-b2e2a10ac733-public-tls-certs\") pod \"placement-5d65d69454-8bzbn\" (UID: \"9bcc69ce-fb1a-4265-a6de-b2e2a10ac733\") " pod="openstack/placement-5d65d69454-8bzbn" Oct 02 02:05:03 crc kubenswrapper[4885]: I1002 02:05:03.695329 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9wb98\" (UniqueName: \"kubernetes.io/projected/9bcc69ce-fb1a-4265-a6de-b2e2a10ac733-kube-api-access-9wb98\") pod \"placement-5d65d69454-8bzbn\" (UID: \"9bcc69ce-fb1a-4265-a6de-b2e2a10ac733\") " pod="openstack/placement-5d65d69454-8bzbn" Oct 02 02:05:03 crc kubenswrapper[4885]: I1002 02:05:03.695378 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9bcc69ce-fb1a-4265-a6de-b2e2a10ac733-config-data\") pod \"placement-5d65d69454-8bzbn\" (UID: \"9bcc69ce-fb1a-4265-a6de-b2e2a10ac733\") " pod="openstack/placement-5d65d69454-8bzbn" Oct 02 02:05:03 crc kubenswrapper[4885]: I1002 02:05:03.695417 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9bcc69ce-fb1a-4265-a6de-b2e2a10ac733-combined-ca-bundle\") pod \"placement-5d65d69454-8bzbn\" (UID: \"9bcc69ce-fb1a-4265-a6de-b2e2a10ac733\") " pod="openstack/placement-5d65d69454-8bzbn" Oct 02 02:05:03 crc kubenswrapper[4885]: I1002 02:05:03.695480 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9bcc69ce-fb1a-4265-a6de-b2e2a10ac733-scripts\") pod \"placement-5d65d69454-8bzbn\" (UID: \"9bcc69ce-fb1a-4265-a6de-b2e2a10ac733\") " pod="openstack/placement-5d65d69454-8bzbn" Oct 02 02:05:03 crc kubenswrapper[4885]: I1002 02:05:03.695511 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9bcc69ce-fb1a-4265-a6de-b2e2a10ac733-internal-tls-certs\") pod \"placement-5d65d69454-8bzbn\" (UID: \"9bcc69ce-fb1a-4265-a6de-b2e2a10ac733\") " pod="openstack/placement-5d65d69454-8bzbn" Oct 02 02:05:03 crc kubenswrapper[4885]: I1002 02:05:03.695535 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9bcc69ce-fb1a-4265-a6de-b2e2a10ac733-logs\") pod \"placement-5d65d69454-8bzbn\" (UID: \"9bcc69ce-fb1a-4265-a6de-b2e2a10ac733\") " pod="openstack/placement-5d65d69454-8bzbn" Oct 02 02:05:03 crc kubenswrapper[4885]: I1002 02:05:03.740702 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-9c5d5c444-vkjrr" Oct 02 02:05:03 crc kubenswrapper[4885]: I1002 02:05:03.796519 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9bcc69ce-fb1a-4265-a6de-b2e2a10ac733-internal-tls-certs\") pod \"placement-5d65d69454-8bzbn\" (UID: \"9bcc69ce-fb1a-4265-a6de-b2e2a10ac733\") " pod="openstack/placement-5d65d69454-8bzbn" Oct 02 02:05:03 crc kubenswrapper[4885]: I1002 02:05:03.796567 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9bcc69ce-fb1a-4265-a6de-b2e2a10ac733-logs\") pod \"placement-5d65d69454-8bzbn\" (UID: \"9bcc69ce-fb1a-4265-a6de-b2e2a10ac733\") " pod="openstack/placement-5d65d69454-8bzbn" Oct 02 02:05:03 crc kubenswrapper[4885]: I1002 02:05:03.796605 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9bcc69ce-fb1a-4265-a6de-b2e2a10ac733-public-tls-certs\") pod \"placement-5d65d69454-8bzbn\" (UID: \"9bcc69ce-fb1a-4265-a6de-b2e2a10ac733\") " pod="openstack/placement-5d65d69454-8bzbn" Oct 02 02:05:03 crc kubenswrapper[4885]: I1002 02:05:03.796667 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9wb98\" (UniqueName: \"kubernetes.io/projected/9bcc69ce-fb1a-4265-a6de-b2e2a10ac733-kube-api-access-9wb98\") pod \"placement-5d65d69454-8bzbn\" (UID: \"9bcc69ce-fb1a-4265-a6de-b2e2a10ac733\") " pod="openstack/placement-5d65d69454-8bzbn" Oct 02 02:05:03 crc kubenswrapper[4885]: I1002 02:05:03.796690 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9bcc69ce-fb1a-4265-a6de-b2e2a10ac733-config-data\") pod \"placement-5d65d69454-8bzbn\" (UID: \"9bcc69ce-fb1a-4265-a6de-b2e2a10ac733\") " pod="openstack/placement-5d65d69454-8bzbn" Oct 02 02:05:03 crc kubenswrapper[4885]: I1002 02:05:03.796705 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9bcc69ce-fb1a-4265-a6de-b2e2a10ac733-combined-ca-bundle\") pod \"placement-5d65d69454-8bzbn\" (UID: \"9bcc69ce-fb1a-4265-a6de-b2e2a10ac733\") " pod="openstack/placement-5d65d69454-8bzbn" Oct 02 02:05:03 crc kubenswrapper[4885]: I1002 02:05:03.796756 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9bcc69ce-fb1a-4265-a6de-b2e2a10ac733-scripts\") pod \"placement-5d65d69454-8bzbn\" (UID: \"9bcc69ce-fb1a-4265-a6de-b2e2a10ac733\") " pod="openstack/placement-5d65d69454-8bzbn" Oct 02 02:05:03 crc kubenswrapper[4885]: I1002 02:05:03.801596 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9bcc69ce-fb1a-4265-a6de-b2e2a10ac733-public-tls-certs\") pod \"placement-5d65d69454-8bzbn\" (UID: \"9bcc69ce-fb1a-4265-a6de-b2e2a10ac733\") " pod="openstack/placement-5d65d69454-8bzbn" Oct 02 02:05:03 crc kubenswrapper[4885]: I1002 02:05:03.802011 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9bcc69ce-fb1a-4265-a6de-b2e2a10ac733-internal-tls-certs\") pod \"placement-5d65d69454-8bzbn\" (UID: \"9bcc69ce-fb1a-4265-a6de-b2e2a10ac733\") " pod="openstack/placement-5d65d69454-8bzbn" Oct 02 02:05:03 crc kubenswrapper[4885]: I1002 02:05:03.803693 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9bcc69ce-fb1a-4265-a6de-b2e2a10ac733-logs\") pod \"placement-5d65d69454-8bzbn\" (UID: \"9bcc69ce-fb1a-4265-a6de-b2e2a10ac733\") " pod="openstack/placement-5d65d69454-8bzbn" Oct 02 02:05:03 crc kubenswrapper[4885]: I1002 02:05:03.803979 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9bcc69ce-fb1a-4265-a6de-b2e2a10ac733-config-data\") pod \"placement-5d65d69454-8bzbn\" (UID: \"9bcc69ce-fb1a-4265-a6de-b2e2a10ac733\") " pod="openstack/placement-5d65d69454-8bzbn" Oct 02 02:05:03 crc kubenswrapper[4885]: I1002 02:05:03.809215 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9bcc69ce-fb1a-4265-a6de-b2e2a10ac733-combined-ca-bundle\") pod \"placement-5d65d69454-8bzbn\" (UID: \"9bcc69ce-fb1a-4265-a6de-b2e2a10ac733\") " pod="openstack/placement-5d65d69454-8bzbn" Oct 02 02:05:03 crc kubenswrapper[4885]: I1002 02:05:03.814469 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9bcc69ce-fb1a-4265-a6de-b2e2a10ac733-scripts\") pod \"placement-5d65d69454-8bzbn\" (UID: \"9bcc69ce-fb1a-4265-a6de-b2e2a10ac733\") " pod="openstack/placement-5d65d69454-8bzbn" Oct 02 02:05:03 crc kubenswrapper[4885]: I1002 02:05:03.818180 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9wb98\" (UniqueName: \"kubernetes.io/projected/9bcc69ce-fb1a-4265-a6de-b2e2a10ac733-kube-api-access-9wb98\") pod \"placement-5d65d69454-8bzbn\" (UID: \"9bcc69ce-fb1a-4265-a6de-b2e2a10ac733\") " pod="openstack/placement-5d65d69454-8bzbn" Oct 02 02:05:03 crc kubenswrapper[4885]: I1002 02:05:03.946402 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-5d65d69454-8bzbn" Oct 02 02:05:04 crc kubenswrapper[4885]: I1002 02:05:04.295485 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-7474b946b7-s2w7j"] Oct 02 02:05:04 crc kubenswrapper[4885]: I1002 02:05:04.298712 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-7474b946b7-s2w7j" Oct 02 02:05:04 crc kubenswrapper[4885]: I1002 02:05:04.301736 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Oct 02 02:05:04 crc kubenswrapper[4885]: I1002 02:05:04.301878 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Oct 02 02:05:04 crc kubenswrapper[4885]: I1002 02:05:04.301736 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Oct 02 02:05:04 crc kubenswrapper[4885]: I1002 02:05:04.316342 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-7474b946b7-s2w7j"] Oct 02 02:05:04 crc kubenswrapper[4885]: I1002 02:05:04.409787 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b22bebc-2f6c-4981-8166-4ccaf19f9727-config-data\") pod \"swift-proxy-7474b946b7-s2w7j\" (UID: \"3b22bebc-2f6c-4981-8166-4ccaf19f9727\") " pod="openstack/swift-proxy-7474b946b7-s2w7j" Oct 02 02:05:04 crc kubenswrapper[4885]: I1002 02:05:04.409895 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3b22bebc-2f6c-4981-8166-4ccaf19f9727-public-tls-certs\") pod \"swift-proxy-7474b946b7-s2w7j\" (UID: \"3b22bebc-2f6c-4981-8166-4ccaf19f9727\") " pod="openstack/swift-proxy-7474b946b7-s2w7j" Oct 02 02:05:04 crc kubenswrapper[4885]: I1002 02:05:04.409936 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3b22bebc-2f6c-4981-8166-4ccaf19f9727-internal-tls-certs\") pod \"swift-proxy-7474b946b7-s2w7j\" (UID: \"3b22bebc-2f6c-4981-8166-4ccaf19f9727\") " pod="openstack/swift-proxy-7474b946b7-s2w7j" Oct 02 02:05:04 crc kubenswrapper[4885]: I1002 02:05:04.410190 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3b22bebc-2f6c-4981-8166-4ccaf19f9727-run-httpd\") pod \"swift-proxy-7474b946b7-s2w7j\" (UID: \"3b22bebc-2f6c-4981-8166-4ccaf19f9727\") " pod="openstack/swift-proxy-7474b946b7-s2w7j" Oct 02 02:05:04 crc kubenswrapper[4885]: I1002 02:05:04.410314 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/3b22bebc-2f6c-4981-8166-4ccaf19f9727-etc-swift\") pod \"swift-proxy-7474b946b7-s2w7j\" (UID: \"3b22bebc-2f6c-4981-8166-4ccaf19f9727\") " pod="openstack/swift-proxy-7474b946b7-s2w7j" Oct 02 02:05:04 crc kubenswrapper[4885]: I1002 02:05:04.410370 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3b22bebc-2f6c-4981-8166-4ccaf19f9727-log-httpd\") pod \"swift-proxy-7474b946b7-s2w7j\" (UID: \"3b22bebc-2f6c-4981-8166-4ccaf19f9727\") " pod="openstack/swift-proxy-7474b946b7-s2w7j" Oct 02 02:05:04 crc kubenswrapper[4885]: I1002 02:05:04.410388 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b22bebc-2f6c-4981-8166-4ccaf19f9727-combined-ca-bundle\") pod \"swift-proxy-7474b946b7-s2w7j\" (UID: \"3b22bebc-2f6c-4981-8166-4ccaf19f9727\") " pod="openstack/swift-proxy-7474b946b7-s2w7j" Oct 02 02:05:04 crc kubenswrapper[4885]: I1002 02:05:04.410439 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tk2db\" (UniqueName: \"kubernetes.io/projected/3b22bebc-2f6c-4981-8166-4ccaf19f9727-kube-api-access-tk2db\") pod \"swift-proxy-7474b946b7-s2w7j\" (UID: \"3b22bebc-2f6c-4981-8166-4ccaf19f9727\") " pod="openstack/swift-proxy-7474b946b7-s2w7j" Oct 02 02:05:04 crc kubenswrapper[4885]: I1002 02:05:04.438373 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-5d65d69454-8bzbn"] Oct 02 02:05:04 crc kubenswrapper[4885]: I1002 02:05:04.512466 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3b22bebc-2f6c-4981-8166-4ccaf19f9727-run-httpd\") pod \"swift-proxy-7474b946b7-s2w7j\" (UID: \"3b22bebc-2f6c-4981-8166-4ccaf19f9727\") " pod="openstack/swift-proxy-7474b946b7-s2w7j" Oct 02 02:05:04 crc kubenswrapper[4885]: I1002 02:05:04.512565 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/3b22bebc-2f6c-4981-8166-4ccaf19f9727-etc-swift\") pod \"swift-proxy-7474b946b7-s2w7j\" (UID: \"3b22bebc-2f6c-4981-8166-4ccaf19f9727\") " pod="openstack/swift-proxy-7474b946b7-s2w7j" Oct 02 02:05:04 crc kubenswrapper[4885]: I1002 02:05:04.512615 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3b22bebc-2f6c-4981-8166-4ccaf19f9727-log-httpd\") pod \"swift-proxy-7474b946b7-s2w7j\" (UID: \"3b22bebc-2f6c-4981-8166-4ccaf19f9727\") " pod="openstack/swift-proxy-7474b946b7-s2w7j" Oct 02 02:05:04 crc kubenswrapper[4885]: I1002 02:05:04.512640 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b22bebc-2f6c-4981-8166-4ccaf19f9727-combined-ca-bundle\") pod \"swift-proxy-7474b946b7-s2w7j\" (UID: \"3b22bebc-2f6c-4981-8166-4ccaf19f9727\") " pod="openstack/swift-proxy-7474b946b7-s2w7j" Oct 02 02:05:04 crc kubenswrapper[4885]: I1002 02:05:04.512667 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tk2db\" (UniqueName: \"kubernetes.io/projected/3b22bebc-2f6c-4981-8166-4ccaf19f9727-kube-api-access-tk2db\") pod \"swift-proxy-7474b946b7-s2w7j\" (UID: \"3b22bebc-2f6c-4981-8166-4ccaf19f9727\") " pod="openstack/swift-proxy-7474b946b7-s2w7j" Oct 02 02:05:04 crc kubenswrapper[4885]: I1002 02:05:04.512708 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b22bebc-2f6c-4981-8166-4ccaf19f9727-config-data\") pod \"swift-proxy-7474b946b7-s2w7j\" (UID: \"3b22bebc-2f6c-4981-8166-4ccaf19f9727\") " pod="openstack/swift-proxy-7474b946b7-s2w7j" Oct 02 02:05:04 crc kubenswrapper[4885]: I1002 02:05:04.512764 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3b22bebc-2f6c-4981-8166-4ccaf19f9727-public-tls-certs\") pod \"swift-proxy-7474b946b7-s2w7j\" (UID: \"3b22bebc-2f6c-4981-8166-4ccaf19f9727\") " pod="openstack/swift-proxy-7474b946b7-s2w7j" Oct 02 02:05:04 crc kubenswrapper[4885]: I1002 02:05:04.512847 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3b22bebc-2f6c-4981-8166-4ccaf19f9727-internal-tls-certs\") pod \"swift-proxy-7474b946b7-s2w7j\" (UID: \"3b22bebc-2f6c-4981-8166-4ccaf19f9727\") " pod="openstack/swift-proxy-7474b946b7-s2w7j" Oct 02 02:05:04 crc kubenswrapper[4885]: I1002 02:05:04.513057 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3b22bebc-2f6c-4981-8166-4ccaf19f9727-run-httpd\") pod \"swift-proxy-7474b946b7-s2w7j\" (UID: \"3b22bebc-2f6c-4981-8166-4ccaf19f9727\") " pod="openstack/swift-proxy-7474b946b7-s2w7j" Oct 02 02:05:04 crc kubenswrapper[4885]: I1002 02:05:04.513728 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3b22bebc-2f6c-4981-8166-4ccaf19f9727-log-httpd\") pod \"swift-proxy-7474b946b7-s2w7j\" (UID: \"3b22bebc-2f6c-4981-8166-4ccaf19f9727\") " pod="openstack/swift-proxy-7474b946b7-s2w7j" Oct 02 02:05:04 crc kubenswrapper[4885]: I1002 02:05:04.517499 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3b22bebc-2f6c-4981-8166-4ccaf19f9727-internal-tls-certs\") pod \"swift-proxy-7474b946b7-s2w7j\" (UID: \"3b22bebc-2f6c-4981-8166-4ccaf19f9727\") " pod="openstack/swift-proxy-7474b946b7-s2w7j" Oct 02 02:05:04 crc kubenswrapper[4885]: I1002 02:05:04.518514 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b22bebc-2f6c-4981-8166-4ccaf19f9727-config-data\") pod \"swift-proxy-7474b946b7-s2w7j\" (UID: \"3b22bebc-2f6c-4981-8166-4ccaf19f9727\") " pod="openstack/swift-proxy-7474b946b7-s2w7j" Oct 02 02:05:04 crc kubenswrapper[4885]: I1002 02:05:04.520054 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b22bebc-2f6c-4981-8166-4ccaf19f9727-combined-ca-bundle\") pod \"swift-proxy-7474b946b7-s2w7j\" (UID: \"3b22bebc-2f6c-4981-8166-4ccaf19f9727\") " pod="openstack/swift-proxy-7474b946b7-s2w7j" Oct 02 02:05:04 crc kubenswrapper[4885]: I1002 02:05:04.520687 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/3b22bebc-2f6c-4981-8166-4ccaf19f9727-etc-swift\") pod \"swift-proxy-7474b946b7-s2w7j\" (UID: \"3b22bebc-2f6c-4981-8166-4ccaf19f9727\") " pod="openstack/swift-proxy-7474b946b7-s2w7j" Oct 02 02:05:04 crc kubenswrapper[4885]: I1002 02:05:04.520750 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3b22bebc-2f6c-4981-8166-4ccaf19f9727-public-tls-certs\") pod \"swift-proxy-7474b946b7-s2w7j\" (UID: \"3b22bebc-2f6c-4981-8166-4ccaf19f9727\") " pod="openstack/swift-proxy-7474b946b7-s2w7j" Oct 02 02:05:04 crc kubenswrapper[4885]: I1002 02:05:04.531839 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tk2db\" (UniqueName: \"kubernetes.io/projected/3b22bebc-2f6c-4981-8166-4ccaf19f9727-kube-api-access-tk2db\") pod \"swift-proxy-7474b946b7-s2w7j\" (UID: \"3b22bebc-2f6c-4981-8166-4ccaf19f9727\") " pod="openstack/swift-proxy-7474b946b7-s2w7j" Oct 02 02:05:04 crc kubenswrapper[4885]: I1002 02:05:04.557194 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5d65d69454-8bzbn" event={"ID":"9bcc69ce-fb1a-4265-a6de-b2e2a10ac733","Type":"ContainerStarted","Data":"22b2c3c0c0821e1242672559267d18aaec2906d4177723f1a150dd5915123515"} Oct 02 02:05:04 crc kubenswrapper[4885]: I1002 02:05:04.618629 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-7474b946b7-s2w7j" Oct 02 02:05:05 crc kubenswrapper[4885]: I1002 02:05:05.320338 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-7474b946b7-s2w7j"] Oct 02 02:05:05 crc kubenswrapper[4885]: I1002 02:05:05.569067 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5d65d69454-8bzbn" event={"ID":"9bcc69ce-fb1a-4265-a6de-b2e2a10ac733","Type":"ContainerStarted","Data":"b730c3a5ff59f8916fd7cabaae66833c501036fdff7c6f13ae03a1442804c7ab"} Oct 02 02:05:06 crc kubenswrapper[4885]: I1002 02:05:06.301219 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-684bd7f7b9-p7lvf" Oct 02 02:05:06 crc kubenswrapper[4885]: I1002 02:05:06.364001 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-9c5d5c444-vkjrr"] Oct 02 02:05:06 crc kubenswrapper[4885]: I1002 02:05:06.364574 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-9c5d5c444-vkjrr" podUID="887b9c10-e0ae-41b2-bfc3-a329f6013d98" containerName="neutron-api" containerID="cri-o://01ca921473539e0109a018c1729a703dbf2f939cc7f6e4997e2af653d9fc5def" gracePeriod=30 Oct 02 02:05:06 crc kubenswrapper[4885]: I1002 02:05:06.364657 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-9c5d5c444-vkjrr" podUID="887b9c10-e0ae-41b2-bfc3-a329f6013d98" containerName="neutron-httpd" containerID="cri-o://772e2dc86bdea03785060468cdc4451b409b9c8bea7dad586dca7ef5e11995a4" gracePeriod=30 Oct 02 02:05:06 crc kubenswrapper[4885]: I1002 02:05:06.380575 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 02:05:06 crc kubenswrapper[4885]: I1002 02:05:06.380821 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d1811543-6bba-4856-b8e1-814c7592723b" containerName="ceilometer-central-agent" containerID="cri-o://698a9529d687dfa8a21be0b32df8bc1601b7907df0cc82a1db5354711128fd20" gracePeriod=30 Oct 02 02:05:06 crc kubenswrapper[4885]: I1002 02:05:06.381453 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d1811543-6bba-4856-b8e1-814c7592723b" containerName="proxy-httpd" containerID="cri-o://af43c4281d0097fcff46438db9415467ee7c17c8499187886167ff52afb265c4" gracePeriod=30 Oct 02 02:05:06 crc kubenswrapper[4885]: I1002 02:05:06.381508 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d1811543-6bba-4856-b8e1-814c7592723b" containerName="sg-core" containerID="cri-o://232833a51fcd5052f21baebfd44bdd73a1477e207c7fe5b7f94ecac2710bcbdf" gracePeriod=30 Oct 02 02:05:06 crc kubenswrapper[4885]: I1002 02:05:06.381542 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d1811543-6bba-4856-b8e1-814c7592723b" containerName="ceilometer-notification-agent" containerID="cri-o://0e30240269c0fe1f2ef2a216874adfdd1ec919db92613966d00fbd96c5b1ed42" gracePeriod=30 Oct 02 02:05:06 crc kubenswrapper[4885]: I1002 02:05:06.394933 4885 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="d1811543-6bba-4856-b8e1-814c7592723b" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.158:3000/\": EOF" Oct 02 02:05:06 crc kubenswrapper[4885]: I1002 02:05:06.584232 4885 generic.go:334] "Generic (PLEG): container finished" podID="887b9c10-e0ae-41b2-bfc3-a329f6013d98" containerID="772e2dc86bdea03785060468cdc4451b409b9c8bea7dad586dca7ef5e11995a4" exitCode=0 Oct 02 02:05:06 crc kubenswrapper[4885]: I1002 02:05:06.584302 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-9c5d5c444-vkjrr" event={"ID":"887b9c10-e0ae-41b2-bfc3-a329f6013d98","Type":"ContainerDied","Data":"772e2dc86bdea03785060468cdc4451b409b9c8bea7dad586dca7ef5e11995a4"} Oct 02 02:05:06 crc kubenswrapper[4885]: I1002 02:05:06.588743 4885 generic.go:334] "Generic (PLEG): container finished" podID="d1811543-6bba-4856-b8e1-814c7592723b" containerID="af43c4281d0097fcff46438db9415467ee7c17c8499187886167ff52afb265c4" exitCode=0 Oct 02 02:05:06 crc kubenswrapper[4885]: I1002 02:05:06.588779 4885 generic.go:334] "Generic (PLEG): container finished" podID="d1811543-6bba-4856-b8e1-814c7592723b" containerID="232833a51fcd5052f21baebfd44bdd73a1477e207c7fe5b7f94ecac2710bcbdf" exitCode=2 Oct 02 02:05:06 crc kubenswrapper[4885]: I1002 02:05:06.588800 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d1811543-6bba-4856-b8e1-814c7592723b","Type":"ContainerDied","Data":"af43c4281d0097fcff46438db9415467ee7c17c8499187886167ff52afb265c4"} Oct 02 02:05:06 crc kubenswrapper[4885]: I1002 02:05:06.588832 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d1811543-6bba-4856-b8e1-814c7592723b","Type":"ContainerDied","Data":"232833a51fcd5052f21baebfd44bdd73a1477e207c7fe5b7f94ecac2710bcbdf"} Oct 02 02:05:07 crc kubenswrapper[4885]: I1002 02:05:07.598980 4885 generic.go:334] "Generic (PLEG): container finished" podID="d1811543-6bba-4856-b8e1-814c7592723b" containerID="698a9529d687dfa8a21be0b32df8bc1601b7907df0cc82a1db5354711128fd20" exitCode=0 Oct 02 02:05:07 crc kubenswrapper[4885]: I1002 02:05:07.599027 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d1811543-6bba-4856-b8e1-814c7592723b","Type":"ContainerDied","Data":"698a9529d687dfa8a21be0b32df8bc1601b7907df0cc82a1db5354711128fd20"} Oct 02 02:05:07 crc kubenswrapper[4885]: I1002 02:05:07.985109 4885 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-7c49b4cffb-brdz9" podUID="b0fddcb3-3649-48f4-aa8f-85dd1c8137cf" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.146:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.146:8443: connect: connection refused" Oct 02 02:05:08 crc kubenswrapper[4885]: I1002 02:05:08.611242 4885 generic.go:334] "Generic (PLEG): container finished" podID="887b9c10-e0ae-41b2-bfc3-a329f6013d98" containerID="01ca921473539e0109a018c1729a703dbf2f939cc7f6e4997e2af653d9fc5def" exitCode=0 Oct 02 02:05:08 crc kubenswrapper[4885]: I1002 02:05:08.611309 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-9c5d5c444-vkjrr" event={"ID":"887b9c10-e0ae-41b2-bfc3-a329f6013d98","Type":"ContainerDied","Data":"01ca921473539e0109a018c1729a703dbf2f939cc7f6e4997e2af653d9fc5def"} Oct 02 02:05:09 crc kubenswrapper[4885]: I1002 02:05:09.626046 4885 generic.go:334] "Generic (PLEG): container finished" podID="d1811543-6bba-4856-b8e1-814c7592723b" containerID="0e30240269c0fe1f2ef2a216874adfdd1ec919db92613966d00fbd96c5b1ed42" exitCode=0 Oct 02 02:05:09 crc kubenswrapper[4885]: I1002 02:05:09.626116 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d1811543-6bba-4856-b8e1-814c7592723b","Type":"ContainerDied","Data":"0e30240269c0fe1f2ef2a216874adfdd1ec919db92613966d00fbd96c5b1ed42"} Oct 02 02:05:09 crc kubenswrapper[4885]: I1002 02:05:09.628684 4885 generic.go:334] "Generic (PLEG): container finished" podID="0feba5bf-20c5-4578-9cfc-f37d6511c2d6" containerID="0e6ab596da7147218797e106083b8cfd31507cec30d5f1645bcc1a5f69320aac" exitCode=0 Oct 02 02:05:09 crc kubenswrapper[4885]: I1002 02:05:09.628715 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-gp9st" event={"ID":"0feba5bf-20c5-4578-9cfc-f37d6511c2d6","Type":"ContainerDied","Data":"0e6ab596da7147218797e106083b8cfd31507cec30d5f1645bcc1a5f69320aac"} Oct 02 02:05:10 crc kubenswrapper[4885]: I1002 02:05:10.157307 4885 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="d1811543-6bba-4856-b8e1-814c7592723b" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.158:3000/\": dial tcp 10.217.0.158:3000: connect: connection refused" Oct 02 02:05:10 crc kubenswrapper[4885]: W1002 02:05:10.288238 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3b22bebc_2f6c_4981_8166_4ccaf19f9727.slice/crio-71d945de2dfe6a862a90b5416e46fcfd45fe47869e1985db3d8f25a5459870cc WatchSource:0}: Error finding container 71d945de2dfe6a862a90b5416e46fcfd45fe47869e1985db3d8f25a5459870cc: Status 404 returned error can't find the container with id 71d945de2dfe6a862a90b5416e46fcfd45fe47869e1985db3d8f25a5459870cc Oct 02 02:05:10 crc kubenswrapper[4885]: I1002 02:05:10.682027 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 02:05:10 crc kubenswrapper[4885]: I1002 02:05:10.697827 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-7474b946b7-s2w7j" event={"ID":"3b22bebc-2f6c-4981-8166-4ccaf19f9727","Type":"ContainerStarted","Data":"71d945de2dfe6a862a90b5416e46fcfd45fe47869e1985db3d8f25a5459870cc"} Oct 02 02:05:10 crc kubenswrapper[4885]: I1002 02:05:10.734252 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-9c5d5c444-vkjrr" Oct 02 02:05:10 crc kubenswrapper[4885]: I1002 02:05:10.737195 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 02:05:10 crc kubenswrapper[4885]: I1002 02:05:10.737191 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d1811543-6bba-4856-b8e1-814c7592723b","Type":"ContainerDied","Data":"30faa88c919983f7626a53ca812e2ca526eae85993762541b5a6fd6c39e4db3f"} Oct 02 02:05:10 crc kubenswrapper[4885]: I1002 02:05:10.739435 4885 scope.go:117] "RemoveContainer" containerID="af43c4281d0097fcff46438db9415467ee7c17c8499187886167ff52afb265c4" Oct 02 02:05:10 crc kubenswrapper[4885]: I1002 02:05:10.747629 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-5d65d69454-8bzbn" Oct 02 02:05:10 crc kubenswrapper[4885]: I1002 02:05:10.748140 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-5d65d69454-8bzbn" Oct 02 02:05:10 crc kubenswrapper[4885]: I1002 02:05:10.752164 4885 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/placement-5d65d69454-8bzbn" podUID="9bcc69ce-fb1a-4265-a6de-b2e2a10ac733" containerName="placement-log" probeResult="failure" output="Get \"https://10.217.0.160:8778/\": dial tcp 10.217.0.160:8778: connect: connection refused" Oct 02 02:05:10 crc kubenswrapper[4885]: I1002 02:05:10.791117 4885 scope.go:117] "RemoveContainer" containerID="232833a51fcd5052f21baebfd44bdd73a1477e207c7fe5b7f94ecac2710bcbdf" Oct 02 02:05:10 crc kubenswrapper[4885]: I1002 02:05:10.792695 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-5d65d69454-8bzbn" podStartSLOduration=7.792674107 podStartE2EDuration="7.792674107s" podCreationTimestamp="2025-10-02 02:05:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:05:10.779369915 +0000 UTC m=+1099.591117314" watchObservedRunningTime="2025-10-02 02:05:10.792674107 +0000 UTC m=+1099.604421506" Oct 02 02:05:10 crc kubenswrapper[4885]: I1002 02:05:10.809484 4885 scope.go:117] "RemoveContainer" containerID="0e30240269c0fe1f2ef2a216874adfdd1ec919db92613966d00fbd96c5b1ed42" Oct 02 02:05:10 crc kubenswrapper[4885]: I1002 02:05:10.832462 4885 scope.go:117] "RemoveContainer" containerID="698a9529d687dfa8a21be0b32df8bc1601b7907df0cc82a1db5354711128fd20" Oct 02 02:05:10 crc kubenswrapper[4885]: I1002 02:05:10.835802 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1811543-6bba-4856-b8e1-814c7592723b-config-data\") pod \"d1811543-6bba-4856-b8e1-814c7592723b\" (UID: \"d1811543-6bba-4856-b8e1-814c7592723b\") " Oct 02 02:05:10 crc kubenswrapper[4885]: I1002 02:05:10.835926 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d1811543-6bba-4856-b8e1-814c7592723b-log-httpd\") pod \"d1811543-6bba-4856-b8e1-814c7592723b\" (UID: \"d1811543-6bba-4856-b8e1-814c7592723b\") " Oct 02 02:05:10 crc kubenswrapper[4885]: I1002 02:05:10.835994 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/887b9c10-e0ae-41b2-bfc3-a329f6013d98-ovndb-tls-certs\") pod \"887b9c10-e0ae-41b2-bfc3-a329f6013d98\" (UID: \"887b9c10-e0ae-41b2-bfc3-a329f6013d98\") " Oct 02 02:05:10 crc kubenswrapper[4885]: I1002 02:05:10.836028 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/887b9c10-e0ae-41b2-bfc3-a329f6013d98-httpd-config\") pod \"887b9c10-e0ae-41b2-bfc3-a329f6013d98\" (UID: \"887b9c10-e0ae-41b2-bfc3-a329f6013d98\") " Oct 02 02:05:10 crc kubenswrapper[4885]: I1002 02:05:10.836046 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1811543-6bba-4856-b8e1-814c7592723b-combined-ca-bundle\") pod \"d1811543-6bba-4856-b8e1-814c7592723b\" (UID: \"d1811543-6bba-4856-b8e1-814c7592723b\") " Oct 02 02:05:10 crc kubenswrapper[4885]: I1002 02:05:10.836070 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nj4k4\" (UniqueName: \"kubernetes.io/projected/d1811543-6bba-4856-b8e1-814c7592723b-kube-api-access-nj4k4\") pod \"d1811543-6bba-4856-b8e1-814c7592723b\" (UID: \"d1811543-6bba-4856-b8e1-814c7592723b\") " Oct 02 02:05:10 crc kubenswrapper[4885]: I1002 02:05:10.836116 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sh2v7\" (UniqueName: \"kubernetes.io/projected/887b9c10-e0ae-41b2-bfc3-a329f6013d98-kube-api-access-sh2v7\") pod \"887b9c10-e0ae-41b2-bfc3-a329f6013d98\" (UID: \"887b9c10-e0ae-41b2-bfc3-a329f6013d98\") " Oct 02 02:05:10 crc kubenswrapper[4885]: I1002 02:05:10.836147 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d1811543-6bba-4856-b8e1-814c7592723b-sg-core-conf-yaml\") pod \"d1811543-6bba-4856-b8e1-814c7592723b\" (UID: \"d1811543-6bba-4856-b8e1-814c7592723b\") " Oct 02 02:05:10 crc kubenswrapper[4885]: I1002 02:05:10.836181 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/887b9c10-e0ae-41b2-bfc3-a329f6013d98-combined-ca-bundle\") pod \"887b9c10-e0ae-41b2-bfc3-a329f6013d98\" (UID: \"887b9c10-e0ae-41b2-bfc3-a329f6013d98\") " Oct 02 02:05:10 crc kubenswrapper[4885]: I1002 02:05:10.836206 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/887b9c10-e0ae-41b2-bfc3-a329f6013d98-config\") pod \"887b9c10-e0ae-41b2-bfc3-a329f6013d98\" (UID: \"887b9c10-e0ae-41b2-bfc3-a329f6013d98\") " Oct 02 02:05:10 crc kubenswrapper[4885]: I1002 02:05:10.836230 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d1811543-6bba-4856-b8e1-814c7592723b-scripts\") pod \"d1811543-6bba-4856-b8e1-814c7592723b\" (UID: \"d1811543-6bba-4856-b8e1-814c7592723b\") " Oct 02 02:05:10 crc kubenswrapper[4885]: I1002 02:05:10.836250 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d1811543-6bba-4856-b8e1-814c7592723b-run-httpd\") pod \"d1811543-6bba-4856-b8e1-814c7592723b\" (UID: \"d1811543-6bba-4856-b8e1-814c7592723b\") " Oct 02 02:05:10 crc kubenswrapper[4885]: I1002 02:05:10.837742 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d1811543-6bba-4856-b8e1-814c7592723b-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "d1811543-6bba-4856-b8e1-814c7592723b" (UID: "d1811543-6bba-4856-b8e1-814c7592723b"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:05:10 crc kubenswrapper[4885]: I1002 02:05:10.839998 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/887b9c10-e0ae-41b2-bfc3-a329f6013d98-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "887b9c10-e0ae-41b2-bfc3-a329f6013d98" (UID: "887b9c10-e0ae-41b2-bfc3-a329f6013d98"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:05:10 crc kubenswrapper[4885]: I1002 02:05:10.841970 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d1811543-6bba-4856-b8e1-814c7592723b-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "d1811543-6bba-4856-b8e1-814c7592723b" (UID: "d1811543-6bba-4856-b8e1-814c7592723b"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:05:10 crc kubenswrapper[4885]: I1002 02:05:10.857886 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d1811543-6bba-4856-b8e1-814c7592723b-kube-api-access-nj4k4" (OuterVolumeSpecName: "kube-api-access-nj4k4") pod "d1811543-6bba-4856-b8e1-814c7592723b" (UID: "d1811543-6bba-4856-b8e1-814c7592723b"). InnerVolumeSpecName "kube-api-access-nj4k4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:05:10 crc kubenswrapper[4885]: I1002 02:05:10.857932 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/887b9c10-e0ae-41b2-bfc3-a329f6013d98-kube-api-access-sh2v7" (OuterVolumeSpecName: "kube-api-access-sh2v7") pod "887b9c10-e0ae-41b2-bfc3-a329f6013d98" (UID: "887b9c10-e0ae-41b2-bfc3-a329f6013d98"). InnerVolumeSpecName "kube-api-access-sh2v7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:05:10 crc kubenswrapper[4885]: I1002 02:05:10.861802 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1811543-6bba-4856-b8e1-814c7592723b-scripts" (OuterVolumeSpecName: "scripts") pod "d1811543-6bba-4856-b8e1-814c7592723b" (UID: "d1811543-6bba-4856-b8e1-814c7592723b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:05:10 crc kubenswrapper[4885]: I1002 02:05:10.895211 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1811543-6bba-4856-b8e1-814c7592723b-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "d1811543-6bba-4856-b8e1-814c7592723b" (UID: "d1811543-6bba-4856-b8e1-814c7592723b"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:05:10 crc kubenswrapper[4885]: I1002 02:05:10.921157 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/887b9c10-e0ae-41b2-bfc3-a329f6013d98-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "887b9c10-e0ae-41b2-bfc3-a329f6013d98" (UID: "887b9c10-e0ae-41b2-bfc3-a329f6013d98"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:05:10 crc kubenswrapper[4885]: I1002 02:05:10.940224 4885 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/887b9c10-e0ae-41b2-bfc3-a329f6013d98-httpd-config\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:10 crc kubenswrapper[4885]: I1002 02:05:10.940502 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nj4k4\" (UniqueName: \"kubernetes.io/projected/d1811543-6bba-4856-b8e1-814c7592723b-kube-api-access-nj4k4\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:10 crc kubenswrapper[4885]: I1002 02:05:10.940597 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sh2v7\" (UniqueName: \"kubernetes.io/projected/887b9c10-e0ae-41b2-bfc3-a329f6013d98-kube-api-access-sh2v7\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:10 crc kubenswrapper[4885]: I1002 02:05:10.940677 4885 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d1811543-6bba-4856-b8e1-814c7592723b-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:10 crc kubenswrapper[4885]: I1002 02:05:10.940789 4885 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/887b9c10-e0ae-41b2-bfc3-a329f6013d98-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:10 crc kubenswrapper[4885]: I1002 02:05:10.940867 4885 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d1811543-6bba-4856-b8e1-814c7592723b-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:10 crc kubenswrapper[4885]: I1002 02:05:10.940941 4885 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d1811543-6bba-4856-b8e1-814c7592723b-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:10 crc kubenswrapper[4885]: I1002 02:05:10.941013 4885 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d1811543-6bba-4856-b8e1-814c7592723b-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:10 crc kubenswrapper[4885]: I1002 02:05:10.955001 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/887b9c10-e0ae-41b2-bfc3-a329f6013d98-config" (OuterVolumeSpecName: "config") pod "887b9c10-e0ae-41b2-bfc3-a329f6013d98" (UID: "887b9c10-e0ae-41b2-bfc3-a329f6013d98"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:05:10 crc kubenswrapper[4885]: I1002 02:05:10.972121 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1811543-6bba-4856-b8e1-814c7592723b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d1811543-6bba-4856-b8e1-814c7592723b" (UID: "d1811543-6bba-4856-b8e1-814c7592723b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:05:10 crc kubenswrapper[4885]: I1002 02:05:10.985393 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/887b9c10-e0ae-41b2-bfc3-a329f6013d98-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "887b9c10-e0ae-41b2-bfc3-a329f6013d98" (UID: "887b9c10-e0ae-41b2-bfc3-a329f6013d98"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:05:10 crc kubenswrapper[4885]: I1002 02:05:10.991198 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1811543-6bba-4856-b8e1-814c7592723b-config-data" (OuterVolumeSpecName: "config-data") pod "d1811543-6bba-4856-b8e1-814c7592723b" (UID: "d1811543-6bba-4856-b8e1-814c7592723b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:05:11 crc kubenswrapper[4885]: I1002 02:05:11.030536 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-gp9st" Oct 02 02:05:11 crc kubenswrapper[4885]: I1002 02:05:11.043333 4885 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/887b9c10-e0ae-41b2-bfc3-a329f6013d98-config\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:11 crc kubenswrapper[4885]: I1002 02:05:11.043371 4885 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1811543-6bba-4856-b8e1-814c7592723b-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:11 crc kubenswrapper[4885]: I1002 02:05:11.043380 4885 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/887b9c10-e0ae-41b2-bfc3-a329f6013d98-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:11 crc kubenswrapper[4885]: I1002 02:05:11.043389 4885 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1811543-6bba-4856-b8e1-814c7592723b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:11 crc kubenswrapper[4885]: I1002 02:05:11.146673 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0feba5bf-20c5-4578-9cfc-f37d6511c2d6-db-sync-config-data\") pod \"0feba5bf-20c5-4578-9cfc-f37d6511c2d6\" (UID: \"0feba5bf-20c5-4578-9cfc-f37d6511c2d6\") " Oct 02 02:05:11 crc kubenswrapper[4885]: I1002 02:05:11.146745 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tf2lz\" (UniqueName: \"kubernetes.io/projected/0feba5bf-20c5-4578-9cfc-f37d6511c2d6-kube-api-access-tf2lz\") pod \"0feba5bf-20c5-4578-9cfc-f37d6511c2d6\" (UID: \"0feba5bf-20c5-4578-9cfc-f37d6511c2d6\") " Oct 02 02:05:11 crc kubenswrapper[4885]: I1002 02:05:11.146833 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0feba5bf-20c5-4578-9cfc-f37d6511c2d6-combined-ca-bundle\") pod \"0feba5bf-20c5-4578-9cfc-f37d6511c2d6\" (UID: \"0feba5bf-20c5-4578-9cfc-f37d6511c2d6\") " Oct 02 02:05:11 crc kubenswrapper[4885]: I1002 02:05:11.166768 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0feba5bf-20c5-4578-9cfc-f37d6511c2d6-kube-api-access-tf2lz" (OuterVolumeSpecName: "kube-api-access-tf2lz") pod "0feba5bf-20c5-4578-9cfc-f37d6511c2d6" (UID: "0feba5bf-20c5-4578-9cfc-f37d6511c2d6"). InnerVolumeSpecName "kube-api-access-tf2lz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:05:11 crc kubenswrapper[4885]: I1002 02:05:11.171124 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0feba5bf-20c5-4578-9cfc-f37d6511c2d6-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "0feba5bf-20c5-4578-9cfc-f37d6511c2d6" (UID: "0feba5bf-20c5-4578-9cfc-f37d6511c2d6"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:05:11 crc kubenswrapper[4885]: I1002 02:05:11.186405 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 02:05:11 crc kubenswrapper[4885]: I1002 02:05:11.243140 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 02 02:05:11 crc kubenswrapper[4885]: I1002 02:05:11.251385 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 02 02:05:11 crc kubenswrapper[4885]: E1002 02:05:11.251757 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0feba5bf-20c5-4578-9cfc-f37d6511c2d6" containerName="barbican-db-sync" Oct 02 02:05:11 crc kubenswrapper[4885]: I1002 02:05:11.251775 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="0feba5bf-20c5-4578-9cfc-f37d6511c2d6" containerName="barbican-db-sync" Oct 02 02:05:11 crc kubenswrapper[4885]: E1002 02:05:11.251796 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1811543-6bba-4856-b8e1-814c7592723b" containerName="sg-core" Oct 02 02:05:11 crc kubenswrapper[4885]: I1002 02:05:11.251803 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1811543-6bba-4856-b8e1-814c7592723b" containerName="sg-core" Oct 02 02:05:11 crc kubenswrapper[4885]: E1002 02:05:11.251820 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="887b9c10-e0ae-41b2-bfc3-a329f6013d98" containerName="neutron-httpd" Oct 02 02:05:11 crc kubenswrapper[4885]: I1002 02:05:11.251825 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="887b9c10-e0ae-41b2-bfc3-a329f6013d98" containerName="neutron-httpd" Oct 02 02:05:11 crc kubenswrapper[4885]: E1002 02:05:11.251840 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1811543-6bba-4856-b8e1-814c7592723b" containerName="ceilometer-central-agent" Oct 02 02:05:11 crc kubenswrapper[4885]: I1002 02:05:11.251845 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1811543-6bba-4856-b8e1-814c7592723b" containerName="ceilometer-central-agent" Oct 02 02:05:11 crc kubenswrapper[4885]: E1002 02:05:11.251855 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="887b9c10-e0ae-41b2-bfc3-a329f6013d98" containerName="neutron-api" Oct 02 02:05:11 crc kubenswrapper[4885]: I1002 02:05:11.251860 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="887b9c10-e0ae-41b2-bfc3-a329f6013d98" containerName="neutron-api" Oct 02 02:05:11 crc kubenswrapper[4885]: E1002 02:05:11.251880 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1811543-6bba-4856-b8e1-814c7592723b" containerName="ceilometer-notification-agent" Oct 02 02:05:11 crc kubenswrapper[4885]: I1002 02:05:11.251886 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1811543-6bba-4856-b8e1-814c7592723b" containerName="ceilometer-notification-agent" Oct 02 02:05:11 crc kubenswrapper[4885]: E1002 02:05:11.251892 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1811543-6bba-4856-b8e1-814c7592723b" containerName="proxy-httpd" Oct 02 02:05:11 crc kubenswrapper[4885]: I1002 02:05:11.251898 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1811543-6bba-4856-b8e1-814c7592723b" containerName="proxy-httpd" Oct 02 02:05:11 crc kubenswrapper[4885]: I1002 02:05:11.252070 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1811543-6bba-4856-b8e1-814c7592723b" containerName="sg-core" Oct 02 02:05:11 crc kubenswrapper[4885]: I1002 02:05:11.252080 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="0feba5bf-20c5-4578-9cfc-f37d6511c2d6" containerName="barbican-db-sync" Oct 02 02:05:11 crc kubenswrapper[4885]: I1002 02:05:11.252089 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1811543-6bba-4856-b8e1-814c7592723b" containerName="proxy-httpd" Oct 02 02:05:11 crc kubenswrapper[4885]: I1002 02:05:11.252102 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1811543-6bba-4856-b8e1-814c7592723b" containerName="ceilometer-notification-agent" Oct 02 02:05:11 crc kubenswrapper[4885]: I1002 02:05:11.252114 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="887b9c10-e0ae-41b2-bfc3-a329f6013d98" containerName="neutron-httpd" Oct 02 02:05:11 crc kubenswrapper[4885]: I1002 02:05:11.252126 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="887b9c10-e0ae-41b2-bfc3-a329f6013d98" containerName="neutron-api" Oct 02 02:05:11 crc kubenswrapper[4885]: I1002 02:05:11.252138 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1811543-6bba-4856-b8e1-814c7592723b" containerName="ceilometer-central-agent" Oct 02 02:05:11 crc kubenswrapper[4885]: I1002 02:05:11.253395 4885 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0feba5bf-20c5-4578-9cfc-f37d6511c2d6-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:11 crc kubenswrapper[4885]: I1002 02:05:11.253455 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tf2lz\" (UniqueName: \"kubernetes.io/projected/0feba5bf-20c5-4578-9cfc-f37d6511c2d6-kube-api-access-tf2lz\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:11 crc kubenswrapper[4885]: I1002 02:05:11.253737 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 02:05:11 crc kubenswrapper[4885]: I1002 02:05:11.258481 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0feba5bf-20c5-4578-9cfc-f37d6511c2d6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0feba5bf-20c5-4578-9cfc-f37d6511c2d6" (UID: "0feba5bf-20c5-4578-9cfc-f37d6511c2d6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:05:11 crc kubenswrapper[4885]: I1002 02:05:11.258633 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 02 02:05:11 crc kubenswrapper[4885]: I1002 02:05:11.258874 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 02 02:05:11 crc kubenswrapper[4885]: I1002 02:05:11.260221 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 02:05:11 crc kubenswrapper[4885]: I1002 02:05:11.354522 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9df73c9d-22c1-47c0-a996-0983f206dad5-run-httpd\") pod \"ceilometer-0\" (UID: \"9df73c9d-22c1-47c0-a996-0983f206dad5\") " pod="openstack/ceilometer-0" Oct 02 02:05:11 crc kubenswrapper[4885]: I1002 02:05:11.354655 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9df73c9d-22c1-47c0-a996-0983f206dad5-scripts\") pod \"ceilometer-0\" (UID: \"9df73c9d-22c1-47c0-a996-0983f206dad5\") " pod="openstack/ceilometer-0" Oct 02 02:05:11 crc kubenswrapper[4885]: I1002 02:05:11.354688 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9df73c9d-22c1-47c0-a996-0983f206dad5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9df73c9d-22c1-47c0-a996-0983f206dad5\") " pod="openstack/ceilometer-0" Oct 02 02:05:11 crc kubenswrapper[4885]: I1002 02:05:11.354729 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9df73c9d-22c1-47c0-a996-0983f206dad5-log-httpd\") pod \"ceilometer-0\" (UID: \"9df73c9d-22c1-47c0-a996-0983f206dad5\") " pod="openstack/ceilometer-0" Oct 02 02:05:11 crc kubenswrapper[4885]: I1002 02:05:11.354745 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6jr2d\" (UniqueName: \"kubernetes.io/projected/9df73c9d-22c1-47c0-a996-0983f206dad5-kube-api-access-6jr2d\") pod \"ceilometer-0\" (UID: \"9df73c9d-22c1-47c0-a996-0983f206dad5\") " pod="openstack/ceilometer-0" Oct 02 02:05:11 crc kubenswrapper[4885]: I1002 02:05:11.354792 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9df73c9d-22c1-47c0-a996-0983f206dad5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9df73c9d-22c1-47c0-a996-0983f206dad5\") " pod="openstack/ceilometer-0" Oct 02 02:05:11 crc kubenswrapper[4885]: I1002 02:05:11.354865 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9df73c9d-22c1-47c0-a996-0983f206dad5-config-data\") pod \"ceilometer-0\" (UID: \"9df73c9d-22c1-47c0-a996-0983f206dad5\") " pod="openstack/ceilometer-0" Oct 02 02:05:11 crc kubenswrapper[4885]: I1002 02:05:11.354921 4885 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0feba5bf-20c5-4578-9cfc-f37d6511c2d6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:11 crc kubenswrapper[4885]: I1002 02:05:11.456758 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9df73c9d-22c1-47c0-a996-0983f206dad5-log-httpd\") pod \"ceilometer-0\" (UID: \"9df73c9d-22c1-47c0-a996-0983f206dad5\") " pod="openstack/ceilometer-0" Oct 02 02:05:11 crc kubenswrapper[4885]: I1002 02:05:11.456796 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6jr2d\" (UniqueName: \"kubernetes.io/projected/9df73c9d-22c1-47c0-a996-0983f206dad5-kube-api-access-6jr2d\") pod \"ceilometer-0\" (UID: \"9df73c9d-22c1-47c0-a996-0983f206dad5\") " pod="openstack/ceilometer-0" Oct 02 02:05:11 crc kubenswrapper[4885]: I1002 02:05:11.456838 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9df73c9d-22c1-47c0-a996-0983f206dad5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9df73c9d-22c1-47c0-a996-0983f206dad5\") " pod="openstack/ceilometer-0" Oct 02 02:05:11 crc kubenswrapper[4885]: I1002 02:05:11.456892 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9df73c9d-22c1-47c0-a996-0983f206dad5-config-data\") pod \"ceilometer-0\" (UID: \"9df73c9d-22c1-47c0-a996-0983f206dad5\") " pod="openstack/ceilometer-0" Oct 02 02:05:11 crc kubenswrapper[4885]: I1002 02:05:11.456924 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9df73c9d-22c1-47c0-a996-0983f206dad5-run-httpd\") pod \"ceilometer-0\" (UID: \"9df73c9d-22c1-47c0-a996-0983f206dad5\") " pod="openstack/ceilometer-0" Oct 02 02:05:11 crc kubenswrapper[4885]: I1002 02:05:11.456977 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9df73c9d-22c1-47c0-a996-0983f206dad5-scripts\") pod \"ceilometer-0\" (UID: \"9df73c9d-22c1-47c0-a996-0983f206dad5\") " pod="openstack/ceilometer-0" Oct 02 02:05:11 crc kubenswrapper[4885]: I1002 02:05:11.457000 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9df73c9d-22c1-47c0-a996-0983f206dad5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9df73c9d-22c1-47c0-a996-0983f206dad5\") " pod="openstack/ceilometer-0" Oct 02 02:05:11 crc kubenswrapper[4885]: I1002 02:05:11.457405 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9df73c9d-22c1-47c0-a996-0983f206dad5-run-httpd\") pod \"ceilometer-0\" (UID: \"9df73c9d-22c1-47c0-a996-0983f206dad5\") " pod="openstack/ceilometer-0" Oct 02 02:05:11 crc kubenswrapper[4885]: I1002 02:05:11.457564 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9df73c9d-22c1-47c0-a996-0983f206dad5-log-httpd\") pod \"ceilometer-0\" (UID: \"9df73c9d-22c1-47c0-a996-0983f206dad5\") " pod="openstack/ceilometer-0" Oct 02 02:05:11 crc kubenswrapper[4885]: I1002 02:05:11.461029 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9df73c9d-22c1-47c0-a996-0983f206dad5-scripts\") pod \"ceilometer-0\" (UID: \"9df73c9d-22c1-47c0-a996-0983f206dad5\") " pod="openstack/ceilometer-0" Oct 02 02:05:11 crc kubenswrapper[4885]: I1002 02:05:11.462797 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9df73c9d-22c1-47c0-a996-0983f206dad5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9df73c9d-22c1-47c0-a996-0983f206dad5\") " pod="openstack/ceilometer-0" Oct 02 02:05:11 crc kubenswrapper[4885]: I1002 02:05:11.463391 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9df73c9d-22c1-47c0-a996-0983f206dad5-config-data\") pod \"ceilometer-0\" (UID: \"9df73c9d-22c1-47c0-a996-0983f206dad5\") " pod="openstack/ceilometer-0" Oct 02 02:05:11 crc kubenswrapper[4885]: I1002 02:05:11.466561 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9df73c9d-22c1-47c0-a996-0983f206dad5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9df73c9d-22c1-47c0-a996-0983f206dad5\") " pod="openstack/ceilometer-0" Oct 02 02:05:11 crc kubenswrapper[4885]: I1002 02:05:11.474301 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6jr2d\" (UniqueName: \"kubernetes.io/projected/9df73c9d-22c1-47c0-a996-0983f206dad5-kube-api-access-6jr2d\") pod \"ceilometer-0\" (UID: \"9df73c9d-22c1-47c0-a996-0983f206dad5\") " pod="openstack/ceilometer-0" Oct 02 02:05:11 crc kubenswrapper[4885]: I1002 02:05:11.622544 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 02:05:11 crc kubenswrapper[4885]: I1002 02:05:11.764076 4885 generic.go:334] "Generic (PLEG): container finished" podID="b0fddcb3-3649-48f4-aa8f-85dd1c8137cf" containerID="9a7c1423da72336ecfa2cfe108bb2a287a4349ab5629c7bfb13486150d01b56a" exitCode=137 Oct 02 02:05:11 crc kubenswrapper[4885]: I1002 02:05:11.764352 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7c49b4cffb-brdz9" event={"ID":"b0fddcb3-3649-48f4-aa8f-85dd1c8137cf","Type":"ContainerDied","Data":"9a7c1423da72336ecfa2cfe108bb2a287a4349ab5629c7bfb13486150d01b56a"} Oct 02 02:05:11 crc kubenswrapper[4885]: I1002 02:05:11.767002 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-gp9st" Oct 02 02:05:11 crc kubenswrapper[4885]: I1002 02:05:11.767281 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-gp9st" event={"ID":"0feba5bf-20c5-4578-9cfc-f37d6511c2d6","Type":"ContainerDied","Data":"31f95e8399c06106eae6157550f827c372a5d14208ea2c35ccb560d237e7e9dc"} Oct 02 02:05:11 crc kubenswrapper[4885]: I1002 02:05:11.767323 4885 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="31f95e8399c06106eae6157550f827c372a5d14208ea2c35ccb560d237e7e9dc" Oct 02 02:05:11 crc kubenswrapper[4885]: I1002 02:05:11.786426 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"ab251703-d0d6-49cf-bd2a-b74dee9e48f7","Type":"ContainerStarted","Data":"1057673ec4ebe707c573e0f91f840cc904a92a55934d3dbd6532595865f710e4"} Oct 02 02:05:11 crc kubenswrapper[4885]: I1002 02:05:11.808975 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-9c5d5c444-vkjrr" Oct 02 02:05:11 crc kubenswrapper[4885]: I1002 02:05:11.808986 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-9c5d5c444-vkjrr" event={"ID":"887b9c10-e0ae-41b2-bfc3-a329f6013d98","Type":"ContainerDied","Data":"0b5865dc809e8e154588c9a9e4cdf2257e724e2d55495dd83fa32e279c2e4fc8"} Oct 02 02:05:11 crc kubenswrapper[4885]: I1002 02:05:11.809032 4885 scope.go:117] "RemoveContainer" containerID="772e2dc86bdea03785060468cdc4451b409b9c8bea7dad586dca7ef5e11995a4" Oct 02 02:05:11 crc kubenswrapper[4885]: I1002 02:05:11.814450 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.421673006 podStartE2EDuration="12.810928257s" podCreationTimestamp="2025-10-02 02:04:59 +0000 UTC" firstStartedPulling="2025-10-02 02:05:00.088940119 +0000 UTC m=+1088.900687518" lastFinishedPulling="2025-10-02 02:05:10.47819537 +0000 UTC m=+1099.289942769" observedRunningTime="2025-10-02 02:05:11.806756751 +0000 UTC m=+1100.618504150" watchObservedRunningTime="2025-10-02 02:05:11.810928257 +0000 UTC m=+1100.622675656" Oct 02 02:05:11 crc kubenswrapper[4885]: I1002 02:05:11.821464 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5d65d69454-8bzbn" event={"ID":"9bcc69ce-fb1a-4265-a6de-b2e2a10ac733","Type":"ContainerStarted","Data":"5e87d8939518e74e02c55864767059aea97ad28ff49c5c95ee1e9f12004aa868"} Oct 02 02:05:11 crc kubenswrapper[4885]: I1002 02:05:11.834936 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-7474b946b7-s2w7j" event={"ID":"3b22bebc-2f6c-4981-8166-4ccaf19f9727","Type":"ContainerStarted","Data":"00c31a47299fc1c7bcd9cc60096b03509334b0ba6070c0461899045cb370be4e"} Oct 02 02:05:11 crc kubenswrapper[4885]: I1002 02:05:11.835003 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-7474b946b7-s2w7j" event={"ID":"3b22bebc-2f6c-4981-8166-4ccaf19f9727","Type":"ContainerStarted","Data":"6cd966a683460e66919677e0a0f50caaeb63851b89c9f4e7e75a757234819f04"} Oct 02 02:05:11 crc kubenswrapper[4885]: I1002 02:05:11.836432 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-7474b946b7-s2w7j" Oct 02 02:05:11 crc kubenswrapper[4885]: I1002 02:05:11.836490 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-7474b946b7-s2w7j" Oct 02 02:05:11 crc kubenswrapper[4885]: I1002 02:05:11.851752 4885 scope.go:117] "RemoveContainer" containerID="01ca921473539e0109a018c1729a703dbf2f939cc7f6e4997e2af653d9fc5def" Oct 02 02:05:11 crc kubenswrapper[4885]: I1002 02:05:11.853296 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-9c5d5c444-vkjrr"] Oct 02 02:05:11 crc kubenswrapper[4885]: I1002 02:05:11.875941 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-9c5d5c444-vkjrr"] Oct 02 02:05:11 crc kubenswrapper[4885]: I1002 02:05:11.879943 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-7474b946b7-s2w7j" podStartSLOduration=7.879922418 podStartE2EDuration="7.879922418s" podCreationTimestamp="2025-10-02 02:05:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:05:11.87108779 +0000 UTC m=+1100.682835189" watchObservedRunningTime="2025-10-02 02:05:11.879922418 +0000 UTC m=+1100.691669817" Oct 02 02:05:11 crc kubenswrapper[4885]: I1002 02:05:11.946792 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-5c6b9856b7-vhzx6"] Oct 02 02:05:11 crc kubenswrapper[4885]: I1002 02:05:11.948405 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-5c6b9856b7-vhzx6" Oct 02 02:05:11 crc kubenswrapper[4885]: I1002 02:05:11.959024 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-2djz5" Oct 02 02:05:11 crc kubenswrapper[4885]: I1002 02:05:11.959320 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 02 02:05:11 crc kubenswrapper[4885]: I1002 02:05:11.959448 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Oct 02 02:05:11 crc kubenswrapper[4885]: I1002 02:05:11.995653 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-5c6b9856b7-vhzx6"] Oct 02 02:05:12 crc kubenswrapper[4885]: I1002 02:05:12.013082 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-6b9ddcb5cd-2r5wd"] Oct 02 02:05:12 crc kubenswrapper[4885]: I1002 02:05:12.014484 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-6b9ddcb5cd-2r5wd" Oct 02 02:05:12 crc kubenswrapper[4885]: I1002 02:05:12.024185 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Oct 02 02:05:12 crc kubenswrapper[4885]: I1002 02:05:12.071365 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="887b9c10-e0ae-41b2-bfc3-a329f6013d98" path="/var/lib/kubelet/pods/887b9c10-e0ae-41b2-bfc3-a329f6013d98/volumes" Oct 02 02:05:12 crc kubenswrapper[4885]: I1002 02:05:12.071967 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d1811543-6bba-4856-b8e1-814c7592723b" path="/var/lib/kubelet/pods/d1811543-6bba-4856-b8e1-814c7592723b/volumes" Oct 02 02:05:12 crc kubenswrapper[4885]: I1002 02:05:12.072672 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-6b9ddcb5cd-2r5wd"] Oct 02 02:05:12 crc kubenswrapper[4885]: I1002 02:05:12.075607 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/075adfcf-950d-44e6-ab5e-826f3b06c05e-combined-ca-bundle\") pod \"barbican-worker-5c6b9856b7-vhzx6\" (UID: \"075adfcf-950d-44e6-ab5e-826f3b06c05e\") " pod="openstack/barbican-worker-5c6b9856b7-vhzx6" Oct 02 02:05:12 crc kubenswrapper[4885]: I1002 02:05:12.075739 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/075adfcf-950d-44e6-ab5e-826f3b06c05e-logs\") pod \"barbican-worker-5c6b9856b7-vhzx6\" (UID: \"075adfcf-950d-44e6-ab5e-826f3b06c05e\") " pod="openstack/barbican-worker-5c6b9856b7-vhzx6" Oct 02 02:05:12 crc kubenswrapper[4885]: I1002 02:05:12.075785 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hzhpp\" (UniqueName: \"kubernetes.io/projected/075adfcf-950d-44e6-ab5e-826f3b06c05e-kube-api-access-hzhpp\") pod \"barbican-worker-5c6b9856b7-vhzx6\" (UID: \"075adfcf-950d-44e6-ab5e-826f3b06c05e\") " pod="openstack/barbican-worker-5c6b9856b7-vhzx6" Oct 02 02:05:12 crc kubenswrapper[4885]: I1002 02:05:12.075840 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/075adfcf-950d-44e6-ab5e-826f3b06c05e-config-data\") pod \"barbican-worker-5c6b9856b7-vhzx6\" (UID: \"075adfcf-950d-44e6-ab5e-826f3b06c05e\") " pod="openstack/barbican-worker-5c6b9856b7-vhzx6" Oct 02 02:05:12 crc kubenswrapper[4885]: I1002 02:05:12.075911 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/075adfcf-950d-44e6-ab5e-826f3b06c05e-config-data-custom\") pod \"barbican-worker-5c6b9856b7-vhzx6\" (UID: \"075adfcf-950d-44e6-ab5e-826f3b06c05e\") " pod="openstack/barbican-worker-5c6b9856b7-vhzx6" Oct 02 02:05:12 crc kubenswrapper[4885]: I1002 02:05:12.099316 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-s4xb6"] Oct 02 02:05:12 crc kubenswrapper[4885]: I1002 02:05:12.100758 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75c8ddd69c-s4xb6" Oct 02 02:05:12 crc kubenswrapper[4885]: I1002 02:05:12.102531 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-s4xb6"] Oct 02 02:05:12 crc kubenswrapper[4885]: I1002 02:05:12.109187 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 02:05:12 crc kubenswrapper[4885]: I1002 02:05:12.171163 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-6c87b5899d-lvdbp"] Oct 02 02:05:12 crc kubenswrapper[4885]: I1002 02:05:12.187232 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c416d7da-12d2-4aee-a9c3-5f171a791b43-combined-ca-bundle\") pod \"barbican-keystone-listener-6b9ddcb5cd-2r5wd\" (UID: \"c416d7da-12d2-4aee-a9c3-5f171a791b43\") " pod="openstack/barbican-keystone-listener-6b9ddcb5cd-2r5wd" Oct 02 02:05:12 crc kubenswrapper[4885]: I1002 02:05:12.187317 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/075adfcf-950d-44e6-ab5e-826f3b06c05e-config-data-custom\") pod \"barbican-worker-5c6b9856b7-vhzx6\" (UID: \"075adfcf-950d-44e6-ab5e-826f3b06c05e\") " pod="openstack/barbican-worker-5c6b9856b7-vhzx6" Oct 02 02:05:12 crc kubenswrapper[4885]: I1002 02:05:12.187352 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c416d7da-12d2-4aee-a9c3-5f171a791b43-config-data-custom\") pod \"barbican-keystone-listener-6b9ddcb5cd-2r5wd\" (UID: \"c416d7da-12d2-4aee-a9c3-5f171a791b43\") " pod="openstack/barbican-keystone-listener-6b9ddcb5cd-2r5wd" Oct 02 02:05:12 crc kubenswrapper[4885]: I1002 02:05:12.187376 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vf5b9\" (UniqueName: \"kubernetes.io/projected/c416d7da-12d2-4aee-a9c3-5f171a791b43-kube-api-access-vf5b9\") pod \"barbican-keystone-listener-6b9ddcb5cd-2r5wd\" (UID: \"c416d7da-12d2-4aee-a9c3-5f171a791b43\") " pod="openstack/barbican-keystone-listener-6b9ddcb5cd-2r5wd" Oct 02 02:05:12 crc kubenswrapper[4885]: I1002 02:05:12.187395 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c416d7da-12d2-4aee-a9c3-5f171a791b43-logs\") pod \"barbican-keystone-listener-6b9ddcb5cd-2r5wd\" (UID: \"c416d7da-12d2-4aee-a9c3-5f171a791b43\") " pod="openstack/barbican-keystone-listener-6b9ddcb5cd-2r5wd" Oct 02 02:05:12 crc kubenswrapper[4885]: I1002 02:05:12.187422 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/075adfcf-950d-44e6-ab5e-826f3b06c05e-combined-ca-bundle\") pod \"barbican-worker-5c6b9856b7-vhzx6\" (UID: \"075adfcf-950d-44e6-ab5e-826f3b06c05e\") " pod="openstack/barbican-worker-5c6b9856b7-vhzx6" Oct 02 02:05:12 crc kubenswrapper[4885]: I1002 02:05:12.187487 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/075adfcf-950d-44e6-ab5e-826f3b06c05e-logs\") pod \"barbican-worker-5c6b9856b7-vhzx6\" (UID: \"075adfcf-950d-44e6-ab5e-826f3b06c05e\") " pod="openstack/barbican-worker-5c6b9856b7-vhzx6" Oct 02 02:05:12 crc kubenswrapper[4885]: I1002 02:05:12.187504 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c416d7da-12d2-4aee-a9c3-5f171a791b43-config-data\") pod \"barbican-keystone-listener-6b9ddcb5cd-2r5wd\" (UID: \"c416d7da-12d2-4aee-a9c3-5f171a791b43\") " pod="openstack/barbican-keystone-listener-6b9ddcb5cd-2r5wd" Oct 02 02:05:12 crc kubenswrapper[4885]: I1002 02:05:12.187534 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hzhpp\" (UniqueName: \"kubernetes.io/projected/075adfcf-950d-44e6-ab5e-826f3b06c05e-kube-api-access-hzhpp\") pod \"barbican-worker-5c6b9856b7-vhzx6\" (UID: \"075adfcf-950d-44e6-ab5e-826f3b06c05e\") " pod="openstack/barbican-worker-5c6b9856b7-vhzx6" Oct 02 02:05:12 crc kubenswrapper[4885]: I1002 02:05:12.187558 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/075adfcf-950d-44e6-ab5e-826f3b06c05e-config-data\") pod \"barbican-worker-5c6b9856b7-vhzx6\" (UID: \"075adfcf-950d-44e6-ab5e-826f3b06c05e\") " pod="openstack/barbican-worker-5c6b9856b7-vhzx6" Oct 02 02:05:12 crc kubenswrapper[4885]: I1002 02:05:12.190606 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6c87b5899d-lvdbp" Oct 02 02:05:12 crc kubenswrapper[4885]: I1002 02:05:12.194576 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/075adfcf-950d-44e6-ab5e-826f3b06c05e-config-data\") pod \"barbican-worker-5c6b9856b7-vhzx6\" (UID: \"075adfcf-950d-44e6-ab5e-826f3b06c05e\") " pod="openstack/barbican-worker-5c6b9856b7-vhzx6" Oct 02 02:05:12 crc kubenswrapper[4885]: I1002 02:05:12.197827 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/075adfcf-950d-44e6-ab5e-826f3b06c05e-logs\") pod \"barbican-worker-5c6b9856b7-vhzx6\" (UID: \"075adfcf-950d-44e6-ab5e-826f3b06c05e\") " pod="openstack/barbican-worker-5c6b9856b7-vhzx6" Oct 02 02:05:12 crc kubenswrapper[4885]: I1002 02:05:12.212863 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Oct 02 02:05:12 crc kubenswrapper[4885]: I1002 02:05:12.213249 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/075adfcf-950d-44e6-ab5e-826f3b06c05e-config-data-custom\") pod \"barbican-worker-5c6b9856b7-vhzx6\" (UID: \"075adfcf-950d-44e6-ab5e-826f3b06c05e\") " pod="openstack/barbican-worker-5c6b9856b7-vhzx6" Oct 02 02:05:12 crc kubenswrapper[4885]: I1002 02:05:12.212875 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/075adfcf-950d-44e6-ab5e-826f3b06c05e-combined-ca-bundle\") pod \"barbican-worker-5c6b9856b7-vhzx6\" (UID: \"075adfcf-950d-44e6-ab5e-826f3b06c05e\") " pod="openstack/barbican-worker-5c6b9856b7-vhzx6" Oct 02 02:05:12 crc kubenswrapper[4885]: I1002 02:05:12.214813 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6c87b5899d-lvdbp"] Oct 02 02:05:12 crc kubenswrapper[4885]: I1002 02:05:12.229843 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hzhpp\" (UniqueName: \"kubernetes.io/projected/075adfcf-950d-44e6-ab5e-826f3b06c05e-kube-api-access-hzhpp\") pod \"barbican-worker-5c6b9856b7-vhzx6\" (UID: \"075adfcf-950d-44e6-ab5e-826f3b06c05e\") " pod="openstack/barbican-worker-5c6b9856b7-vhzx6" Oct 02 02:05:12 crc kubenswrapper[4885]: I1002 02:05:12.292095 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c416d7da-12d2-4aee-a9c3-5f171a791b43-combined-ca-bundle\") pod \"barbican-keystone-listener-6b9ddcb5cd-2r5wd\" (UID: \"c416d7da-12d2-4aee-a9c3-5f171a791b43\") " pod="openstack/barbican-keystone-listener-6b9ddcb5cd-2r5wd" Oct 02 02:05:12 crc kubenswrapper[4885]: I1002 02:05:12.292144 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bc7146f7-6d12-4bd2-8fd2-0604f358bedf-dns-swift-storage-0\") pod \"dnsmasq-dns-75c8ddd69c-s4xb6\" (UID: \"bc7146f7-6d12-4bd2-8fd2-0604f358bedf\") " pod="openstack/dnsmasq-dns-75c8ddd69c-s4xb6" Oct 02 02:05:12 crc kubenswrapper[4885]: I1002 02:05:12.292171 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bc7146f7-6d12-4bd2-8fd2-0604f358bedf-config\") pod \"dnsmasq-dns-75c8ddd69c-s4xb6\" (UID: \"bc7146f7-6d12-4bd2-8fd2-0604f358bedf\") " pod="openstack/dnsmasq-dns-75c8ddd69c-s4xb6" Oct 02 02:05:12 crc kubenswrapper[4885]: I1002 02:05:12.292199 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb9e27f4-9f80-40ff-af3e-fe5aca9f5755-combined-ca-bundle\") pod \"barbican-api-6c87b5899d-lvdbp\" (UID: \"bb9e27f4-9f80-40ff-af3e-fe5aca9f5755\") " pod="openstack/barbican-api-6c87b5899d-lvdbp" Oct 02 02:05:12 crc kubenswrapper[4885]: I1002 02:05:12.292216 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c416d7da-12d2-4aee-a9c3-5f171a791b43-config-data-custom\") pod \"barbican-keystone-listener-6b9ddcb5cd-2r5wd\" (UID: \"c416d7da-12d2-4aee-a9c3-5f171a791b43\") " pod="openstack/barbican-keystone-listener-6b9ddcb5cd-2r5wd" Oct 02 02:05:12 crc kubenswrapper[4885]: I1002 02:05:12.292243 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vf5b9\" (UniqueName: \"kubernetes.io/projected/c416d7da-12d2-4aee-a9c3-5f171a791b43-kube-api-access-vf5b9\") pod \"barbican-keystone-listener-6b9ddcb5cd-2r5wd\" (UID: \"c416d7da-12d2-4aee-a9c3-5f171a791b43\") " pod="openstack/barbican-keystone-listener-6b9ddcb5cd-2r5wd" Oct 02 02:05:12 crc kubenswrapper[4885]: I1002 02:05:12.292285 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bc7146f7-6d12-4bd2-8fd2-0604f358bedf-dns-svc\") pod \"dnsmasq-dns-75c8ddd69c-s4xb6\" (UID: \"bc7146f7-6d12-4bd2-8fd2-0604f358bedf\") " pod="openstack/dnsmasq-dns-75c8ddd69c-s4xb6" Oct 02 02:05:12 crc kubenswrapper[4885]: I1002 02:05:12.292309 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c416d7da-12d2-4aee-a9c3-5f171a791b43-logs\") pod \"barbican-keystone-listener-6b9ddcb5cd-2r5wd\" (UID: \"c416d7da-12d2-4aee-a9c3-5f171a791b43\") " pod="openstack/barbican-keystone-listener-6b9ddcb5cd-2r5wd" Oct 02 02:05:12 crc kubenswrapper[4885]: I1002 02:05:12.292333 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb9e27f4-9f80-40ff-af3e-fe5aca9f5755-config-data\") pod \"barbican-api-6c87b5899d-lvdbp\" (UID: \"bb9e27f4-9f80-40ff-af3e-fe5aca9f5755\") " pod="openstack/barbican-api-6c87b5899d-lvdbp" Oct 02 02:05:12 crc kubenswrapper[4885]: I1002 02:05:12.292365 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bb9e27f4-9f80-40ff-af3e-fe5aca9f5755-config-data-custom\") pod \"barbican-api-6c87b5899d-lvdbp\" (UID: \"bb9e27f4-9f80-40ff-af3e-fe5aca9f5755\") " pod="openstack/barbican-api-6c87b5899d-lvdbp" Oct 02 02:05:12 crc kubenswrapper[4885]: I1002 02:05:12.292388 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bc7146f7-6d12-4bd2-8fd2-0604f358bedf-ovsdbserver-sb\") pod \"dnsmasq-dns-75c8ddd69c-s4xb6\" (UID: \"bc7146f7-6d12-4bd2-8fd2-0604f358bedf\") " pod="openstack/dnsmasq-dns-75c8ddd69c-s4xb6" Oct 02 02:05:12 crc kubenswrapper[4885]: I1002 02:05:12.292424 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xf65m\" (UniqueName: \"kubernetes.io/projected/bc7146f7-6d12-4bd2-8fd2-0604f358bedf-kube-api-access-xf65m\") pod \"dnsmasq-dns-75c8ddd69c-s4xb6\" (UID: \"bc7146f7-6d12-4bd2-8fd2-0604f358bedf\") " pod="openstack/dnsmasq-dns-75c8ddd69c-s4xb6" Oct 02 02:05:12 crc kubenswrapper[4885]: I1002 02:05:12.292455 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bb9e27f4-9f80-40ff-af3e-fe5aca9f5755-logs\") pod \"barbican-api-6c87b5899d-lvdbp\" (UID: \"bb9e27f4-9f80-40ff-af3e-fe5aca9f5755\") " pod="openstack/barbican-api-6c87b5899d-lvdbp" Oct 02 02:05:12 crc kubenswrapper[4885]: I1002 02:05:12.292482 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c416d7da-12d2-4aee-a9c3-5f171a791b43-config-data\") pod \"barbican-keystone-listener-6b9ddcb5cd-2r5wd\" (UID: \"c416d7da-12d2-4aee-a9c3-5f171a791b43\") " pod="openstack/barbican-keystone-listener-6b9ddcb5cd-2r5wd" Oct 02 02:05:12 crc kubenswrapper[4885]: I1002 02:05:12.292505 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-72p9b\" (UniqueName: \"kubernetes.io/projected/bb9e27f4-9f80-40ff-af3e-fe5aca9f5755-kube-api-access-72p9b\") pod \"barbican-api-6c87b5899d-lvdbp\" (UID: \"bb9e27f4-9f80-40ff-af3e-fe5aca9f5755\") " pod="openstack/barbican-api-6c87b5899d-lvdbp" Oct 02 02:05:12 crc kubenswrapper[4885]: I1002 02:05:12.292541 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bc7146f7-6d12-4bd2-8fd2-0604f358bedf-ovsdbserver-nb\") pod \"dnsmasq-dns-75c8ddd69c-s4xb6\" (UID: \"bc7146f7-6d12-4bd2-8fd2-0604f358bedf\") " pod="openstack/dnsmasq-dns-75c8ddd69c-s4xb6" Oct 02 02:05:12 crc kubenswrapper[4885]: I1002 02:05:12.293706 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c416d7da-12d2-4aee-a9c3-5f171a791b43-logs\") pod \"barbican-keystone-listener-6b9ddcb5cd-2r5wd\" (UID: \"c416d7da-12d2-4aee-a9c3-5f171a791b43\") " pod="openstack/barbican-keystone-listener-6b9ddcb5cd-2r5wd" Oct 02 02:05:12 crc kubenswrapper[4885]: I1002 02:05:12.293835 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-5c6b9856b7-vhzx6" Oct 02 02:05:12 crc kubenswrapper[4885]: I1002 02:05:12.301534 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c416d7da-12d2-4aee-a9c3-5f171a791b43-combined-ca-bundle\") pod \"barbican-keystone-listener-6b9ddcb5cd-2r5wd\" (UID: \"c416d7da-12d2-4aee-a9c3-5f171a791b43\") " pod="openstack/barbican-keystone-listener-6b9ddcb5cd-2r5wd" Oct 02 02:05:12 crc kubenswrapper[4885]: I1002 02:05:12.304828 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c416d7da-12d2-4aee-a9c3-5f171a791b43-config-data-custom\") pod \"barbican-keystone-listener-6b9ddcb5cd-2r5wd\" (UID: \"c416d7da-12d2-4aee-a9c3-5f171a791b43\") " pod="openstack/barbican-keystone-listener-6b9ddcb5cd-2r5wd" Oct 02 02:05:12 crc kubenswrapper[4885]: I1002 02:05:12.307905 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c416d7da-12d2-4aee-a9c3-5f171a791b43-config-data\") pod \"barbican-keystone-listener-6b9ddcb5cd-2r5wd\" (UID: \"c416d7da-12d2-4aee-a9c3-5f171a791b43\") " pod="openstack/barbican-keystone-listener-6b9ddcb5cd-2r5wd" Oct 02 02:05:12 crc kubenswrapper[4885]: I1002 02:05:12.318920 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vf5b9\" (UniqueName: \"kubernetes.io/projected/c416d7da-12d2-4aee-a9c3-5f171a791b43-kube-api-access-vf5b9\") pod \"barbican-keystone-listener-6b9ddcb5cd-2r5wd\" (UID: \"c416d7da-12d2-4aee-a9c3-5f171a791b43\") " pod="openstack/barbican-keystone-listener-6b9ddcb5cd-2r5wd" Oct 02 02:05:12 crc kubenswrapper[4885]: I1002 02:05:12.365238 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-6b9ddcb5cd-2r5wd" Oct 02 02:05:12 crc kubenswrapper[4885]: I1002 02:05:12.397376 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bc7146f7-6d12-4bd2-8fd2-0604f358bedf-config\") pod \"dnsmasq-dns-75c8ddd69c-s4xb6\" (UID: \"bc7146f7-6d12-4bd2-8fd2-0604f358bedf\") " pod="openstack/dnsmasq-dns-75c8ddd69c-s4xb6" Oct 02 02:05:12 crc kubenswrapper[4885]: I1002 02:05:12.397428 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb9e27f4-9f80-40ff-af3e-fe5aca9f5755-combined-ca-bundle\") pod \"barbican-api-6c87b5899d-lvdbp\" (UID: \"bb9e27f4-9f80-40ff-af3e-fe5aca9f5755\") " pod="openstack/barbican-api-6c87b5899d-lvdbp" Oct 02 02:05:12 crc kubenswrapper[4885]: I1002 02:05:12.397452 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bc7146f7-6d12-4bd2-8fd2-0604f358bedf-dns-svc\") pod \"dnsmasq-dns-75c8ddd69c-s4xb6\" (UID: \"bc7146f7-6d12-4bd2-8fd2-0604f358bedf\") " pod="openstack/dnsmasq-dns-75c8ddd69c-s4xb6" Oct 02 02:05:12 crc kubenswrapper[4885]: I1002 02:05:12.397476 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb9e27f4-9f80-40ff-af3e-fe5aca9f5755-config-data\") pod \"barbican-api-6c87b5899d-lvdbp\" (UID: \"bb9e27f4-9f80-40ff-af3e-fe5aca9f5755\") " pod="openstack/barbican-api-6c87b5899d-lvdbp" Oct 02 02:05:12 crc kubenswrapper[4885]: I1002 02:05:12.397501 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bb9e27f4-9f80-40ff-af3e-fe5aca9f5755-config-data-custom\") pod \"barbican-api-6c87b5899d-lvdbp\" (UID: \"bb9e27f4-9f80-40ff-af3e-fe5aca9f5755\") " pod="openstack/barbican-api-6c87b5899d-lvdbp" Oct 02 02:05:12 crc kubenswrapper[4885]: I1002 02:05:12.397522 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bc7146f7-6d12-4bd2-8fd2-0604f358bedf-ovsdbserver-sb\") pod \"dnsmasq-dns-75c8ddd69c-s4xb6\" (UID: \"bc7146f7-6d12-4bd2-8fd2-0604f358bedf\") " pod="openstack/dnsmasq-dns-75c8ddd69c-s4xb6" Oct 02 02:05:12 crc kubenswrapper[4885]: I1002 02:05:12.397556 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xf65m\" (UniqueName: \"kubernetes.io/projected/bc7146f7-6d12-4bd2-8fd2-0604f358bedf-kube-api-access-xf65m\") pod \"dnsmasq-dns-75c8ddd69c-s4xb6\" (UID: \"bc7146f7-6d12-4bd2-8fd2-0604f358bedf\") " pod="openstack/dnsmasq-dns-75c8ddd69c-s4xb6" Oct 02 02:05:12 crc kubenswrapper[4885]: I1002 02:05:12.397584 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bb9e27f4-9f80-40ff-af3e-fe5aca9f5755-logs\") pod \"barbican-api-6c87b5899d-lvdbp\" (UID: \"bb9e27f4-9f80-40ff-af3e-fe5aca9f5755\") " pod="openstack/barbican-api-6c87b5899d-lvdbp" Oct 02 02:05:12 crc kubenswrapper[4885]: I1002 02:05:12.397623 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-72p9b\" (UniqueName: \"kubernetes.io/projected/bb9e27f4-9f80-40ff-af3e-fe5aca9f5755-kube-api-access-72p9b\") pod \"barbican-api-6c87b5899d-lvdbp\" (UID: \"bb9e27f4-9f80-40ff-af3e-fe5aca9f5755\") " pod="openstack/barbican-api-6c87b5899d-lvdbp" Oct 02 02:05:12 crc kubenswrapper[4885]: I1002 02:05:12.397658 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bc7146f7-6d12-4bd2-8fd2-0604f358bedf-ovsdbserver-nb\") pod \"dnsmasq-dns-75c8ddd69c-s4xb6\" (UID: \"bc7146f7-6d12-4bd2-8fd2-0604f358bedf\") " pod="openstack/dnsmasq-dns-75c8ddd69c-s4xb6" Oct 02 02:05:12 crc kubenswrapper[4885]: I1002 02:05:12.397693 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bc7146f7-6d12-4bd2-8fd2-0604f358bedf-dns-swift-storage-0\") pod \"dnsmasq-dns-75c8ddd69c-s4xb6\" (UID: \"bc7146f7-6d12-4bd2-8fd2-0604f358bedf\") " pod="openstack/dnsmasq-dns-75c8ddd69c-s4xb6" Oct 02 02:05:12 crc kubenswrapper[4885]: I1002 02:05:12.398467 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bc7146f7-6d12-4bd2-8fd2-0604f358bedf-dns-swift-storage-0\") pod \"dnsmasq-dns-75c8ddd69c-s4xb6\" (UID: \"bc7146f7-6d12-4bd2-8fd2-0604f358bedf\") " pod="openstack/dnsmasq-dns-75c8ddd69c-s4xb6" Oct 02 02:05:12 crc kubenswrapper[4885]: I1002 02:05:12.398946 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bc7146f7-6d12-4bd2-8fd2-0604f358bedf-ovsdbserver-sb\") pod \"dnsmasq-dns-75c8ddd69c-s4xb6\" (UID: \"bc7146f7-6d12-4bd2-8fd2-0604f358bedf\") " pod="openstack/dnsmasq-dns-75c8ddd69c-s4xb6" Oct 02 02:05:12 crc kubenswrapper[4885]: I1002 02:05:12.399428 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bb9e27f4-9f80-40ff-af3e-fe5aca9f5755-logs\") pod \"barbican-api-6c87b5899d-lvdbp\" (UID: \"bb9e27f4-9f80-40ff-af3e-fe5aca9f5755\") " pod="openstack/barbican-api-6c87b5899d-lvdbp" Oct 02 02:05:12 crc kubenswrapper[4885]: I1002 02:05:12.400033 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bc7146f7-6d12-4bd2-8fd2-0604f358bedf-ovsdbserver-nb\") pod \"dnsmasq-dns-75c8ddd69c-s4xb6\" (UID: \"bc7146f7-6d12-4bd2-8fd2-0604f358bedf\") " pod="openstack/dnsmasq-dns-75c8ddd69c-s4xb6" Oct 02 02:05:12 crc kubenswrapper[4885]: I1002 02:05:12.400927 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bc7146f7-6d12-4bd2-8fd2-0604f358bedf-config\") pod \"dnsmasq-dns-75c8ddd69c-s4xb6\" (UID: \"bc7146f7-6d12-4bd2-8fd2-0604f358bedf\") " pod="openstack/dnsmasq-dns-75c8ddd69c-s4xb6" Oct 02 02:05:12 crc kubenswrapper[4885]: I1002 02:05:12.401616 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bc7146f7-6d12-4bd2-8fd2-0604f358bedf-dns-svc\") pod \"dnsmasq-dns-75c8ddd69c-s4xb6\" (UID: \"bc7146f7-6d12-4bd2-8fd2-0604f358bedf\") " pod="openstack/dnsmasq-dns-75c8ddd69c-s4xb6" Oct 02 02:05:12 crc kubenswrapper[4885]: I1002 02:05:12.409299 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb9e27f4-9f80-40ff-af3e-fe5aca9f5755-config-data\") pod \"barbican-api-6c87b5899d-lvdbp\" (UID: \"bb9e27f4-9f80-40ff-af3e-fe5aca9f5755\") " pod="openstack/barbican-api-6c87b5899d-lvdbp" Oct 02 02:05:12 crc kubenswrapper[4885]: I1002 02:05:12.409987 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bb9e27f4-9f80-40ff-af3e-fe5aca9f5755-config-data-custom\") pod \"barbican-api-6c87b5899d-lvdbp\" (UID: \"bb9e27f4-9f80-40ff-af3e-fe5aca9f5755\") " pod="openstack/barbican-api-6c87b5899d-lvdbp" Oct 02 02:05:12 crc kubenswrapper[4885]: I1002 02:05:12.413382 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb9e27f4-9f80-40ff-af3e-fe5aca9f5755-combined-ca-bundle\") pod \"barbican-api-6c87b5899d-lvdbp\" (UID: \"bb9e27f4-9f80-40ff-af3e-fe5aca9f5755\") " pod="openstack/barbican-api-6c87b5899d-lvdbp" Oct 02 02:05:12 crc kubenswrapper[4885]: I1002 02:05:12.423965 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xf65m\" (UniqueName: \"kubernetes.io/projected/bc7146f7-6d12-4bd2-8fd2-0604f358bedf-kube-api-access-xf65m\") pod \"dnsmasq-dns-75c8ddd69c-s4xb6\" (UID: \"bc7146f7-6d12-4bd2-8fd2-0604f358bedf\") " pod="openstack/dnsmasq-dns-75c8ddd69c-s4xb6" Oct 02 02:05:12 crc kubenswrapper[4885]: I1002 02:05:12.426449 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75c8ddd69c-s4xb6" Oct 02 02:05:12 crc kubenswrapper[4885]: I1002 02:05:12.429854 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-72p9b\" (UniqueName: \"kubernetes.io/projected/bb9e27f4-9f80-40ff-af3e-fe5aca9f5755-kube-api-access-72p9b\") pod \"barbican-api-6c87b5899d-lvdbp\" (UID: \"bb9e27f4-9f80-40ff-af3e-fe5aca9f5755\") " pod="openstack/barbican-api-6c87b5899d-lvdbp" Oct 02 02:05:12 crc kubenswrapper[4885]: I1002 02:05:12.441051 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6c87b5899d-lvdbp" Oct 02 02:05:12 crc kubenswrapper[4885]: I1002 02:05:12.886929 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7c49b4cffb-brdz9" Oct 02 02:05:12 crc kubenswrapper[4885]: I1002 02:05:12.946698 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9df73c9d-22c1-47c0-a996-0983f206dad5","Type":"ContainerStarted","Data":"7afdecbaf1bfab89170f887053ae340028a863e6810764f52b01aebac9f7ee08"} Oct 02 02:05:13 crc kubenswrapper[4885]: I1002 02:05:13.011971 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b0fddcb3-3649-48f4-aa8f-85dd1c8137cf-logs\") pod \"b0fddcb3-3649-48f4-aa8f-85dd1c8137cf\" (UID: \"b0fddcb3-3649-48f4-aa8f-85dd1c8137cf\") " Oct 02 02:05:13 crc kubenswrapper[4885]: I1002 02:05:13.012316 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b0fddcb3-3649-48f4-aa8f-85dd1c8137cf-scripts\") pod \"b0fddcb3-3649-48f4-aa8f-85dd1c8137cf\" (UID: \"b0fddcb3-3649-48f4-aa8f-85dd1c8137cf\") " Oct 02 02:05:13 crc kubenswrapper[4885]: I1002 02:05:13.012401 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pmfgm\" (UniqueName: \"kubernetes.io/projected/b0fddcb3-3649-48f4-aa8f-85dd1c8137cf-kube-api-access-pmfgm\") pod \"b0fddcb3-3649-48f4-aa8f-85dd1c8137cf\" (UID: \"b0fddcb3-3649-48f4-aa8f-85dd1c8137cf\") " Oct 02 02:05:13 crc kubenswrapper[4885]: I1002 02:05:13.012608 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b0fddcb3-3649-48f4-aa8f-85dd1c8137cf-logs" (OuterVolumeSpecName: "logs") pod "b0fddcb3-3649-48f4-aa8f-85dd1c8137cf" (UID: "b0fddcb3-3649-48f4-aa8f-85dd1c8137cf"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:05:13 crc kubenswrapper[4885]: I1002 02:05:13.013196 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/b0fddcb3-3649-48f4-aa8f-85dd1c8137cf-horizon-tls-certs\") pod \"b0fddcb3-3649-48f4-aa8f-85dd1c8137cf\" (UID: \"b0fddcb3-3649-48f4-aa8f-85dd1c8137cf\") " Oct 02 02:05:13 crc kubenswrapper[4885]: I1002 02:05:13.014888 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/b0fddcb3-3649-48f4-aa8f-85dd1c8137cf-horizon-secret-key\") pod \"b0fddcb3-3649-48f4-aa8f-85dd1c8137cf\" (UID: \"b0fddcb3-3649-48f4-aa8f-85dd1c8137cf\") " Oct 02 02:05:13 crc kubenswrapper[4885]: I1002 02:05:13.014946 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0fddcb3-3649-48f4-aa8f-85dd1c8137cf-combined-ca-bundle\") pod \"b0fddcb3-3649-48f4-aa8f-85dd1c8137cf\" (UID: \"b0fddcb3-3649-48f4-aa8f-85dd1c8137cf\") " Oct 02 02:05:13 crc kubenswrapper[4885]: I1002 02:05:13.014984 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b0fddcb3-3649-48f4-aa8f-85dd1c8137cf-config-data\") pod \"b0fddcb3-3649-48f4-aa8f-85dd1c8137cf\" (UID: \"b0fddcb3-3649-48f4-aa8f-85dd1c8137cf\") " Oct 02 02:05:13 crc kubenswrapper[4885]: I1002 02:05:13.015761 4885 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b0fddcb3-3649-48f4-aa8f-85dd1c8137cf-logs\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:13 crc kubenswrapper[4885]: I1002 02:05:13.020862 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b0fddcb3-3649-48f4-aa8f-85dd1c8137cf-kube-api-access-pmfgm" (OuterVolumeSpecName: "kube-api-access-pmfgm") pod "b0fddcb3-3649-48f4-aa8f-85dd1c8137cf" (UID: "b0fddcb3-3649-48f4-aa8f-85dd1c8137cf"). InnerVolumeSpecName "kube-api-access-pmfgm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:05:13 crc kubenswrapper[4885]: I1002 02:05:13.023541 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b0fddcb3-3649-48f4-aa8f-85dd1c8137cf-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "b0fddcb3-3649-48f4-aa8f-85dd1c8137cf" (UID: "b0fddcb3-3649-48f4-aa8f-85dd1c8137cf"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:05:13 crc kubenswrapper[4885]: I1002 02:05:13.040043 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b0fddcb3-3649-48f4-aa8f-85dd1c8137cf-scripts" (OuterVolumeSpecName: "scripts") pod "b0fddcb3-3649-48f4-aa8f-85dd1c8137cf" (UID: "b0fddcb3-3649-48f4-aa8f-85dd1c8137cf"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:05:13 crc kubenswrapper[4885]: I1002 02:05:13.058294 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b0fddcb3-3649-48f4-aa8f-85dd1c8137cf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b0fddcb3-3649-48f4-aa8f-85dd1c8137cf" (UID: "b0fddcb3-3649-48f4-aa8f-85dd1c8137cf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:05:13 crc kubenswrapper[4885]: I1002 02:05:13.058808 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b0fddcb3-3649-48f4-aa8f-85dd1c8137cf-config-data" (OuterVolumeSpecName: "config-data") pod "b0fddcb3-3649-48f4-aa8f-85dd1c8137cf" (UID: "b0fddcb3-3649-48f4-aa8f-85dd1c8137cf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:05:13 crc kubenswrapper[4885]: I1002 02:05:13.097461 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-6b9ddcb5cd-2r5wd"] Oct 02 02:05:13 crc kubenswrapper[4885]: I1002 02:05:13.105430 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b0fddcb3-3649-48f4-aa8f-85dd1c8137cf-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "b0fddcb3-3649-48f4-aa8f-85dd1c8137cf" (UID: "b0fddcb3-3649-48f4-aa8f-85dd1c8137cf"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:05:13 crc kubenswrapper[4885]: I1002 02:05:13.122492 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pmfgm\" (UniqueName: \"kubernetes.io/projected/b0fddcb3-3649-48f4-aa8f-85dd1c8137cf-kube-api-access-pmfgm\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:13 crc kubenswrapper[4885]: I1002 02:05:13.122523 4885 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/b0fddcb3-3649-48f4-aa8f-85dd1c8137cf-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:13 crc kubenswrapper[4885]: I1002 02:05:13.122532 4885 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/b0fddcb3-3649-48f4-aa8f-85dd1c8137cf-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:13 crc kubenswrapper[4885]: I1002 02:05:13.122541 4885 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0fddcb3-3649-48f4-aa8f-85dd1c8137cf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:13 crc kubenswrapper[4885]: I1002 02:05:13.122551 4885 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b0fddcb3-3649-48f4-aa8f-85dd1c8137cf-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:13 crc kubenswrapper[4885]: I1002 02:05:13.122562 4885 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b0fddcb3-3649-48f4-aa8f-85dd1c8137cf-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:13 crc kubenswrapper[4885]: W1002 02:05:13.176178 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod075adfcf_950d_44e6_ab5e_826f3b06c05e.slice/crio-d4c435f04bfe002cb5fad30dd02fdbe00e86bf8503fe4bac89378dc17919afc7 WatchSource:0}: Error finding container d4c435f04bfe002cb5fad30dd02fdbe00e86bf8503fe4bac89378dc17919afc7: Status 404 returned error can't find the container with id d4c435f04bfe002cb5fad30dd02fdbe00e86bf8503fe4bac89378dc17919afc7 Oct 02 02:05:13 crc kubenswrapper[4885]: I1002 02:05:13.176818 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-5c6b9856b7-vhzx6"] Oct 02 02:05:13 crc kubenswrapper[4885]: W1002 02:05:13.258391 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbb9e27f4_9f80_40ff_af3e_fe5aca9f5755.slice/crio-67f3f877dee69316bb6d1e486462adc80083a21a3f72f2b5be1c92c40c746e3c WatchSource:0}: Error finding container 67f3f877dee69316bb6d1e486462adc80083a21a3f72f2b5be1c92c40c746e3c: Status 404 returned error can't find the container with id 67f3f877dee69316bb6d1e486462adc80083a21a3f72f2b5be1c92c40c746e3c Oct 02 02:05:13 crc kubenswrapper[4885]: I1002 02:05:13.260208 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6c87b5899d-lvdbp"] Oct 02 02:05:13 crc kubenswrapper[4885]: I1002 02:05:13.353149 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-s4xb6"] Oct 02 02:05:13 crc kubenswrapper[4885]: I1002 02:05:13.558307 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 02:05:13 crc kubenswrapper[4885]: I1002 02:05:13.588127 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-5d65d69454-8bzbn" Oct 02 02:05:13 crc kubenswrapper[4885]: I1002 02:05:13.971493 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7c49b4cffb-brdz9" event={"ID":"b0fddcb3-3649-48f4-aa8f-85dd1c8137cf","Type":"ContainerDied","Data":"2f5362e018a9e3621dd43a09aeb9e15a34e904554a2392e255e4061c6a46d062"} Oct 02 02:05:13 crc kubenswrapper[4885]: I1002 02:05:13.971773 4885 scope.go:117] "RemoveContainer" containerID="87f582c0838b65442115c622d40cc23454484a73c597b4939299160d6e09873a" Oct 02 02:05:13 crc kubenswrapper[4885]: I1002 02:05:13.971892 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7c49b4cffb-brdz9" Oct 02 02:05:14 crc kubenswrapper[4885]: I1002 02:05:14.004334 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6c87b5899d-lvdbp" event={"ID":"bb9e27f4-9f80-40ff-af3e-fe5aca9f5755","Type":"ContainerStarted","Data":"2a03a6f00575b3ffc6251f3cfb33c8f81551cffb35a3821957a66e0609a9d69d"} Oct 02 02:05:14 crc kubenswrapper[4885]: I1002 02:05:14.004375 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6c87b5899d-lvdbp" event={"ID":"bb9e27f4-9f80-40ff-af3e-fe5aca9f5755","Type":"ContainerStarted","Data":"6126dae4aa6846b8af80f7984b55f1fce7b7ce164a49363c7ae5e9d582b813a2"} Oct 02 02:05:14 crc kubenswrapper[4885]: I1002 02:05:14.004386 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6c87b5899d-lvdbp" event={"ID":"bb9e27f4-9f80-40ff-af3e-fe5aca9f5755","Type":"ContainerStarted","Data":"67f3f877dee69316bb6d1e486462adc80083a21a3f72f2b5be1c92c40c746e3c"} Oct 02 02:05:14 crc kubenswrapper[4885]: I1002 02:05:14.004465 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6c87b5899d-lvdbp" Oct 02 02:05:14 crc kubenswrapper[4885]: I1002 02:05:14.004480 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6c87b5899d-lvdbp" Oct 02 02:05:14 crc kubenswrapper[4885]: I1002 02:05:14.005922 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6b9ddcb5cd-2r5wd" event={"ID":"c416d7da-12d2-4aee-a9c3-5f171a791b43","Type":"ContainerStarted","Data":"44695bd24766c019acce0e7ef79f70e197fec7d320a664a7e5196296274c207e"} Oct 02 02:05:14 crc kubenswrapper[4885]: I1002 02:05:14.010672 4885 generic.go:334] "Generic (PLEG): container finished" podID="bc7146f7-6d12-4bd2-8fd2-0604f358bedf" containerID="ad63936e2ba60a467001602dab32ef5d5ec0df00c49ea0cf10bb45d14ea406ae" exitCode=0 Oct 02 02:05:14 crc kubenswrapper[4885]: I1002 02:05:14.010791 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c8ddd69c-s4xb6" event={"ID":"bc7146f7-6d12-4bd2-8fd2-0604f358bedf","Type":"ContainerDied","Data":"ad63936e2ba60a467001602dab32ef5d5ec0df00c49ea0cf10bb45d14ea406ae"} Oct 02 02:05:14 crc kubenswrapper[4885]: I1002 02:05:14.010817 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c8ddd69c-s4xb6" event={"ID":"bc7146f7-6d12-4bd2-8fd2-0604f358bedf","Type":"ContainerStarted","Data":"063ceff3164315c0e2bde7e60fce9d38953e290a687949a326aa73ff7d62f4c2"} Oct 02 02:05:14 crc kubenswrapper[4885]: I1002 02:05:14.017712 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7c49b4cffb-brdz9"] Oct 02 02:05:14 crc kubenswrapper[4885]: I1002 02:05:14.023472 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-7c49b4cffb-brdz9"] Oct 02 02:05:14 crc kubenswrapper[4885]: I1002 02:05:14.032739 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-6c87b5899d-lvdbp" podStartSLOduration=2.032720811 podStartE2EDuration="2.032720811s" podCreationTimestamp="2025-10-02 02:05:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:05:14.02641941 +0000 UTC m=+1102.838166809" watchObservedRunningTime="2025-10-02 02:05:14.032720811 +0000 UTC m=+1102.844468210" Oct 02 02:05:14 crc kubenswrapper[4885]: I1002 02:05:14.042907 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9df73c9d-22c1-47c0-a996-0983f206dad5","Type":"ContainerStarted","Data":"597d57d9acdd3cbb0ea1e6a0716e324596b7fdad2025c8334be56d83e0eb7b4a"} Oct 02 02:05:14 crc kubenswrapper[4885]: I1002 02:05:14.042946 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9df73c9d-22c1-47c0-a996-0983f206dad5","Type":"ContainerStarted","Data":"9706f642020809b48c4b255916f1caf766eb34203bb17a12bfc4d4acb363e7d3"} Oct 02 02:05:14 crc kubenswrapper[4885]: I1002 02:05:14.044521 4885 generic.go:334] "Generic (PLEG): container finished" podID="9c66bb6a-5a3c-4223-aba6-c5d9d0ea83ba" containerID="521d56adf2bb767b794788e9e4a1227efa64a91af1a488857a83713bc3a46de9" exitCode=0 Oct 02 02:05:14 crc kubenswrapper[4885]: I1002 02:05:14.044560 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-tjcz4" event={"ID":"9c66bb6a-5a3c-4223-aba6-c5d9d0ea83ba","Type":"ContainerDied","Data":"521d56adf2bb767b794788e9e4a1227efa64a91af1a488857a83713bc3a46de9"} Oct 02 02:05:14 crc kubenswrapper[4885]: I1002 02:05:14.061321 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b0fddcb3-3649-48f4-aa8f-85dd1c8137cf" path="/var/lib/kubelet/pods/b0fddcb3-3649-48f4-aa8f-85dd1c8137cf/volumes" Oct 02 02:05:14 crc kubenswrapper[4885]: I1002 02:05:14.065438 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5c6b9856b7-vhzx6" event={"ID":"075adfcf-950d-44e6-ab5e-826f3b06c05e","Type":"ContainerStarted","Data":"d4c435f04bfe002cb5fad30dd02fdbe00e86bf8503fe4bac89378dc17919afc7"} Oct 02 02:05:14 crc kubenswrapper[4885]: I1002 02:05:14.240056 4885 scope.go:117] "RemoveContainer" containerID="9a7c1423da72336ecfa2cfe108bb2a287a4349ab5629c7bfb13486150d01b56a" Oct 02 02:05:14 crc kubenswrapper[4885]: I1002 02:05:14.351017 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 02:05:14 crc kubenswrapper[4885]: I1002 02:05:14.351493 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="1aff4106-2f4b-48ef-93a1-d51285ce6242" containerName="glance-log" containerID="cri-o://71ffbb4c2cbbe9022cc1a62d8f0fc6cd2ac0fed0d96f405f4ec7589f182ddbbd" gracePeriod=30 Oct 02 02:05:14 crc kubenswrapper[4885]: I1002 02:05:14.351919 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="1aff4106-2f4b-48ef-93a1-d51285ce6242" containerName="glance-httpd" containerID="cri-o://a9cf43c19d008a62e4225357cfd4c6ca237f24f5da97101bd39435507af05ae8" gracePeriod=30 Oct 02 02:05:14 crc kubenswrapper[4885]: I1002 02:05:14.743559 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-6b5d8cdf54-gr28p"] Oct 02 02:05:14 crc kubenswrapper[4885]: E1002 02:05:14.743912 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0fddcb3-3649-48f4-aa8f-85dd1c8137cf" containerName="horizon-log" Oct 02 02:05:14 crc kubenswrapper[4885]: I1002 02:05:14.743931 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0fddcb3-3649-48f4-aa8f-85dd1c8137cf" containerName="horizon-log" Oct 02 02:05:14 crc kubenswrapper[4885]: E1002 02:05:14.743969 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0fddcb3-3649-48f4-aa8f-85dd1c8137cf" containerName="horizon" Oct 02 02:05:14 crc kubenswrapper[4885]: I1002 02:05:14.743975 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0fddcb3-3649-48f4-aa8f-85dd1c8137cf" containerName="horizon" Oct 02 02:05:14 crc kubenswrapper[4885]: I1002 02:05:14.744139 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="b0fddcb3-3649-48f4-aa8f-85dd1c8137cf" containerName="horizon-log" Oct 02 02:05:14 crc kubenswrapper[4885]: I1002 02:05:14.744161 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="b0fddcb3-3649-48f4-aa8f-85dd1c8137cf" containerName="horizon" Oct 02 02:05:14 crc kubenswrapper[4885]: I1002 02:05:14.745201 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6b5d8cdf54-gr28p" Oct 02 02:05:14 crc kubenswrapper[4885]: I1002 02:05:14.748370 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Oct 02 02:05:14 crc kubenswrapper[4885]: I1002 02:05:14.749400 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Oct 02 02:05:14 crc kubenswrapper[4885]: I1002 02:05:14.758429 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6b5d8cdf54-gr28p"] Oct 02 02:05:14 crc kubenswrapper[4885]: I1002 02:05:14.863137 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vm8dz\" (UniqueName: \"kubernetes.io/projected/af70565f-bc37-43bc-97a3-3942c35a9fb2-kube-api-access-vm8dz\") pod \"barbican-api-6b5d8cdf54-gr28p\" (UID: \"af70565f-bc37-43bc-97a3-3942c35a9fb2\") " pod="openstack/barbican-api-6b5d8cdf54-gr28p" Oct 02 02:05:14 crc kubenswrapper[4885]: I1002 02:05:14.863397 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/af70565f-bc37-43bc-97a3-3942c35a9fb2-internal-tls-certs\") pod \"barbican-api-6b5d8cdf54-gr28p\" (UID: \"af70565f-bc37-43bc-97a3-3942c35a9fb2\") " pod="openstack/barbican-api-6b5d8cdf54-gr28p" Oct 02 02:05:14 crc kubenswrapper[4885]: I1002 02:05:14.863424 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af70565f-bc37-43bc-97a3-3942c35a9fb2-combined-ca-bundle\") pod \"barbican-api-6b5d8cdf54-gr28p\" (UID: \"af70565f-bc37-43bc-97a3-3942c35a9fb2\") " pod="openstack/barbican-api-6b5d8cdf54-gr28p" Oct 02 02:05:14 crc kubenswrapper[4885]: I1002 02:05:14.863468 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/af70565f-bc37-43bc-97a3-3942c35a9fb2-logs\") pod \"barbican-api-6b5d8cdf54-gr28p\" (UID: \"af70565f-bc37-43bc-97a3-3942c35a9fb2\") " pod="openstack/barbican-api-6b5d8cdf54-gr28p" Oct 02 02:05:14 crc kubenswrapper[4885]: I1002 02:05:14.863531 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/af70565f-bc37-43bc-97a3-3942c35a9fb2-public-tls-certs\") pod \"barbican-api-6b5d8cdf54-gr28p\" (UID: \"af70565f-bc37-43bc-97a3-3942c35a9fb2\") " pod="openstack/barbican-api-6b5d8cdf54-gr28p" Oct 02 02:05:14 crc kubenswrapper[4885]: I1002 02:05:14.863562 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/af70565f-bc37-43bc-97a3-3942c35a9fb2-config-data-custom\") pod \"barbican-api-6b5d8cdf54-gr28p\" (UID: \"af70565f-bc37-43bc-97a3-3942c35a9fb2\") " pod="openstack/barbican-api-6b5d8cdf54-gr28p" Oct 02 02:05:14 crc kubenswrapper[4885]: I1002 02:05:14.863594 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af70565f-bc37-43bc-97a3-3942c35a9fb2-config-data\") pod \"barbican-api-6b5d8cdf54-gr28p\" (UID: \"af70565f-bc37-43bc-97a3-3942c35a9fb2\") " pod="openstack/barbican-api-6b5d8cdf54-gr28p" Oct 02 02:05:14 crc kubenswrapper[4885]: I1002 02:05:14.964769 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vm8dz\" (UniqueName: \"kubernetes.io/projected/af70565f-bc37-43bc-97a3-3942c35a9fb2-kube-api-access-vm8dz\") pod \"barbican-api-6b5d8cdf54-gr28p\" (UID: \"af70565f-bc37-43bc-97a3-3942c35a9fb2\") " pod="openstack/barbican-api-6b5d8cdf54-gr28p" Oct 02 02:05:14 crc kubenswrapper[4885]: I1002 02:05:14.964817 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/af70565f-bc37-43bc-97a3-3942c35a9fb2-internal-tls-certs\") pod \"barbican-api-6b5d8cdf54-gr28p\" (UID: \"af70565f-bc37-43bc-97a3-3942c35a9fb2\") " pod="openstack/barbican-api-6b5d8cdf54-gr28p" Oct 02 02:05:14 crc kubenswrapper[4885]: I1002 02:05:14.964846 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af70565f-bc37-43bc-97a3-3942c35a9fb2-combined-ca-bundle\") pod \"barbican-api-6b5d8cdf54-gr28p\" (UID: \"af70565f-bc37-43bc-97a3-3942c35a9fb2\") " pod="openstack/barbican-api-6b5d8cdf54-gr28p" Oct 02 02:05:14 crc kubenswrapper[4885]: I1002 02:05:14.964891 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/af70565f-bc37-43bc-97a3-3942c35a9fb2-logs\") pod \"barbican-api-6b5d8cdf54-gr28p\" (UID: \"af70565f-bc37-43bc-97a3-3942c35a9fb2\") " pod="openstack/barbican-api-6b5d8cdf54-gr28p" Oct 02 02:05:14 crc kubenswrapper[4885]: I1002 02:05:14.964933 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/af70565f-bc37-43bc-97a3-3942c35a9fb2-public-tls-certs\") pod \"barbican-api-6b5d8cdf54-gr28p\" (UID: \"af70565f-bc37-43bc-97a3-3942c35a9fb2\") " pod="openstack/barbican-api-6b5d8cdf54-gr28p" Oct 02 02:05:14 crc kubenswrapper[4885]: I1002 02:05:14.964966 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/af70565f-bc37-43bc-97a3-3942c35a9fb2-config-data-custom\") pod \"barbican-api-6b5d8cdf54-gr28p\" (UID: \"af70565f-bc37-43bc-97a3-3942c35a9fb2\") " pod="openstack/barbican-api-6b5d8cdf54-gr28p" Oct 02 02:05:14 crc kubenswrapper[4885]: I1002 02:05:14.965003 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af70565f-bc37-43bc-97a3-3942c35a9fb2-config-data\") pod \"barbican-api-6b5d8cdf54-gr28p\" (UID: \"af70565f-bc37-43bc-97a3-3942c35a9fb2\") " pod="openstack/barbican-api-6b5d8cdf54-gr28p" Oct 02 02:05:14 crc kubenswrapper[4885]: I1002 02:05:14.965786 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/af70565f-bc37-43bc-97a3-3942c35a9fb2-logs\") pod \"barbican-api-6b5d8cdf54-gr28p\" (UID: \"af70565f-bc37-43bc-97a3-3942c35a9fb2\") " pod="openstack/barbican-api-6b5d8cdf54-gr28p" Oct 02 02:05:14 crc kubenswrapper[4885]: I1002 02:05:14.969935 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af70565f-bc37-43bc-97a3-3942c35a9fb2-combined-ca-bundle\") pod \"barbican-api-6b5d8cdf54-gr28p\" (UID: \"af70565f-bc37-43bc-97a3-3942c35a9fb2\") " pod="openstack/barbican-api-6b5d8cdf54-gr28p" Oct 02 02:05:14 crc kubenswrapper[4885]: I1002 02:05:14.973804 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/af70565f-bc37-43bc-97a3-3942c35a9fb2-public-tls-certs\") pod \"barbican-api-6b5d8cdf54-gr28p\" (UID: \"af70565f-bc37-43bc-97a3-3942c35a9fb2\") " pod="openstack/barbican-api-6b5d8cdf54-gr28p" Oct 02 02:05:14 crc kubenswrapper[4885]: I1002 02:05:14.974748 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/af70565f-bc37-43bc-97a3-3942c35a9fb2-internal-tls-certs\") pod \"barbican-api-6b5d8cdf54-gr28p\" (UID: \"af70565f-bc37-43bc-97a3-3942c35a9fb2\") " pod="openstack/barbican-api-6b5d8cdf54-gr28p" Oct 02 02:05:14 crc kubenswrapper[4885]: I1002 02:05:14.975105 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/af70565f-bc37-43bc-97a3-3942c35a9fb2-config-data-custom\") pod \"barbican-api-6b5d8cdf54-gr28p\" (UID: \"af70565f-bc37-43bc-97a3-3942c35a9fb2\") " pod="openstack/barbican-api-6b5d8cdf54-gr28p" Oct 02 02:05:14 crc kubenswrapper[4885]: I1002 02:05:14.975282 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af70565f-bc37-43bc-97a3-3942c35a9fb2-config-data\") pod \"barbican-api-6b5d8cdf54-gr28p\" (UID: \"af70565f-bc37-43bc-97a3-3942c35a9fb2\") " pod="openstack/barbican-api-6b5d8cdf54-gr28p" Oct 02 02:05:14 crc kubenswrapper[4885]: I1002 02:05:14.982338 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vm8dz\" (UniqueName: \"kubernetes.io/projected/af70565f-bc37-43bc-97a3-3942c35a9fb2-kube-api-access-vm8dz\") pod \"barbican-api-6b5d8cdf54-gr28p\" (UID: \"af70565f-bc37-43bc-97a3-3942c35a9fb2\") " pod="openstack/barbican-api-6b5d8cdf54-gr28p" Oct 02 02:05:15 crc kubenswrapper[4885]: I1002 02:05:15.062219 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6b5d8cdf54-gr28p" Oct 02 02:05:15 crc kubenswrapper[4885]: I1002 02:05:15.079831 4885 generic.go:334] "Generic (PLEG): container finished" podID="1aff4106-2f4b-48ef-93a1-d51285ce6242" containerID="71ffbb4c2cbbe9022cc1a62d8f0fc6cd2ac0fed0d96f405f4ec7589f182ddbbd" exitCode=143 Oct 02 02:05:15 crc kubenswrapper[4885]: I1002 02:05:15.079917 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1aff4106-2f4b-48ef-93a1-d51285ce6242","Type":"ContainerDied","Data":"71ffbb4c2cbbe9022cc1a62d8f0fc6cd2ac0fed0d96f405f4ec7589f182ddbbd"} Oct 02 02:05:15 crc kubenswrapper[4885]: I1002 02:05:15.084865 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c8ddd69c-s4xb6" event={"ID":"bc7146f7-6d12-4bd2-8fd2-0604f358bedf","Type":"ContainerStarted","Data":"aae1c6ba38e1514704db03f0c5e9604bed0365df873bcb5f4016a948b1d19310"} Oct 02 02:05:15 crc kubenswrapper[4885]: I1002 02:05:15.085645 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-75c8ddd69c-s4xb6" Oct 02 02:05:15 crc kubenswrapper[4885]: I1002 02:05:15.108121 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-75c8ddd69c-s4xb6" podStartSLOduration=3.108101122 podStartE2EDuration="3.108101122s" podCreationTimestamp="2025-10-02 02:05:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:05:15.101985566 +0000 UTC m=+1103.913732965" watchObservedRunningTime="2025-10-02 02:05:15.108101122 +0000 UTC m=+1103.919848531" Oct 02 02:05:17 crc kubenswrapper[4885]: I1002 02:05:17.745240 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-tjcz4" Oct 02 02:05:17 crc kubenswrapper[4885]: I1002 02:05:17.918559 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9c66bb6a-5a3c-4223-aba6-c5d9d0ea83ba-etc-machine-id\") pod \"9c66bb6a-5a3c-4223-aba6-c5d9d0ea83ba\" (UID: \"9c66bb6a-5a3c-4223-aba6-c5d9d0ea83ba\") " Oct 02 02:05:17 crc kubenswrapper[4885]: I1002 02:05:17.918617 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqdf4\" (UniqueName: \"kubernetes.io/projected/9c66bb6a-5a3c-4223-aba6-c5d9d0ea83ba-kube-api-access-fqdf4\") pod \"9c66bb6a-5a3c-4223-aba6-c5d9d0ea83ba\" (UID: \"9c66bb6a-5a3c-4223-aba6-c5d9d0ea83ba\") " Oct 02 02:05:17 crc kubenswrapper[4885]: I1002 02:05:17.918654 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c66bb6a-5a3c-4223-aba6-c5d9d0ea83ba-combined-ca-bundle\") pod \"9c66bb6a-5a3c-4223-aba6-c5d9d0ea83ba\" (UID: \"9c66bb6a-5a3c-4223-aba6-c5d9d0ea83ba\") " Oct 02 02:05:17 crc kubenswrapper[4885]: I1002 02:05:17.918691 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c66bb6a-5a3c-4223-aba6-c5d9d0ea83ba-config-data\") pod \"9c66bb6a-5a3c-4223-aba6-c5d9d0ea83ba\" (UID: \"9c66bb6a-5a3c-4223-aba6-c5d9d0ea83ba\") " Oct 02 02:05:17 crc kubenswrapper[4885]: I1002 02:05:17.918812 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/9c66bb6a-5a3c-4223-aba6-c5d9d0ea83ba-db-sync-config-data\") pod \"9c66bb6a-5a3c-4223-aba6-c5d9d0ea83ba\" (UID: \"9c66bb6a-5a3c-4223-aba6-c5d9d0ea83ba\") " Oct 02 02:05:17 crc kubenswrapper[4885]: I1002 02:05:17.918869 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c66bb6a-5a3c-4223-aba6-c5d9d0ea83ba-scripts\") pod \"9c66bb6a-5a3c-4223-aba6-c5d9d0ea83ba\" (UID: \"9c66bb6a-5a3c-4223-aba6-c5d9d0ea83ba\") " Oct 02 02:05:17 crc kubenswrapper[4885]: I1002 02:05:17.921201 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9c66bb6a-5a3c-4223-aba6-c5d9d0ea83ba-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "9c66bb6a-5a3c-4223-aba6-c5d9d0ea83ba" (UID: "9c66bb6a-5a3c-4223-aba6-c5d9d0ea83ba"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 02:05:17 crc kubenswrapper[4885]: I1002 02:05:17.927351 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c66bb6a-5a3c-4223-aba6-c5d9d0ea83ba-kube-api-access-fqdf4" (OuterVolumeSpecName: "kube-api-access-fqdf4") pod "9c66bb6a-5a3c-4223-aba6-c5d9d0ea83ba" (UID: "9c66bb6a-5a3c-4223-aba6-c5d9d0ea83ba"). InnerVolumeSpecName "kube-api-access-fqdf4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:05:17 crc kubenswrapper[4885]: I1002 02:05:17.939518 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c66bb6a-5a3c-4223-aba6-c5d9d0ea83ba-scripts" (OuterVolumeSpecName: "scripts") pod "9c66bb6a-5a3c-4223-aba6-c5d9d0ea83ba" (UID: "9c66bb6a-5a3c-4223-aba6-c5d9d0ea83ba"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:05:17 crc kubenswrapper[4885]: I1002 02:05:17.943648 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c66bb6a-5a3c-4223-aba6-c5d9d0ea83ba-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "9c66bb6a-5a3c-4223-aba6-c5d9d0ea83ba" (UID: "9c66bb6a-5a3c-4223-aba6-c5d9d0ea83ba"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:05:17 crc kubenswrapper[4885]: I1002 02:05:17.948448 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c66bb6a-5a3c-4223-aba6-c5d9d0ea83ba-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9c66bb6a-5a3c-4223-aba6-c5d9d0ea83ba" (UID: "9c66bb6a-5a3c-4223-aba6-c5d9d0ea83ba"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:05:17 crc kubenswrapper[4885]: I1002 02:05:17.999498 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c66bb6a-5a3c-4223-aba6-c5d9d0ea83ba-config-data" (OuterVolumeSpecName: "config-data") pod "9c66bb6a-5a3c-4223-aba6-c5d9d0ea83ba" (UID: "9c66bb6a-5a3c-4223-aba6-c5d9d0ea83ba"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:05:18 crc kubenswrapper[4885]: I1002 02:05:18.021365 4885 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9c66bb6a-5a3c-4223-aba6-c5d9d0ea83ba-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:18 crc kubenswrapper[4885]: I1002 02:05:18.021393 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqdf4\" (UniqueName: \"kubernetes.io/projected/9c66bb6a-5a3c-4223-aba6-c5d9d0ea83ba-kube-api-access-fqdf4\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:18 crc kubenswrapper[4885]: I1002 02:05:18.021406 4885 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c66bb6a-5a3c-4223-aba6-c5d9d0ea83ba-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:18 crc kubenswrapper[4885]: I1002 02:05:18.021415 4885 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c66bb6a-5a3c-4223-aba6-c5d9d0ea83ba-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:18 crc kubenswrapper[4885]: I1002 02:05:18.021424 4885 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/9c66bb6a-5a3c-4223-aba6-c5d9d0ea83ba-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:18 crc kubenswrapper[4885]: I1002 02:05:18.021431 4885 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c66bb6a-5a3c-4223-aba6-c5d9d0ea83ba-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:18 crc kubenswrapper[4885]: I1002 02:05:18.115954 4885 generic.go:334] "Generic (PLEG): container finished" podID="1aff4106-2f4b-48ef-93a1-d51285ce6242" containerID="a9cf43c19d008a62e4225357cfd4c6ca237f24f5da97101bd39435507af05ae8" exitCode=0 Oct 02 02:05:18 crc kubenswrapper[4885]: I1002 02:05:18.116063 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1aff4106-2f4b-48ef-93a1-d51285ce6242","Type":"ContainerDied","Data":"a9cf43c19d008a62e4225357cfd4c6ca237f24f5da97101bd39435507af05ae8"} Oct 02 02:05:18 crc kubenswrapper[4885]: I1002 02:05:18.117192 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 02:05:18 crc kubenswrapper[4885]: I1002 02:05:18.117442 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="345ef025-b8fb-4955-baec-9af9be17792e" containerName="glance-log" containerID="cri-o://0a4f01b0e93fc3d8b30b56a1c50f9a5b6be0c884595ab22522466917a5ed9451" gracePeriod=30 Oct 02 02:05:18 crc kubenswrapper[4885]: I1002 02:05:18.117553 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="345ef025-b8fb-4955-baec-9af9be17792e" containerName="glance-httpd" containerID="cri-o://c2dd3394b4b25538f87dffc1aa40df5a3104efa6e9908de25784a09ee89b7f7b" gracePeriod=30 Oct 02 02:05:18 crc kubenswrapper[4885]: I1002 02:05:18.120217 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-tjcz4" event={"ID":"9c66bb6a-5a3c-4223-aba6-c5d9d0ea83ba","Type":"ContainerDied","Data":"a265f07b2a178becf2d4f64243b607392df0bcc0cd1797f70e9daa3dd886a014"} Oct 02 02:05:18 crc kubenswrapper[4885]: I1002 02:05:18.120243 4885 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a265f07b2a178becf2d4f64243b607392df0bcc0cd1797f70e9daa3dd886a014" Oct 02 02:05:18 crc kubenswrapper[4885]: I1002 02:05:18.120255 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-tjcz4" Oct 02 02:05:19 crc kubenswrapper[4885]: I1002 02:05:19.142898 4885 generic.go:334] "Generic (PLEG): container finished" podID="345ef025-b8fb-4955-baec-9af9be17792e" containerID="0a4f01b0e93fc3d8b30b56a1c50f9a5b6be0c884595ab22522466917a5ed9451" exitCode=143 Oct 02 02:05:19 crc kubenswrapper[4885]: I1002 02:05:19.143015 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"345ef025-b8fb-4955-baec-9af9be17792e","Type":"ContainerDied","Data":"0a4f01b0e93fc3d8b30b56a1c50f9a5b6be0c884595ab22522466917a5ed9451"} Oct 02 02:05:19 crc kubenswrapper[4885]: I1002 02:05:19.214406 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-s4xb6"] Oct 02 02:05:19 crc kubenswrapper[4885]: I1002 02:05:19.214603 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-75c8ddd69c-s4xb6" podUID="bc7146f7-6d12-4bd2-8fd2-0604f358bedf" containerName="dnsmasq-dns" containerID="cri-o://aae1c6ba38e1514704db03f0c5e9604bed0365df873bcb5f4016a948b1d19310" gracePeriod=10 Oct 02 02:05:19 crc kubenswrapper[4885]: I1002 02:05:19.216434 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-75c8ddd69c-s4xb6" Oct 02 02:05:19 crc kubenswrapper[4885]: I1002 02:05:19.243528 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-85ntb"] Oct 02 02:05:19 crc kubenswrapper[4885]: E1002 02:05:19.243888 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c66bb6a-5a3c-4223-aba6-c5d9d0ea83ba" containerName="cinder-db-sync" Oct 02 02:05:19 crc kubenswrapper[4885]: I1002 02:05:19.243904 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c66bb6a-5a3c-4223-aba6-c5d9d0ea83ba" containerName="cinder-db-sync" Oct 02 02:05:19 crc kubenswrapper[4885]: I1002 02:05:19.244099 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c66bb6a-5a3c-4223-aba6-c5d9d0ea83ba" containerName="cinder-db-sync" Oct 02 02:05:19 crc kubenswrapper[4885]: I1002 02:05:19.244989 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-85ntb" Oct 02 02:05:19 crc kubenswrapper[4885]: I1002 02:05:19.306563 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-85ntb"] Oct 02 02:05:19 crc kubenswrapper[4885]: I1002 02:05:19.355489 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8h5f4\" (UniqueName: \"kubernetes.io/projected/7a9422e4-cbc9-4eb7-885d-eb42b1017cab-kube-api-access-8h5f4\") pod \"dnsmasq-dns-5784cf869f-85ntb\" (UID: \"7a9422e4-cbc9-4eb7-885d-eb42b1017cab\") " pod="openstack/dnsmasq-dns-5784cf869f-85ntb" Oct 02 02:05:19 crc kubenswrapper[4885]: I1002 02:05:19.355716 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7a9422e4-cbc9-4eb7-885d-eb42b1017cab-ovsdbserver-nb\") pod \"dnsmasq-dns-5784cf869f-85ntb\" (UID: \"7a9422e4-cbc9-4eb7-885d-eb42b1017cab\") " pod="openstack/dnsmasq-dns-5784cf869f-85ntb" Oct 02 02:05:19 crc kubenswrapper[4885]: I1002 02:05:19.355792 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7a9422e4-cbc9-4eb7-885d-eb42b1017cab-dns-svc\") pod \"dnsmasq-dns-5784cf869f-85ntb\" (UID: \"7a9422e4-cbc9-4eb7-885d-eb42b1017cab\") " pod="openstack/dnsmasq-dns-5784cf869f-85ntb" Oct 02 02:05:19 crc kubenswrapper[4885]: I1002 02:05:19.355900 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a9422e4-cbc9-4eb7-885d-eb42b1017cab-config\") pod \"dnsmasq-dns-5784cf869f-85ntb\" (UID: \"7a9422e4-cbc9-4eb7-885d-eb42b1017cab\") " pod="openstack/dnsmasq-dns-5784cf869f-85ntb" Oct 02 02:05:19 crc kubenswrapper[4885]: I1002 02:05:19.355967 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7a9422e4-cbc9-4eb7-885d-eb42b1017cab-ovsdbserver-sb\") pod \"dnsmasq-dns-5784cf869f-85ntb\" (UID: \"7a9422e4-cbc9-4eb7-885d-eb42b1017cab\") " pod="openstack/dnsmasq-dns-5784cf869f-85ntb" Oct 02 02:05:19 crc kubenswrapper[4885]: I1002 02:05:19.356046 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7a9422e4-cbc9-4eb7-885d-eb42b1017cab-dns-swift-storage-0\") pod \"dnsmasq-dns-5784cf869f-85ntb\" (UID: \"7a9422e4-cbc9-4eb7-885d-eb42b1017cab\") " pod="openstack/dnsmasq-dns-5784cf869f-85ntb" Oct 02 02:05:19 crc kubenswrapper[4885]: I1002 02:05:19.359939 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 02:05:19 crc kubenswrapper[4885]: I1002 02:05:19.372999 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 02:05:19 crc kubenswrapper[4885]: I1002 02:05:19.373099 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 02 02:05:19 crc kubenswrapper[4885]: I1002 02:05:19.384200 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 02 02:05:19 crc kubenswrapper[4885]: I1002 02:05:19.385033 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 02 02:05:19 crc kubenswrapper[4885]: I1002 02:05:19.385114 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 02 02:05:19 crc kubenswrapper[4885]: I1002 02:05:19.385241 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-sdgb9" Oct 02 02:05:19 crc kubenswrapper[4885]: I1002 02:05:19.410308 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 02 02:05:19 crc kubenswrapper[4885]: I1002 02:05:19.411917 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 02 02:05:19 crc kubenswrapper[4885]: I1002 02:05:19.414460 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 02 02:05:19 crc kubenswrapper[4885]: I1002 02:05:19.419844 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 02 02:05:19 crc kubenswrapper[4885]: I1002 02:05:19.458102 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8h5f4\" (UniqueName: \"kubernetes.io/projected/7a9422e4-cbc9-4eb7-885d-eb42b1017cab-kube-api-access-8h5f4\") pod \"dnsmasq-dns-5784cf869f-85ntb\" (UID: \"7a9422e4-cbc9-4eb7-885d-eb42b1017cab\") " pod="openstack/dnsmasq-dns-5784cf869f-85ntb" Oct 02 02:05:19 crc kubenswrapper[4885]: I1002 02:05:19.458143 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7a9422e4-cbc9-4eb7-885d-eb42b1017cab-ovsdbserver-nb\") pod \"dnsmasq-dns-5784cf869f-85ntb\" (UID: \"7a9422e4-cbc9-4eb7-885d-eb42b1017cab\") " pod="openstack/dnsmasq-dns-5784cf869f-85ntb" Oct 02 02:05:19 crc kubenswrapper[4885]: I1002 02:05:19.458161 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7a9422e4-cbc9-4eb7-885d-eb42b1017cab-dns-svc\") pod \"dnsmasq-dns-5784cf869f-85ntb\" (UID: \"7a9422e4-cbc9-4eb7-885d-eb42b1017cab\") " pod="openstack/dnsmasq-dns-5784cf869f-85ntb" Oct 02 02:05:19 crc kubenswrapper[4885]: I1002 02:05:19.458182 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a9422e4-cbc9-4eb7-885d-eb42b1017cab-config\") pod \"dnsmasq-dns-5784cf869f-85ntb\" (UID: \"7a9422e4-cbc9-4eb7-885d-eb42b1017cab\") " pod="openstack/dnsmasq-dns-5784cf869f-85ntb" Oct 02 02:05:19 crc kubenswrapper[4885]: I1002 02:05:19.458197 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7a9422e4-cbc9-4eb7-885d-eb42b1017cab-ovsdbserver-sb\") pod \"dnsmasq-dns-5784cf869f-85ntb\" (UID: \"7a9422e4-cbc9-4eb7-885d-eb42b1017cab\") " pod="openstack/dnsmasq-dns-5784cf869f-85ntb" Oct 02 02:05:19 crc kubenswrapper[4885]: I1002 02:05:19.458219 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7a9422e4-cbc9-4eb7-885d-eb42b1017cab-dns-swift-storage-0\") pod \"dnsmasq-dns-5784cf869f-85ntb\" (UID: \"7a9422e4-cbc9-4eb7-885d-eb42b1017cab\") " pod="openstack/dnsmasq-dns-5784cf869f-85ntb" Oct 02 02:05:19 crc kubenswrapper[4885]: I1002 02:05:19.458983 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7a9422e4-cbc9-4eb7-885d-eb42b1017cab-dns-swift-storage-0\") pod \"dnsmasq-dns-5784cf869f-85ntb\" (UID: \"7a9422e4-cbc9-4eb7-885d-eb42b1017cab\") " pod="openstack/dnsmasq-dns-5784cf869f-85ntb" Oct 02 02:05:19 crc kubenswrapper[4885]: I1002 02:05:19.459745 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7a9422e4-cbc9-4eb7-885d-eb42b1017cab-ovsdbserver-nb\") pod \"dnsmasq-dns-5784cf869f-85ntb\" (UID: \"7a9422e4-cbc9-4eb7-885d-eb42b1017cab\") " pod="openstack/dnsmasq-dns-5784cf869f-85ntb" Oct 02 02:05:19 crc kubenswrapper[4885]: I1002 02:05:19.460141 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a9422e4-cbc9-4eb7-885d-eb42b1017cab-config\") pod \"dnsmasq-dns-5784cf869f-85ntb\" (UID: \"7a9422e4-cbc9-4eb7-885d-eb42b1017cab\") " pod="openstack/dnsmasq-dns-5784cf869f-85ntb" Oct 02 02:05:19 crc kubenswrapper[4885]: I1002 02:05:19.460313 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7a9422e4-cbc9-4eb7-885d-eb42b1017cab-ovsdbserver-sb\") pod \"dnsmasq-dns-5784cf869f-85ntb\" (UID: \"7a9422e4-cbc9-4eb7-885d-eb42b1017cab\") " pod="openstack/dnsmasq-dns-5784cf869f-85ntb" Oct 02 02:05:19 crc kubenswrapper[4885]: I1002 02:05:19.460571 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7a9422e4-cbc9-4eb7-885d-eb42b1017cab-dns-svc\") pod \"dnsmasq-dns-5784cf869f-85ntb\" (UID: \"7a9422e4-cbc9-4eb7-885d-eb42b1017cab\") " pod="openstack/dnsmasq-dns-5784cf869f-85ntb" Oct 02 02:05:19 crc kubenswrapper[4885]: I1002 02:05:19.481076 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8h5f4\" (UniqueName: \"kubernetes.io/projected/7a9422e4-cbc9-4eb7-885d-eb42b1017cab-kube-api-access-8h5f4\") pod \"dnsmasq-dns-5784cf869f-85ntb\" (UID: \"7a9422e4-cbc9-4eb7-885d-eb42b1017cab\") " pod="openstack/dnsmasq-dns-5784cf869f-85ntb" Oct 02 02:05:19 crc kubenswrapper[4885]: I1002 02:05:19.559431 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f985ab4a-2149-40c7-9af9-5ac242974f28-config-data-custom\") pod \"cinder-api-0\" (UID: \"f985ab4a-2149-40c7-9af9-5ac242974f28\") " pod="openstack/cinder-api-0" Oct 02 02:05:19 crc kubenswrapper[4885]: I1002 02:05:19.559526 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hc8gw\" (UniqueName: \"kubernetes.io/projected/f985ab4a-2149-40c7-9af9-5ac242974f28-kube-api-access-hc8gw\") pod \"cinder-api-0\" (UID: \"f985ab4a-2149-40c7-9af9-5ac242974f28\") " pod="openstack/cinder-api-0" Oct 02 02:05:19 crc kubenswrapper[4885]: I1002 02:05:19.559546 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f985ab4a-2149-40c7-9af9-5ac242974f28-scripts\") pod \"cinder-api-0\" (UID: \"f985ab4a-2149-40c7-9af9-5ac242974f28\") " pod="openstack/cinder-api-0" Oct 02 02:05:19 crc kubenswrapper[4885]: I1002 02:05:19.559598 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/72053d12-6b1a-4df1-9fbe-ba2973d53dc3-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"72053d12-6b1a-4df1-9fbe-ba2973d53dc3\") " pod="openstack/cinder-scheduler-0" Oct 02 02:05:19 crc kubenswrapper[4885]: I1002 02:05:19.559627 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72053d12-6b1a-4df1-9fbe-ba2973d53dc3-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"72053d12-6b1a-4df1-9fbe-ba2973d53dc3\") " pod="openstack/cinder-scheduler-0" Oct 02 02:05:19 crc kubenswrapper[4885]: I1002 02:05:19.559666 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/72053d12-6b1a-4df1-9fbe-ba2973d53dc3-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"72053d12-6b1a-4df1-9fbe-ba2973d53dc3\") " pod="openstack/cinder-scheduler-0" Oct 02 02:05:19 crc kubenswrapper[4885]: I1002 02:05:19.559695 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f985ab4a-2149-40c7-9af9-5ac242974f28-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"f985ab4a-2149-40c7-9af9-5ac242974f28\") " pod="openstack/cinder-api-0" Oct 02 02:05:19 crc kubenswrapper[4885]: I1002 02:05:19.559749 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f985ab4a-2149-40c7-9af9-5ac242974f28-config-data\") pod \"cinder-api-0\" (UID: \"f985ab4a-2149-40c7-9af9-5ac242974f28\") " pod="openstack/cinder-api-0" Oct 02 02:05:19 crc kubenswrapper[4885]: I1002 02:05:19.559802 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f985ab4a-2149-40c7-9af9-5ac242974f28-etc-machine-id\") pod \"cinder-api-0\" (UID: \"f985ab4a-2149-40c7-9af9-5ac242974f28\") " pod="openstack/cinder-api-0" Oct 02 02:05:19 crc kubenswrapper[4885]: I1002 02:05:19.559822 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72053d12-6b1a-4df1-9fbe-ba2973d53dc3-config-data\") pod \"cinder-scheduler-0\" (UID: \"72053d12-6b1a-4df1-9fbe-ba2973d53dc3\") " pod="openstack/cinder-scheduler-0" Oct 02 02:05:19 crc kubenswrapper[4885]: I1002 02:05:19.559841 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/72053d12-6b1a-4df1-9fbe-ba2973d53dc3-scripts\") pod \"cinder-scheduler-0\" (UID: \"72053d12-6b1a-4df1-9fbe-ba2973d53dc3\") " pod="openstack/cinder-scheduler-0" Oct 02 02:05:19 crc kubenswrapper[4885]: I1002 02:05:19.559918 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f985ab4a-2149-40c7-9af9-5ac242974f28-logs\") pod \"cinder-api-0\" (UID: \"f985ab4a-2149-40c7-9af9-5ac242974f28\") " pod="openstack/cinder-api-0" Oct 02 02:05:19 crc kubenswrapper[4885]: I1002 02:05:19.560117 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wk4tq\" (UniqueName: \"kubernetes.io/projected/72053d12-6b1a-4df1-9fbe-ba2973d53dc3-kube-api-access-wk4tq\") pod \"cinder-scheduler-0\" (UID: \"72053d12-6b1a-4df1-9fbe-ba2973d53dc3\") " pod="openstack/cinder-scheduler-0" Oct 02 02:05:19 crc kubenswrapper[4885]: I1002 02:05:19.567109 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-85ntb" Oct 02 02:05:19 crc kubenswrapper[4885]: I1002 02:05:19.625816 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-7474b946b7-s2w7j" Oct 02 02:05:19 crc kubenswrapper[4885]: I1002 02:05:19.628775 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-7474b946b7-s2w7j" Oct 02 02:05:19 crc kubenswrapper[4885]: I1002 02:05:19.662187 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f985ab4a-2149-40c7-9af9-5ac242974f28-config-data\") pod \"cinder-api-0\" (UID: \"f985ab4a-2149-40c7-9af9-5ac242974f28\") " pod="openstack/cinder-api-0" Oct 02 02:05:19 crc kubenswrapper[4885]: I1002 02:05:19.662256 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f985ab4a-2149-40c7-9af9-5ac242974f28-etc-machine-id\") pod \"cinder-api-0\" (UID: \"f985ab4a-2149-40c7-9af9-5ac242974f28\") " pod="openstack/cinder-api-0" Oct 02 02:05:19 crc kubenswrapper[4885]: I1002 02:05:19.662309 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72053d12-6b1a-4df1-9fbe-ba2973d53dc3-config-data\") pod \"cinder-scheduler-0\" (UID: \"72053d12-6b1a-4df1-9fbe-ba2973d53dc3\") " pod="openstack/cinder-scheduler-0" Oct 02 02:05:19 crc kubenswrapper[4885]: I1002 02:05:19.662338 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/72053d12-6b1a-4df1-9fbe-ba2973d53dc3-scripts\") pod \"cinder-scheduler-0\" (UID: \"72053d12-6b1a-4df1-9fbe-ba2973d53dc3\") " pod="openstack/cinder-scheduler-0" Oct 02 02:05:19 crc kubenswrapper[4885]: I1002 02:05:19.663005 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f985ab4a-2149-40c7-9af9-5ac242974f28-etc-machine-id\") pod \"cinder-api-0\" (UID: \"f985ab4a-2149-40c7-9af9-5ac242974f28\") " pod="openstack/cinder-api-0" Oct 02 02:05:19 crc kubenswrapper[4885]: I1002 02:05:19.663031 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f985ab4a-2149-40c7-9af9-5ac242974f28-logs\") pod \"cinder-api-0\" (UID: \"f985ab4a-2149-40c7-9af9-5ac242974f28\") " pod="openstack/cinder-api-0" Oct 02 02:05:19 crc kubenswrapper[4885]: I1002 02:05:19.663102 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wk4tq\" (UniqueName: \"kubernetes.io/projected/72053d12-6b1a-4df1-9fbe-ba2973d53dc3-kube-api-access-wk4tq\") pod \"cinder-scheduler-0\" (UID: \"72053d12-6b1a-4df1-9fbe-ba2973d53dc3\") " pod="openstack/cinder-scheduler-0" Oct 02 02:05:19 crc kubenswrapper[4885]: I1002 02:05:19.663139 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f985ab4a-2149-40c7-9af9-5ac242974f28-config-data-custom\") pod \"cinder-api-0\" (UID: \"f985ab4a-2149-40c7-9af9-5ac242974f28\") " pod="openstack/cinder-api-0" Oct 02 02:05:19 crc kubenswrapper[4885]: I1002 02:05:19.663167 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hc8gw\" (UniqueName: \"kubernetes.io/projected/f985ab4a-2149-40c7-9af9-5ac242974f28-kube-api-access-hc8gw\") pod \"cinder-api-0\" (UID: \"f985ab4a-2149-40c7-9af9-5ac242974f28\") " pod="openstack/cinder-api-0" Oct 02 02:05:19 crc kubenswrapper[4885]: I1002 02:05:19.663188 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f985ab4a-2149-40c7-9af9-5ac242974f28-scripts\") pod \"cinder-api-0\" (UID: \"f985ab4a-2149-40c7-9af9-5ac242974f28\") " pod="openstack/cinder-api-0" Oct 02 02:05:19 crc kubenswrapper[4885]: I1002 02:05:19.663212 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/72053d12-6b1a-4df1-9fbe-ba2973d53dc3-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"72053d12-6b1a-4df1-9fbe-ba2973d53dc3\") " pod="openstack/cinder-scheduler-0" Oct 02 02:05:19 crc kubenswrapper[4885]: I1002 02:05:19.663242 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72053d12-6b1a-4df1-9fbe-ba2973d53dc3-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"72053d12-6b1a-4df1-9fbe-ba2973d53dc3\") " pod="openstack/cinder-scheduler-0" Oct 02 02:05:19 crc kubenswrapper[4885]: I1002 02:05:19.663282 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/72053d12-6b1a-4df1-9fbe-ba2973d53dc3-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"72053d12-6b1a-4df1-9fbe-ba2973d53dc3\") " pod="openstack/cinder-scheduler-0" Oct 02 02:05:19 crc kubenswrapper[4885]: I1002 02:05:19.663335 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f985ab4a-2149-40c7-9af9-5ac242974f28-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"f985ab4a-2149-40c7-9af9-5ac242974f28\") " pod="openstack/cinder-api-0" Oct 02 02:05:19 crc kubenswrapper[4885]: I1002 02:05:19.663442 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f985ab4a-2149-40c7-9af9-5ac242974f28-logs\") pod \"cinder-api-0\" (UID: \"f985ab4a-2149-40c7-9af9-5ac242974f28\") " pod="openstack/cinder-api-0" Oct 02 02:05:19 crc kubenswrapper[4885]: I1002 02:05:19.666526 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/72053d12-6b1a-4df1-9fbe-ba2973d53dc3-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"72053d12-6b1a-4df1-9fbe-ba2973d53dc3\") " pod="openstack/cinder-scheduler-0" Oct 02 02:05:19 crc kubenswrapper[4885]: I1002 02:05:19.666757 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/72053d12-6b1a-4df1-9fbe-ba2973d53dc3-scripts\") pod \"cinder-scheduler-0\" (UID: \"72053d12-6b1a-4df1-9fbe-ba2973d53dc3\") " pod="openstack/cinder-scheduler-0" Oct 02 02:05:19 crc kubenswrapper[4885]: I1002 02:05:19.666992 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72053d12-6b1a-4df1-9fbe-ba2973d53dc3-config-data\") pod \"cinder-scheduler-0\" (UID: \"72053d12-6b1a-4df1-9fbe-ba2973d53dc3\") " pod="openstack/cinder-scheduler-0" Oct 02 02:05:19 crc kubenswrapper[4885]: I1002 02:05:19.667082 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f985ab4a-2149-40c7-9af9-5ac242974f28-config-data\") pod \"cinder-api-0\" (UID: \"f985ab4a-2149-40c7-9af9-5ac242974f28\") " pod="openstack/cinder-api-0" Oct 02 02:05:19 crc kubenswrapper[4885]: I1002 02:05:19.667928 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f985ab4a-2149-40c7-9af9-5ac242974f28-scripts\") pod \"cinder-api-0\" (UID: \"f985ab4a-2149-40c7-9af9-5ac242974f28\") " pod="openstack/cinder-api-0" Oct 02 02:05:19 crc kubenswrapper[4885]: I1002 02:05:19.668127 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f985ab4a-2149-40c7-9af9-5ac242974f28-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"f985ab4a-2149-40c7-9af9-5ac242974f28\") " pod="openstack/cinder-api-0" Oct 02 02:05:19 crc kubenswrapper[4885]: I1002 02:05:19.669494 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/72053d12-6b1a-4df1-9fbe-ba2973d53dc3-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"72053d12-6b1a-4df1-9fbe-ba2973d53dc3\") " pod="openstack/cinder-scheduler-0" Oct 02 02:05:19 crc kubenswrapper[4885]: I1002 02:05:19.670392 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f985ab4a-2149-40c7-9af9-5ac242974f28-config-data-custom\") pod \"cinder-api-0\" (UID: \"f985ab4a-2149-40c7-9af9-5ac242974f28\") " pod="openstack/cinder-api-0" Oct 02 02:05:19 crc kubenswrapper[4885]: I1002 02:05:19.693857 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72053d12-6b1a-4df1-9fbe-ba2973d53dc3-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"72053d12-6b1a-4df1-9fbe-ba2973d53dc3\") " pod="openstack/cinder-scheduler-0" Oct 02 02:05:19 crc kubenswrapper[4885]: I1002 02:05:19.699873 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wk4tq\" (UniqueName: \"kubernetes.io/projected/72053d12-6b1a-4df1-9fbe-ba2973d53dc3-kube-api-access-wk4tq\") pod \"cinder-scheduler-0\" (UID: \"72053d12-6b1a-4df1-9fbe-ba2973d53dc3\") " pod="openstack/cinder-scheduler-0" Oct 02 02:05:19 crc kubenswrapper[4885]: I1002 02:05:19.700359 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 02 02:05:19 crc kubenswrapper[4885]: I1002 02:05:19.708982 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hc8gw\" (UniqueName: \"kubernetes.io/projected/f985ab4a-2149-40c7-9af9-5ac242974f28-kube-api-access-hc8gw\") pod \"cinder-api-0\" (UID: \"f985ab4a-2149-40c7-9af9-5ac242974f28\") " pod="openstack/cinder-api-0" Oct 02 02:05:19 crc kubenswrapper[4885]: I1002 02:05:19.741361 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 02 02:05:20 crc kubenswrapper[4885]: I1002 02:05:20.154112 4885 generic.go:334] "Generic (PLEG): container finished" podID="bc7146f7-6d12-4bd2-8fd2-0604f358bedf" containerID="aae1c6ba38e1514704db03f0c5e9604bed0365df873bcb5f4016a948b1d19310" exitCode=0 Oct 02 02:05:20 crc kubenswrapper[4885]: I1002 02:05:20.154190 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c8ddd69c-s4xb6" event={"ID":"bc7146f7-6d12-4bd2-8fd2-0604f358bedf","Type":"ContainerDied","Data":"aae1c6ba38e1514704db03f0c5e9604bed0365df873bcb5f4016a948b1d19310"} Oct 02 02:05:21 crc kubenswrapper[4885]: I1002 02:05:21.459105 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 02 02:05:21 crc kubenswrapper[4885]: I1002 02:05:21.613889 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1aff4106-2f4b-48ef-93a1-d51285ce6242-logs\") pod \"1aff4106-2f4b-48ef-93a1-d51285ce6242\" (UID: \"1aff4106-2f4b-48ef-93a1-d51285ce6242\") " Oct 02 02:05:21 crc kubenswrapper[4885]: I1002 02:05:21.614662 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1aff4106-2f4b-48ef-93a1-d51285ce6242-httpd-run\") pod \"1aff4106-2f4b-48ef-93a1-d51285ce6242\" (UID: \"1aff4106-2f4b-48ef-93a1-d51285ce6242\") " Oct 02 02:05:21 crc kubenswrapper[4885]: I1002 02:05:21.614700 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1aff4106-2f4b-48ef-93a1-d51285ce6242-internal-tls-certs\") pod \"1aff4106-2f4b-48ef-93a1-d51285ce6242\" (UID: \"1aff4106-2f4b-48ef-93a1-d51285ce6242\") " Oct 02 02:05:21 crc kubenswrapper[4885]: I1002 02:05:21.614788 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"1aff4106-2f4b-48ef-93a1-d51285ce6242\" (UID: \"1aff4106-2f4b-48ef-93a1-d51285ce6242\") " Oct 02 02:05:21 crc kubenswrapper[4885]: I1002 02:05:21.614815 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ztxwm\" (UniqueName: \"kubernetes.io/projected/1aff4106-2f4b-48ef-93a1-d51285ce6242-kube-api-access-ztxwm\") pod \"1aff4106-2f4b-48ef-93a1-d51285ce6242\" (UID: \"1aff4106-2f4b-48ef-93a1-d51285ce6242\") " Oct 02 02:05:21 crc kubenswrapper[4885]: I1002 02:05:21.614883 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1aff4106-2f4b-48ef-93a1-d51285ce6242-combined-ca-bundle\") pod \"1aff4106-2f4b-48ef-93a1-d51285ce6242\" (UID: \"1aff4106-2f4b-48ef-93a1-d51285ce6242\") " Oct 02 02:05:21 crc kubenswrapper[4885]: I1002 02:05:21.614937 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1aff4106-2f4b-48ef-93a1-d51285ce6242-config-data\") pod \"1aff4106-2f4b-48ef-93a1-d51285ce6242\" (UID: \"1aff4106-2f4b-48ef-93a1-d51285ce6242\") " Oct 02 02:05:21 crc kubenswrapper[4885]: I1002 02:05:21.614975 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1aff4106-2f4b-48ef-93a1-d51285ce6242-scripts\") pod \"1aff4106-2f4b-48ef-93a1-d51285ce6242\" (UID: \"1aff4106-2f4b-48ef-93a1-d51285ce6242\") " Oct 02 02:05:21 crc kubenswrapper[4885]: I1002 02:05:21.615239 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1aff4106-2f4b-48ef-93a1-d51285ce6242-logs" (OuterVolumeSpecName: "logs") pod "1aff4106-2f4b-48ef-93a1-d51285ce6242" (UID: "1aff4106-2f4b-48ef-93a1-d51285ce6242"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:05:21 crc kubenswrapper[4885]: I1002 02:05:21.615542 4885 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1aff4106-2f4b-48ef-93a1-d51285ce6242-logs\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:21 crc kubenswrapper[4885]: I1002 02:05:21.615770 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1aff4106-2f4b-48ef-93a1-d51285ce6242-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "1aff4106-2f4b-48ef-93a1-d51285ce6242" (UID: "1aff4106-2f4b-48ef-93a1-d51285ce6242"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:05:21 crc kubenswrapper[4885]: I1002 02:05:21.616403 4885 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-6c87b5899d-lvdbp" podUID="bb9e27f4-9f80-40ff-af3e-fe5aca9f5755" containerName="barbican-api-log" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 02:05:21 crc kubenswrapper[4885]: I1002 02:05:21.626299 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "glance") pod "1aff4106-2f4b-48ef-93a1-d51285ce6242" (UID: "1aff4106-2f4b-48ef-93a1-d51285ce6242"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 02 02:05:21 crc kubenswrapper[4885]: I1002 02:05:21.638692 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1aff4106-2f4b-48ef-93a1-d51285ce6242-scripts" (OuterVolumeSpecName: "scripts") pod "1aff4106-2f4b-48ef-93a1-d51285ce6242" (UID: "1aff4106-2f4b-48ef-93a1-d51285ce6242"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:05:21 crc kubenswrapper[4885]: I1002 02:05:21.658313 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1aff4106-2f4b-48ef-93a1-d51285ce6242-kube-api-access-ztxwm" (OuterVolumeSpecName: "kube-api-access-ztxwm") pod "1aff4106-2f4b-48ef-93a1-d51285ce6242" (UID: "1aff4106-2f4b-48ef-93a1-d51285ce6242"). InnerVolumeSpecName "kube-api-access-ztxwm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:05:21 crc kubenswrapper[4885]: I1002 02:05:21.677272 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1aff4106-2f4b-48ef-93a1-d51285ce6242-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1aff4106-2f4b-48ef-93a1-d51285ce6242" (UID: "1aff4106-2f4b-48ef-93a1-d51285ce6242"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:05:21 crc kubenswrapper[4885]: I1002 02:05:21.715123 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1aff4106-2f4b-48ef-93a1-d51285ce6242-config-data" (OuterVolumeSpecName: "config-data") pod "1aff4106-2f4b-48ef-93a1-d51285ce6242" (UID: "1aff4106-2f4b-48ef-93a1-d51285ce6242"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:05:21 crc kubenswrapper[4885]: I1002 02:05:21.719673 4885 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1aff4106-2f4b-48ef-93a1-d51285ce6242-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:21 crc kubenswrapper[4885]: I1002 02:05:21.719889 4885 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Oct 02 02:05:21 crc kubenswrapper[4885]: I1002 02:05:21.719899 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ztxwm\" (UniqueName: \"kubernetes.io/projected/1aff4106-2f4b-48ef-93a1-d51285ce6242-kube-api-access-ztxwm\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:21 crc kubenswrapper[4885]: I1002 02:05:21.719909 4885 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1aff4106-2f4b-48ef-93a1-d51285ce6242-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:21 crc kubenswrapper[4885]: I1002 02:05:21.719917 4885 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1aff4106-2f4b-48ef-93a1-d51285ce6242-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:21 crc kubenswrapper[4885]: I1002 02:05:21.719927 4885 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1aff4106-2f4b-48ef-93a1-d51285ce6242-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:21 crc kubenswrapper[4885]: I1002 02:05:21.758501 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 02 02:05:21 crc kubenswrapper[4885]: I1002 02:05:21.767764 4885 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Oct 02 02:05:21 crc kubenswrapper[4885]: I1002 02:05:21.791400 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1aff4106-2f4b-48ef-93a1-d51285ce6242-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "1aff4106-2f4b-48ef-93a1-d51285ce6242" (UID: "1aff4106-2f4b-48ef-93a1-d51285ce6242"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:05:21 crc kubenswrapper[4885]: I1002 02:05:21.825184 4885 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1aff4106-2f4b-48ef-93a1-d51285ce6242-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:21 crc kubenswrapper[4885]: I1002 02:05:21.825214 4885 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:21 crc kubenswrapper[4885]: I1002 02:05:21.883436 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75c8ddd69c-s4xb6" Oct 02 02:05:21 crc kubenswrapper[4885]: I1002 02:05:21.931144 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bc7146f7-6d12-4bd2-8fd2-0604f358bedf-config\") pod \"bc7146f7-6d12-4bd2-8fd2-0604f358bedf\" (UID: \"bc7146f7-6d12-4bd2-8fd2-0604f358bedf\") " Oct 02 02:05:21 crc kubenswrapper[4885]: I1002 02:05:21.931184 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xf65m\" (UniqueName: \"kubernetes.io/projected/bc7146f7-6d12-4bd2-8fd2-0604f358bedf-kube-api-access-xf65m\") pod \"bc7146f7-6d12-4bd2-8fd2-0604f358bedf\" (UID: \"bc7146f7-6d12-4bd2-8fd2-0604f358bedf\") " Oct 02 02:05:21 crc kubenswrapper[4885]: I1002 02:05:21.931226 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bc7146f7-6d12-4bd2-8fd2-0604f358bedf-ovsdbserver-sb\") pod \"bc7146f7-6d12-4bd2-8fd2-0604f358bedf\" (UID: \"bc7146f7-6d12-4bd2-8fd2-0604f358bedf\") " Oct 02 02:05:21 crc kubenswrapper[4885]: I1002 02:05:21.935288 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bc7146f7-6d12-4bd2-8fd2-0604f358bedf-dns-svc\") pod \"bc7146f7-6d12-4bd2-8fd2-0604f358bedf\" (UID: \"bc7146f7-6d12-4bd2-8fd2-0604f358bedf\") " Oct 02 02:05:21 crc kubenswrapper[4885]: I1002 02:05:21.935371 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bc7146f7-6d12-4bd2-8fd2-0604f358bedf-dns-swift-storage-0\") pod \"bc7146f7-6d12-4bd2-8fd2-0604f358bedf\" (UID: \"bc7146f7-6d12-4bd2-8fd2-0604f358bedf\") " Oct 02 02:05:21 crc kubenswrapper[4885]: I1002 02:05:21.935414 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bc7146f7-6d12-4bd2-8fd2-0604f358bedf-ovsdbserver-nb\") pod \"bc7146f7-6d12-4bd2-8fd2-0604f358bedf\" (UID: \"bc7146f7-6d12-4bd2-8fd2-0604f358bedf\") " Oct 02 02:05:21 crc kubenswrapper[4885]: I1002 02:05:21.959054 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc7146f7-6d12-4bd2-8fd2-0604f358bedf-kube-api-access-xf65m" (OuterVolumeSpecName: "kube-api-access-xf65m") pod "bc7146f7-6d12-4bd2-8fd2-0604f358bedf" (UID: "bc7146f7-6d12-4bd2-8fd2-0604f358bedf"). InnerVolumeSpecName "kube-api-access-xf65m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:05:22 crc kubenswrapper[4885]: I1002 02:05:22.020494 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bc7146f7-6d12-4bd2-8fd2-0604f358bedf-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "bc7146f7-6d12-4bd2-8fd2-0604f358bedf" (UID: "bc7146f7-6d12-4bd2-8fd2-0604f358bedf"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:05:22 crc kubenswrapper[4885]: I1002 02:05:22.026107 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bc7146f7-6d12-4bd2-8fd2-0604f358bedf-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "bc7146f7-6d12-4bd2-8fd2-0604f358bedf" (UID: "bc7146f7-6d12-4bd2-8fd2-0604f358bedf"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:05:22 crc kubenswrapper[4885]: I1002 02:05:22.038161 4885 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bc7146f7-6d12-4bd2-8fd2-0604f358bedf-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:22 crc kubenswrapper[4885]: I1002 02:05:22.038186 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xf65m\" (UniqueName: \"kubernetes.io/projected/bc7146f7-6d12-4bd2-8fd2-0604f358bedf-kube-api-access-xf65m\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:22 crc kubenswrapper[4885]: I1002 02:05:22.038197 4885 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bc7146f7-6d12-4bd2-8fd2-0604f358bedf-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:22 crc kubenswrapper[4885]: I1002 02:05:22.043421 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bc7146f7-6d12-4bd2-8fd2-0604f358bedf-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "bc7146f7-6d12-4bd2-8fd2-0604f358bedf" (UID: "bc7146f7-6d12-4bd2-8fd2-0604f358bedf"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:05:22 crc kubenswrapper[4885]: I1002 02:05:22.109981 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bc7146f7-6d12-4bd2-8fd2-0604f358bedf-config" (OuterVolumeSpecName: "config") pod "bc7146f7-6d12-4bd2-8fd2-0604f358bedf" (UID: "bc7146f7-6d12-4bd2-8fd2-0604f358bedf"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:05:22 crc kubenswrapper[4885]: I1002 02:05:22.114447 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bc7146f7-6d12-4bd2-8fd2-0604f358bedf-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "bc7146f7-6d12-4bd2-8fd2-0604f358bedf" (UID: "bc7146f7-6d12-4bd2-8fd2-0604f358bedf"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:05:22 crc kubenswrapper[4885]: I1002 02:05:22.140436 4885 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bc7146f7-6d12-4bd2-8fd2-0604f358bedf-config\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:22 crc kubenswrapper[4885]: I1002 02:05:22.140467 4885 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bc7146f7-6d12-4bd2-8fd2-0604f358bedf-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:22 crc kubenswrapper[4885]: I1002 02:05:22.140479 4885 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bc7146f7-6d12-4bd2-8fd2-0604f358bedf-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:22 crc kubenswrapper[4885]: I1002 02:05:22.179192 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9df73c9d-22c1-47c0-a996-0983f206dad5","Type":"ContainerStarted","Data":"d8383b35ce2909f3835cd2327a42f284bd94f3880d123202563b36247af92b23"} Oct 02 02:05:22 crc kubenswrapper[4885]: I1002 02:05:22.200080 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5c6b9856b7-vhzx6" event={"ID":"075adfcf-950d-44e6-ab5e-826f3b06c05e","Type":"ContainerStarted","Data":"881cffd0b491e7624f233b475a0e8208b598cc6ee9fb053be8a7ae46ff4d1483"} Oct 02 02:05:22 crc kubenswrapper[4885]: I1002 02:05:22.203402 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1aff4106-2f4b-48ef-93a1-d51285ce6242","Type":"ContainerDied","Data":"19a5f22bd884950be4e1415ac73ee8a37e2dfbc16378d95a7b8d8e6d45ed33e9"} Oct 02 02:05:22 crc kubenswrapper[4885]: I1002 02:05:22.204169 4885 scope.go:117] "RemoveContainer" containerID="a9cf43c19d008a62e4225357cfd4c6ca237f24f5da97101bd39435507af05ae8" Oct 02 02:05:22 crc kubenswrapper[4885]: I1002 02:05:22.204414 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 02 02:05:22 crc kubenswrapper[4885]: I1002 02:05:22.252855 4885 generic.go:334] "Generic (PLEG): container finished" podID="345ef025-b8fb-4955-baec-9af9be17792e" containerID="c2dd3394b4b25538f87dffc1aa40df5a3104efa6e9908de25784a09ee89b7f7b" exitCode=0 Oct 02 02:05:22 crc kubenswrapper[4885]: I1002 02:05:22.252937 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"345ef025-b8fb-4955-baec-9af9be17792e","Type":"ContainerDied","Data":"c2dd3394b4b25538f87dffc1aa40df5a3104efa6e9908de25784a09ee89b7f7b"} Oct 02 02:05:22 crc kubenswrapper[4885]: I1002 02:05:22.264014 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6b9ddcb5cd-2r5wd" event={"ID":"c416d7da-12d2-4aee-a9c3-5f171a791b43","Type":"ContainerStarted","Data":"9e7a1afea0720bec03571282fe310f1784b155808ad861ebd88b3b5d39c31162"} Oct 02 02:05:22 crc kubenswrapper[4885]: I1002 02:05:22.272436 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 02:05:22 crc kubenswrapper[4885]: I1002 02:05:22.281613 4885 scope.go:117] "RemoveContainer" containerID="71ffbb4c2cbbe9022cc1a62d8f0fc6cd2ac0fed0d96f405f4ec7589f182ddbbd" Oct 02 02:05:22 crc kubenswrapper[4885]: I1002 02:05:22.281816 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c8ddd69c-s4xb6" event={"ID":"bc7146f7-6d12-4bd2-8fd2-0604f358bedf","Type":"ContainerDied","Data":"063ceff3164315c0e2bde7e60fce9d38953e290a687949a326aa73ff7d62f4c2"} Oct 02 02:05:22 crc kubenswrapper[4885]: I1002 02:05:22.281918 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75c8ddd69c-s4xb6" Oct 02 02:05:22 crc kubenswrapper[4885]: I1002 02:05:22.283323 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 02:05:22 crc kubenswrapper[4885]: I1002 02:05:22.299415 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 02:05:22 crc kubenswrapper[4885]: E1002 02:05:22.300055 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc7146f7-6d12-4bd2-8fd2-0604f358bedf" containerName="init" Oct 02 02:05:22 crc kubenswrapper[4885]: I1002 02:05:22.300069 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc7146f7-6d12-4bd2-8fd2-0604f358bedf" containerName="init" Oct 02 02:05:22 crc kubenswrapper[4885]: E1002 02:05:22.300088 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1aff4106-2f4b-48ef-93a1-d51285ce6242" containerName="glance-httpd" Oct 02 02:05:22 crc kubenswrapper[4885]: I1002 02:05:22.300095 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="1aff4106-2f4b-48ef-93a1-d51285ce6242" containerName="glance-httpd" Oct 02 02:05:22 crc kubenswrapper[4885]: E1002 02:05:22.300105 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc7146f7-6d12-4bd2-8fd2-0604f358bedf" containerName="dnsmasq-dns" Oct 02 02:05:22 crc kubenswrapper[4885]: I1002 02:05:22.300111 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc7146f7-6d12-4bd2-8fd2-0604f358bedf" containerName="dnsmasq-dns" Oct 02 02:05:22 crc kubenswrapper[4885]: E1002 02:05:22.300208 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1aff4106-2f4b-48ef-93a1-d51285ce6242" containerName="glance-log" Oct 02 02:05:22 crc kubenswrapper[4885]: I1002 02:05:22.300217 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="1aff4106-2f4b-48ef-93a1-d51285ce6242" containerName="glance-log" Oct 02 02:05:22 crc kubenswrapper[4885]: I1002 02:05:22.300467 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="1aff4106-2f4b-48ef-93a1-d51285ce6242" containerName="glance-httpd" Oct 02 02:05:22 crc kubenswrapper[4885]: I1002 02:05:22.300483 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="1aff4106-2f4b-48ef-93a1-d51285ce6242" containerName="glance-log" Oct 02 02:05:22 crc kubenswrapper[4885]: I1002 02:05:22.300501 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc7146f7-6d12-4bd2-8fd2-0604f358bedf" containerName="dnsmasq-dns" Oct 02 02:05:22 crc kubenswrapper[4885]: I1002 02:05:22.301555 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 02 02:05:22 crc kubenswrapper[4885]: I1002 02:05:22.308055 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 02 02:05:22 crc kubenswrapper[4885]: I1002 02:05:22.308195 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 02 02:05:22 crc kubenswrapper[4885]: I1002 02:05:22.313829 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 02 02:05:22 crc kubenswrapper[4885]: I1002 02:05:22.317978 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 02:05:22 crc kubenswrapper[4885]: I1002 02:05:22.331521 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6b5d8cdf54-gr28p"] Oct 02 02:05:22 crc kubenswrapper[4885]: I1002 02:05:22.340425 4885 scope.go:117] "RemoveContainer" containerID="aae1c6ba38e1514704db03f0c5e9604bed0365df873bcb5f4016a948b1d19310" Oct 02 02:05:22 crc kubenswrapper[4885]: I1002 02:05:22.358157 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 02:05:22 crc kubenswrapper[4885]: I1002 02:05:22.388038 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-s4xb6"] Oct 02 02:05:22 crc kubenswrapper[4885]: I1002 02:05:22.404125 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-s4xb6"] Oct 02 02:05:22 crc kubenswrapper[4885]: I1002 02:05:22.447837 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/345ef025-b8fb-4955-baec-9af9be17792e-public-tls-certs\") pod \"345ef025-b8fb-4955-baec-9af9be17792e\" (UID: \"345ef025-b8fb-4955-baec-9af9be17792e\") " Oct 02 02:05:22 crc kubenswrapper[4885]: I1002 02:05:22.447926 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"345ef025-b8fb-4955-baec-9af9be17792e\" (UID: \"345ef025-b8fb-4955-baec-9af9be17792e\") " Oct 02 02:05:22 crc kubenswrapper[4885]: I1002 02:05:22.447979 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/345ef025-b8fb-4955-baec-9af9be17792e-scripts\") pod \"345ef025-b8fb-4955-baec-9af9be17792e\" (UID: \"345ef025-b8fb-4955-baec-9af9be17792e\") " Oct 02 02:05:22 crc kubenswrapper[4885]: I1002 02:05:22.448006 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/345ef025-b8fb-4955-baec-9af9be17792e-logs\") pod \"345ef025-b8fb-4955-baec-9af9be17792e\" (UID: \"345ef025-b8fb-4955-baec-9af9be17792e\") " Oct 02 02:05:22 crc kubenswrapper[4885]: I1002 02:05:22.448038 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/345ef025-b8fb-4955-baec-9af9be17792e-config-data\") pod \"345ef025-b8fb-4955-baec-9af9be17792e\" (UID: \"345ef025-b8fb-4955-baec-9af9be17792e\") " Oct 02 02:05:22 crc kubenswrapper[4885]: I1002 02:05:22.448157 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/345ef025-b8fb-4955-baec-9af9be17792e-httpd-run\") pod \"345ef025-b8fb-4955-baec-9af9be17792e\" (UID: \"345ef025-b8fb-4955-baec-9af9be17792e\") " Oct 02 02:05:22 crc kubenswrapper[4885]: I1002 02:05:22.448197 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/345ef025-b8fb-4955-baec-9af9be17792e-combined-ca-bundle\") pod \"345ef025-b8fb-4955-baec-9af9be17792e\" (UID: \"345ef025-b8fb-4955-baec-9af9be17792e\") " Oct 02 02:05:22 crc kubenswrapper[4885]: I1002 02:05:22.448317 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xkjt2\" (UniqueName: \"kubernetes.io/projected/345ef025-b8fb-4955-baec-9af9be17792e-kube-api-access-xkjt2\") pod \"345ef025-b8fb-4955-baec-9af9be17792e\" (UID: \"345ef025-b8fb-4955-baec-9af9be17792e\") " Oct 02 02:05:22 crc kubenswrapper[4885]: I1002 02:05:22.448616 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/511f1564-a39d-48be-9c35-ed3343e4650c-logs\") pod \"glance-default-internal-api-0\" (UID: \"511f1564-a39d-48be-9c35-ed3343e4650c\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:05:22 crc kubenswrapper[4885]: I1002 02:05:22.448700 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/511f1564-a39d-48be-9c35-ed3343e4650c-config-data\") pod \"glance-default-internal-api-0\" (UID: \"511f1564-a39d-48be-9c35-ed3343e4650c\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:05:22 crc kubenswrapper[4885]: I1002 02:05:22.448735 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/511f1564-a39d-48be-9c35-ed3343e4650c-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"511f1564-a39d-48be-9c35-ed3343e4650c\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:05:22 crc kubenswrapper[4885]: I1002 02:05:22.448784 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/511f1564-a39d-48be-9c35-ed3343e4650c-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"511f1564-a39d-48be-9c35-ed3343e4650c\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:05:22 crc kubenswrapper[4885]: I1002 02:05:22.448824 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"511f1564-a39d-48be-9c35-ed3343e4650c\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:05:22 crc kubenswrapper[4885]: I1002 02:05:22.448882 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2f65g\" (UniqueName: \"kubernetes.io/projected/511f1564-a39d-48be-9c35-ed3343e4650c-kube-api-access-2f65g\") pod \"glance-default-internal-api-0\" (UID: \"511f1564-a39d-48be-9c35-ed3343e4650c\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:05:22 crc kubenswrapper[4885]: I1002 02:05:22.448924 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/511f1564-a39d-48be-9c35-ed3343e4650c-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"511f1564-a39d-48be-9c35-ed3343e4650c\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:05:22 crc kubenswrapper[4885]: I1002 02:05:22.448947 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/511f1564-a39d-48be-9c35-ed3343e4650c-scripts\") pod \"glance-default-internal-api-0\" (UID: \"511f1564-a39d-48be-9c35-ed3343e4650c\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:05:22 crc kubenswrapper[4885]: I1002 02:05:22.456223 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/345ef025-b8fb-4955-baec-9af9be17792e-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "345ef025-b8fb-4955-baec-9af9be17792e" (UID: "345ef025-b8fb-4955-baec-9af9be17792e"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:05:22 crc kubenswrapper[4885]: I1002 02:05:22.456883 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/345ef025-b8fb-4955-baec-9af9be17792e-logs" (OuterVolumeSpecName: "logs") pod "345ef025-b8fb-4955-baec-9af9be17792e" (UID: "345ef025-b8fb-4955-baec-9af9be17792e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:05:22 crc kubenswrapper[4885]: I1002 02:05:22.475442 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance") pod "345ef025-b8fb-4955-baec-9af9be17792e" (UID: "345ef025-b8fb-4955-baec-9af9be17792e"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 02 02:05:22 crc kubenswrapper[4885]: I1002 02:05:22.490480 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/345ef025-b8fb-4955-baec-9af9be17792e-kube-api-access-xkjt2" (OuterVolumeSpecName: "kube-api-access-xkjt2") pod "345ef025-b8fb-4955-baec-9af9be17792e" (UID: "345ef025-b8fb-4955-baec-9af9be17792e"). InnerVolumeSpecName "kube-api-access-xkjt2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:05:22 crc kubenswrapper[4885]: I1002 02:05:22.496680 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/345ef025-b8fb-4955-baec-9af9be17792e-scripts" (OuterVolumeSpecName: "scripts") pod "345ef025-b8fb-4955-baec-9af9be17792e" (UID: "345ef025-b8fb-4955-baec-9af9be17792e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:05:22 crc kubenswrapper[4885]: I1002 02:05:22.542805 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/345ef025-b8fb-4955-baec-9af9be17792e-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "345ef025-b8fb-4955-baec-9af9be17792e" (UID: "345ef025-b8fb-4955-baec-9af9be17792e"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:05:22 crc kubenswrapper[4885]: I1002 02:05:22.550417 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/511f1564-a39d-48be-9c35-ed3343e4650c-scripts\") pod \"glance-default-internal-api-0\" (UID: \"511f1564-a39d-48be-9c35-ed3343e4650c\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:05:22 crc kubenswrapper[4885]: I1002 02:05:22.550451 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/511f1564-a39d-48be-9c35-ed3343e4650c-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"511f1564-a39d-48be-9c35-ed3343e4650c\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:05:22 crc kubenswrapper[4885]: I1002 02:05:22.550522 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/511f1564-a39d-48be-9c35-ed3343e4650c-logs\") pod \"glance-default-internal-api-0\" (UID: \"511f1564-a39d-48be-9c35-ed3343e4650c\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:05:22 crc kubenswrapper[4885]: I1002 02:05:22.550568 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/511f1564-a39d-48be-9c35-ed3343e4650c-config-data\") pod \"glance-default-internal-api-0\" (UID: \"511f1564-a39d-48be-9c35-ed3343e4650c\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:05:22 crc kubenswrapper[4885]: I1002 02:05:22.550590 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/511f1564-a39d-48be-9c35-ed3343e4650c-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"511f1564-a39d-48be-9c35-ed3343e4650c\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:05:22 crc kubenswrapper[4885]: I1002 02:05:22.550618 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/511f1564-a39d-48be-9c35-ed3343e4650c-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"511f1564-a39d-48be-9c35-ed3343e4650c\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:05:22 crc kubenswrapper[4885]: I1002 02:05:22.550644 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"511f1564-a39d-48be-9c35-ed3343e4650c\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:05:22 crc kubenswrapper[4885]: I1002 02:05:22.550677 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2f65g\" (UniqueName: \"kubernetes.io/projected/511f1564-a39d-48be-9c35-ed3343e4650c-kube-api-access-2f65g\") pod \"glance-default-internal-api-0\" (UID: \"511f1564-a39d-48be-9c35-ed3343e4650c\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:05:22 crc kubenswrapper[4885]: I1002 02:05:22.550727 4885 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/345ef025-b8fb-4955-baec-9af9be17792e-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:22 crc kubenswrapper[4885]: I1002 02:05:22.550737 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xkjt2\" (UniqueName: \"kubernetes.io/projected/345ef025-b8fb-4955-baec-9af9be17792e-kube-api-access-xkjt2\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:22 crc kubenswrapper[4885]: I1002 02:05:22.550746 4885 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/345ef025-b8fb-4955-baec-9af9be17792e-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:22 crc kubenswrapper[4885]: I1002 02:05:22.550764 4885 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Oct 02 02:05:22 crc kubenswrapper[4885]: I1002 02:05:22.550772 4885 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/345ef025-b8fb-4955-baec-9af9be17792e-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:22 crc kubenswrapper[4885]: I1002 02:05:22.550780 4885 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/345ef025-b8fb-4955-baec-9af9be17792e-logs\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:22 crc kubenswrapper[4885]: I1002 02:05:22.556362 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/511f1564-a39d-48be-9c35-ed3343e4650c-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"511f1564-a39d-48be-9c35-ed3343e4650c\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:05:22 crc kubenswrapper[4885]: I1002 02:05:22.556622 4885 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"511f1564-a39d-48be-9c35-ed3343e4650c\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/glance-default-internal-api-0" Oct 02 02:05:22 crc kubenswrapper[4885]: I1002 02:05:22.562285 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/511f1564-a39d-48be-9c35-ed3343e4650c-logs\") pod \"glance-default-internal-api-0\" (UID: \"511f1564-a39d-48be-9c35-ed3343e4650c\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:05:22 crc kubenswrapper[4885]: I1002 02:05:22.577222 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/511f1564-a39d-48be-9c35-ed3343e4650c-config-data\") pod \"glance-default-internal-api-0\" (UID: \"511f1564-a39d-48be-9c35-ed3343e4650c\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:05:22 crc kubenswrapper[4885]: I1002 02:05:22.591725 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/511f1564-a39d-48be-9c35-ed3343e4650c-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"511f1564-a39d-48be-9c35-ed3343e4650c\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:05:22 crc kubenswrapper[4885]: I1002 02:05:22.602412 4885 scope.go:117] "RemoveContainer" containerID="ad63936e2ba60a467001602dab32ef5d5ec0df00c49ea0cf10bb45d14ea406ae" Oct 02 02:05:22 crc kubenswrapper[4885]: I1002 02:05:22.633424 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/511f1564-a39d-48be-9c35-ed3343e4650c-scripts\") pod \"glance-default-internal-api-0\" (UID: \"511f1564-a39d-48be-9c35-ed3343e4650c\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:05:22 crc kubenswrapper[4885]: I1002 02:05:22.642944 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/511f1564-a39d-48be-9c35-ed3343e4650c-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"511f1564-a39d-48be-9c35-ed3343e4650c\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:05:22 crc kubenswrapper[4885]: I1002 02:05:22.642957 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2f65g\" (UniqueName: \"kubernetes.io/projected/511f1564-a39d-48be-9c35-ed3343e4650c-kube-api-access-2f65g\") pod \"glance-default-internal-api-0\" (UID: \"511f1564-a39d-48be-9c35-ed3343e4650c\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:05:22 crc kubenswrapper[4885]: I1002 02:05:22.651539 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/345ef025-b8fb-4955-baec-9af9be17792e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "345ef025-b8fb-4955-baec-9af9be17792e" (UID: "345ef025-b8fb-4955-baec-9af9be17792e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:05:22 crc kubenswrapper[4885]: I1002 02:05:22.653201 4885 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/345ef025-b8fb-4955-baec-9af9be17792e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:22 crc kubenswrapper[4885]: I1002 02:05:22.694240 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 02 02:05:22 crc kubenswrapper[4885]: I1002 02:05:22.700809 4885 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Oct 02 02:05:22 crc kubenswrapper[4885]: I1002 02:05:22.747303 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"511f1564-a39d-48be-9c35-ed3343e4650c\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:05:22 crc kubenswrapper[4885]: I1002 02:05:22.748427 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-85ntb"] Oct 02 02:05:22 crc kubenswrapper[4885]: I1002 02:05:22.755629 4885 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:22 crc kubenswrapper[4885]: I1002 02:05:22.774848 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/345ef025-b8fb-4955-baec-9af9be17792e-config-data" (OuterVolumeSpecName: "config-data") pod "345ef025-b8fb-4955-baec-9af9be17792e" (UID: "345ef025-b8fb-4955-baec-9af9be17792e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:05:22 crc kubenswrapper[4885]: I1002 02:05:22.859661 4885 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/345ef025-b8fb-4955-baec-9af9be17792e-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:22 crc kubenswrapper[4885]: I1002 02:05:22.945684 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 02 02:05:23 crc kubenswrapper[4885]: I1002 02:05:23.329514 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"f985ab4a-2149-40c7-9af9-5ac242974f28","Type":"ContainerStarted","Data":"8cd87ad3eebf28cb59f2be5ee00434a3d27fa1575d07f887add30bdbd99425d1"} Oct 02 02:05:23 crc kubenswrapper[4885]: I1002 02:05:23.369403 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6b9ddcb5cd-2r5wd" event={"ID":"c416d7da-12d2-4aee-a9c3-5f171a791b43","Type":"ContainerStarted","Data":"8d8a5c279aba5d4198c4d01a1977228ae63cc30f7fe37d467ee1bd4433c2c485"} Oct 02 02:05:23 crc kubenswrapper[4885]: I1002 02:05:23.386521 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6b5d8cdf54-gr28p" event={"ID":"af70565f-bc37-43bc-97a3-3942c35a9fb2","Type":"ContainerStarted","Data":"36825090bf4e5f1f77e6f7f8542b2ee9812c3e1ed8b6d880efeba7ed54d222bb"} Oct 02 02:05:23 crc kubenswrapper[4885]: I1002 02:05:23.386567 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6b5d8cdf54-gr28p" event={"ID":"af70565f-bc37-43bc-97a3-3942c35a9fb2","Type":"ContainerStarted","Data":"d956b51307c61ba8b71e5e3e55d4b0f2cfb4cad6cf3659ca47a94d0293ad703f"} Oct 02 02:05:23 crc kubenswrapper[4885]: I1002 02:05:23.387455 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6b5d8cdf54-gr28p" Oct 02 02:05:23 crc kubenswrapper[4885]: I1002 02:05:23.387479 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6b5d8cdf54-gr28p" Oct 02 02:05:23 crc kubenswrapper[4885]: I1002 02:05:23.406224 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5c6b9856b7-vhzx6" event={"ID":"075adfcf-950d-44e6-ab5e-826f3b06c05e","Type":"ContainerStarted","Data":"be53a3accbda241476229264d77d995ff9c6c34ee69dd1ddec4cfea523436da7"} Oct 02 02:05:23 crc kubenswrapper[4885]: I1002 02:05:23.424469 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-6b5d8cdf54-gr28p" podStartSLOduration=9.424454947 podStartE2EDuration="9.424454947s" podCreationTimestamp="2025-10-02 02:05:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:05:23.424044964 +0000 UTC m=+1112.235792373" watchObservedRunningTime="2025-10-02 02:05:23.424454947 +0000 UTC m=+1112.236202346" Oct 02 02:05:23 crc kubenswrapper[4885]: I1002 02:05:23.441608 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 02 02:05:23 crc kubenswrapper[4885]: I1002 02:05:23.445207 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"345ef025-b8fb-4955-baec-9af9be17792e","Type":"ContainerDied","Data":"99a5bdbb40f5c2e4ebc7ec16830e9ee8b017533039be1ad5e43e4d261465daf7"} Oct 02 02:05:23 crc kubenswrapper[4885]: I1002 02:05:23.445254 4885 scope.go:117] "RemoveContainer" containerID="c2dd3394b4b25538f87dffc1aa40df5a3104efa6e9908de25784a09ee89b7f7b" Oct 02 02:05:23 crc kubenswrapper[4885]: I1002 02:05:23.455931 4885 generic.go:334] "Generic (PLEG): container finished" podID="7a9422e4-cbc9-4eb7-885d-eb42b1017cab" containerID="fdc94990ef2b2029d14e73d89279ee34dec65d2cb37895e045e79f39df25994e" exitCode=0 Oct 02 02:05:23 crc kubenswrapper[4885]: I1002 02:05:23.456026 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-85ntb" event={"ID":"7a9422e4-cbc9-4eb7-885d-eb42b1017cab","Type":"ContainerDied","Data":"fdc94990ef2b2029d14e73d89279ee34dec65d2cb37895e045e79f39df25994e"} Oct 02 02:05:23 crc kubenswrapper[4885]: I1002 02:05:23.456051 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-85ntb" event={"ID":"7a9422e4-cbc9-4eb7-885d-eb42b1017cab","Type":"ContainerStarted","Data":"37eece54cd8a4c574dcc228201218ea61821ce6ce6068bfa666592c9a6828440"} Oct 02 02:05:23 crc kubenswrapper[4885]: I1002 02:05:23.464953 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-6b9ddcb5cd-2r5wd" podStartSLOduration=4.132168622 podStartE2EDuration="12.464934937s" podCreationTimestamp="2025-10-02 02:05:11 +0000 UTC" firstStartedPulling="2025-10-02 02:05:13.0956324 +0000 UTC m=+1101.907379799" lastFinishedPulling="2025-10-02 02:05:21.428398695 +0000 UTC m=+1110.240146114" observedRunningTime="2025-10-02 02:05:23.395788726 +0000 UTC m=+1112.207536125" watchObservedRunningTime="2025-10-02 02:05:23.464934937 +0000 UTC m=+1112.276682326" Oct 02 02:05:23 crc kubenswrapper[4885]: I1002 02:05:23.467136 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"72053d12-6b1a-4df1-9fbe-ba2973d53dc3","Type":"ContainerStarted","Data":"978695589c6389422d521cee5fa7ae0b367fda9b2ccfb5c01a5c38aeebd5308c"} Oct 02 02:05:23 crc kubenswrapper[4885]: I1002 02:05:23.504520 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-5c6b9856b7-vhzx6" podStartSLOduration=4.295087976 podStartE2EDuration="12.50449655s" podCreationTimestamp="2025-10-02 02:05:11 +0000 UTC" firstStartedPulling="2025-10-02 02:05:13.178953714 +0000 UTC m=+1101.990701103" lastFinishedPulling="2025-10-02 02:05:21.388362268 +0000 UTC m=+1110.200109677" observedRunningTime="2025-10-02 02:05:23.441212277 +0000 UTC m=+1112.252959676" watchObservedRunningTime="2025-10-02 02:05:23.50449655 +0000 UTC m=+1112.316243949" Oct 02 02:05:23 crc kubenswrapper[4885]: I1002 02:05:23.608434 4885 scope.go:117] "RemoveContainer" containerID="0a4f01b0e93fc3d8b30b56a1c50f9a5b6be0c884595ab22522466917a5ed9451" Oct 02 02:05:23 crc kubenswrapper[4885]: I1002 02:05:23.613130 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 02:05:23 crc kubenswrapper[4885]: I1002 02:05:23.675184 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 02:05:23 crc kubenswrapper[4885]: I1002 02:05:23.709949 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 02:05:23 crc kubenswrapper[4885]: E1002 02:05:23.710369 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="345ef025-b8fb-4955-baec-9af9be17792e" containerName="glance-log" Oct 02 02:05:23 crc kubenswrapper[4885]: I1002 02:05:23.710384 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="345ef025-b8fb-4955-baec-9af9be17792e" containerName="glance-log" Oct 02 02:05:23 crc kubenswrapper[4885]: E1002 02:05:23.710413 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="345ef025-b8fb-4955-baec-9af9be17792e" containerName="glance-httpd" Oct 02 02:05:23 crc kubenswrapper[4885]: I1002 02:05:23.710419 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="345ef025-b8fb-4955-baec-9af9be17792e" containerName="glance-httpd" Oct 02 02:05:23 crc kubenswrapper[4885]: I1002 02:05:23.710596 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="345ef025-b8fb-4955-baec-9af9be17792e" containerName="glance-httpd" Oct 02 02:05:23 crc kubenswrapper[4885]: I1002 02:05:23.710622 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="345ef025-b8fb-4955-baec-9af9be17792e" containerName="glance-log" Oct 02 02:05:23 crc kubenswrapper[4885]: I1002 02:05:23.711529 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 02 02:05:23 crc kubenswrapper[4885]: I1002 02:05:23.713502 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 02 02:05:23 crc kubenswrapper[4885]: I1002 02:05:23.719867 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 02 02:05:23 crc kubenswrapper[4885]: I1002 02:05:23.734297 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 02:05:23 crc kubenswrapper[4885]: I1002 02:05:23.751791 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 02:05:23 crc kubenswrapper[4885]: I1002 02:05:23.790225 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6e0fd15-e949-444a-86e2-894963c1267d-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"d6e0fd15-e949-444a-86e2-894963c1267d\") " pod="openstack/glance-default-external-api-0" Oct 02 02:05:23 crc kubenswrapper[4885]: I1002 02:05:23.794762 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6e0fd15-e949-444a-86e2-894963c1267d-config-data\") pod \"glance-default-external-api-0\" (UID: \"d6e0fd15-e949-444a-86e2-894963c1267d\") " pod="openstack/glance-default-external-api-0" Oct 02 02:05:23 crc kubenswrapper[4885]: I1002 02:05:23.794804 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6e0fd15-e949-444a-86e2-894963c1267d-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"d6e0fd15-e949-444a-86e2-894963c1267d\") " pod="openstack/glance-default-external-api-0" Oct 02 02:05:23 crc kubenswrapper[4885]: I1002 02:05:23.794874 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"d6e0fd15-e949-444a-86e2-894963c1267d\") " pod="openstack/glance-default-external-api-0" Oct 02 02:05:23 crc kubenswrapper[4885]: I1002 02:05:23.795002 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d6e0fd15-e949-444a-86e2-894963c1267d-scripts\") pod \"glance-default-external-api-0\" (UID: \"d6e0fd15-e949-444a-86e2-894963c1267d\") " pod="openstack/glance-default-external-api-0" Oct 02 02:05:23 crc kubenswrapper[4885]: I1002 02:05:23.795019 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d6e0fd15-e949-444a-86e2-894963c1267d-logs\") pod \"glance-default-external-api-0\" (UID: \"d6e0fd15-e949-444a-86e2-894963c1267d\") " pod="openstack/glance-default-external-api-0" Oct 02 02:05:23 crc kubenswrapper[4885]: I1002 02:05:23.795064 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f9wm6\" (UniqueName: \"kubernetes.io/projected/d6e0fd15-e949-444a-86e2-894963c1267d-kube-api-access-f9wm6\") pod \"glance-default-external-api-0\" (UID: \"d6e0fd15-e949-444a-86e2-894963c1267d\") " pod="openstack/glance-default-external-api-0" Oct 02 02:05:23 crc kubenswrapper[4885]: I1002 02:05:23.795103 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d6e0fd15-e949-444a-86e2-894963c1267d-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d6e0fd15-e949-444a-86e2-894963c1267d\") " pod="openstack/glance-default-external-api-0" Oct 02 02:05:23 crc kubenswrapper[4885]: I1002 02:05:23.896238 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6e0fd15-e949-444a-86e2-894963c1267d-config-data\") pod \"glance-default-external-api-0\" (UID: \"d6e0fd15-e949-444a-86e2-894963c1267d\") " pod="openstack/glance-default-external-api-0" Oct 02 02:05:23 crc kubenswrapper[4885]: I1002 02:05:23.896291 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6e0fd15-e949-444a-86e2-894963c1267d-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"d6e0fd15-e949-444a-86e2-894963c1267d\") " pod="openstack/glance-default-external-api-0" Oct 02 02:05:23 crc kubenswrapper[4885]: I1002 02:05:23.896331 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"d6e0fd15-e949-444a-86e2-894963c1267d\") " pod="openstack/glance-default-external-api-0" Oct 02 02:05:23 crc kubenswrapper[4885]: I1002 02:05:23.896398 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d6e0fd15-e949-444a-86e2-894963c1267d-scripts\") pod \"glance-default-external-api-0\" (UID: \"d6e0fd15-e949-444a-86e2-894963c1267d\") " pod="openstack/glance-default-external-api-0" Oct 02 02:05:23 crc kubenswrapper[4885]: I1002 02:05:23.896417 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d6e0fd15-e949-444a-86e2-894963c1267d-logs\") pod \"glance-default-external-api-0\" (UID: \"d6e0fd15-e949-444a-86e2-894963c1267d\") " pod="openstack/glance-default-external-api-0" Oct 02 02:05:23 crc kubenswrapper[4885]: I1002 02:05:23.896445 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f9wm6\" (UniqueName: \"kubernetes.io/projected/d6e0fd15-e949-444a-86e2-894963c1267d-kube-api-access-f9wm6\") pod \"glance-default-external-api-0\" (UID: \"d6e0fd15-e949-444a-86e2-894963c1267d\") " pod="openstack/glance-default-external-api-0" Oct 02 02:05:23 crc kubenswrapper[4885]: I1002 02:05:23.896473 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d6e0fd15-e949-444a-86e2-894963c1267d-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d6e0fd15-e949-444a-86e2-894963c1267d\") " pod="openstack/glance-default-external-api-0" Oct 02 02:05:23 crc kubenswrapper[4885]: I1002 02:05:23.896519 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6e0fd15-e949-444a-86e2-894963c1267d-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"d6e0fd15-e949-444a-86e2-894963c1267d\") " pod="openstack/glance-default-external-api-0" Oct 02 02:05:23 crc kubenswrapper[4885]: I1002 02:05:23.900472 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d6e0fd15-e949-444a-86e2-894963c1267d-scripts\") pod \"glance-default-external-api-0\" (UID: \"d6e0fd15-e949-444a-86e2-894963c1267d\") " pod="openstack/glance-default-external-api-0" Oct 02 02:05:23 crc kubenswrapper[4885]: I1002 02:05:23.901573 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6e0fd15-e949-444a-86e2-894963c1267d-config-data\") pod \"glance-default-external-api-0\" (UID: \"d6e0fd15-e949-444a-86e2-894963c1267d\") " pod="openstack/glance-default-external-api-0" Oct 02 02:05:23 crc kubenswrapper[4885]: I1002 02:05:23.901705 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d6e0fd15-e949-444a-86e2-894963c1267d-logs\") pod \"glance-default-external-api-0\" (UID: \"d6e0fd15-e949-444a-86e2-894963c1267d\") " pod="openstack/glance-default-external-api-0" Oct 02 02:05:23 crc kubenswrapper[4885]: I1002 02:05:23.902110 4885 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"d6e0fd15-e949-444a-86e2-894963c1267d\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/glance-default-external-api-0" Oct 02 02:05:23 crc kubenswrapper[4885]: I1002 02:05:23.902346 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d6e0fd15-e949-444a-86e2-894963c1267d-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d6e0fd15-e949-444a-86e2-894963c1267d\") " pod="openstack/glance-default-external-api-0" Oct 02 02:05:23 crc kubenswrapper[4885]: I1002 02:05:23.910744 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6e0fd15-e949-444a-86e2-894963c1267d-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"d6e0fd15-e949-444a-86e2-894963c1267d\") " pod="openstack/glance-default-external-api-0" Oct 02 02:05:23 crc kubenswrapper[4885]: I1002 02:05:23.911065 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6e0fd15-e949-444a-86e2-894963c1267d-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"d6e0fd15-e949-444a-86e2-894963c1267d\") " pod="openstack/glance-default-external-api-0" Oct 02 02:05:23 crc kubenswrapper[4885]: I1002 02:05:23.938619 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f9wm6\" (UniqueName: \"kubernetes.io/projected/d6e0fd15-e949-444a-86e2-894963c1267d-kube-api-access-f9wm6\") pod \"glance-default-external-api-0\" (UID: \"d6e0fd15-e949-444a-86e2-894963c1267d\") " pod="openstack/glance-default-external-api-0" Oct 02 02:05:23 crc kubenswrapper[4885]: I1002 02:05:23.996590 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"d6e0fd15-e949-444a-86e2-894963c1267d\") " pod="openstack/glance-default-external-api-0" Oct 02 02:05:24 crc kubenswrapper[4885]: I1002 02:05:24.063599 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 02 02:05:24 crc kubenswrapper[4885]: I1002 02:05:24.065673 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1aff4106-2f4b-48ef-93a1-d51285ce6242" path="/var/lib/kubelet/pods/1aff4106-2f4b-48ef-93a1-d51285ce6242/volumes" Oct 02 02:05:24 crc kubenswrapper[4885]: I1002 02:05:24.067099 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="345ef025-b8fb-4955-baec-9af9be17792e" path="/var/lib/kubelet/pods/345ef025-b8fb-4955-baec-9af9be17792e/volumes" Oct 02 02:05:24 crc kubenswrapper[4885]: I1002 02:05:24.067850 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc7146f7-6d12-4bd2-8fd2-0604f358bedf" path="/var/lib/kubelet/pods/bc7146f7-6d12-4bd2-8fd2-0604f358bedf/volumes" Oct 02 02:05:24 crc kubenswrapper[4885]: I1002 02:05:24.522011 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-85ntb" event={"ID":"7a9422e4-cbc9-4eb7-885d-eb42b1017cab","Type":"ContainerStarted","Data":"347578f3963a42175c1618b8d901e7db107730b159cc81edce216f280b6faa66"} Oct 02 02:05:24 crc kubenswrapper[4885]: I1002 02:05:24.522287 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5784cf869f-85ntb" Oct 02 02:05:24 crc kubenswrapper[4885]: I1002 02:05:24.541698 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6b5d8cdf54-gr28p" event={"ID":"af70565f-bc37-43bc-97a3-3942c35a9fb2","Type":"ContainerStarted","Data":"b96a3ab5bc2ed8125ca3b8ac2979f801688d6727e8622cc8d5f3f9d8c8fca7aa"} Oct 02 02:05:24 crc kubenswrapper[4885]: I1002 02:05:24.552019 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"511f1564-a39d-48be-9c35-ed3343e4650c","Type":"ContainerStarted","Data":"58f306617c5b625e09ad84669a73a1c70469b63ee670656eedef1a9576e0679a"} Oct 02 02:05:24 crc kubenswrapper[4885]: I1002 02:05:24.591022 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9df73c9d-22c1-47c0-a996-0983f206dad5","Type":"ContainerStarted","Data":"f2f63cc40789d147d4c97ba73e7b6f04c28d928c576d2b12411df129a24ece8f"} Oct 02 02:05:24 crc kubenswrapper[4885]: I1002 02:05:24.591293 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9df73c9d-22c1-47c0-a996-0983f206dad5" containerName="ceilometer-central-agent" containerID="cri-o://9706f642020809b48c4b255916f1caf766eb34203bb17a12bfc4d4acb363e7d3" gracePeriod=30 Oct 02 02:05:24 crc kubenswrapper[4885]: I1002 02:05:24.591338 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 02 02:05:24 crc kubenswrapper[4885]: I1002 02:05:24.591361 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9df73c9d-22c1-47c0-a996-0983f206dad5" containerName="ceilometer-notification-agent" containerID="cri-o://597d57d9acdd3cbb0ea1e6a0716e324596b7fdad2025c8334be56d83e0eb7b4a" gracePeriod=30 Oct 02 02:05:24 crc kubenswrapper[4885]: I1002 02:05:24.591376 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9df73c9d-22c1-47c0-a996-0983f206dad5" containerName="sg-core" containerID="cri-o://d8383b35ce2909f3835cd2327a42f284bd94f3880d123202563b36247af92b23" gracePeriod=30 Oct 02 02:05:24 crc kubenswrapper[4885]: I1002 02:05:24.591409 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9df73c9d-22c1-47c0-a996-0983f206dad5" containerName="proxy-httpd" containerID="cri-o://f2f63cc40789d147d4c97ba73e7b6f04c28d928c576d2b12411df129a24ece8f" gracePeriod=30 Oct 02 02:05:24 crc kubenswrapper[4885]: I1002 02:05:24.608797 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"f985ab4a-2149-40c7-9af9-5ac242974f28","Type":"ContainerStarted","Data":"2a5ee4d55456ceb8c5773da638d05738c8b83706c49ead5b8fabb99d0934b508"} Oct 02 02:05:24 crc kubenswrapper[4885]: I1002 02:05:24.649643 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5784cf869f-85ntb" podStartSLOduration=5.649625255 podStartE2EDuration="5.649625255s" podCreationTimestamp="2025-10-02 02:05:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:05:24.563687192 +0000 UTC m=+1113.375434591" watchObservedRunningTime="2025-10-02 02:05:24.649625255 +0000 UTC m=+1113.461372654" Oct 02 02:05:24 crc kubenswrapper[4885]: I1002 02:05:24.762674 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.196899999 podStartE2EDuration="13.762658391s" podCreationTimestamp="2025-10-02 02:05:11 +0000 UTC" firstStartedPulling="2025-10-02 02:05:12.139594795 +0000 UTC m=+1100.951342194" lastFinishedPulling="2025-10-02 02:05:23.705353197 +0000 UTC m=+1112.517100586" observedRunningTime="2025-10-02 02:05:24.649453009 +0000 UTC m=+1113.461200408" watchObservedRunningTime="2025-10-02 02:05:24.762658391 +0000 UTC m=+1113.574405790" Oct 02 02:05:24 crc kubenswrapper[4885]: I1002 02:05:24.768434 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 02:05:24 crc kubenswrapper[4885]: I1002 02:05:24.771532 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6c87b5899d-lvdbp" Oct 02 02:05:25 crc kubenswrapper[4885]: I1002 02:05:25.153811 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6c87b5899d-lvdbp" Oct 02 02:05:25 crc kubenswrapper[4885]: I1002 02:05:25.641909 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d6e0fd15-e949-444a-86e2-894963c1267d","Type":"ContainerStarted","Data":"a80cd6b954f8999b312c85b6417e2d970b05a22d52171241f6dce15bc6ec1ef2"} Oct 02 02:05:25 crc kubenswrapper[4885]: I1002 02:05:25.645074 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"511f1564-a39d-48be-9c35-ed3343e4650c","Type":"ContainerStarted","Data":"fb7cf60d4c4741727f52f4864b6a687f80942cb6be3592b2075d5710aa786267"} Oct 02 02:05:25 crc kubenswrapper[4885]: I1002 02:05:25.650418 4885 generic.go:334] "Generic (PLEG): container finished" podID="9df73c9d-22c1-47c0-a996-0983f206dad5" containerID="f2f63cc40789d147d4c97ba73e7b6f04c28d928c576d2b12411df129a24ece8f" exitCode=0 Oct 02 02:05:25 crc kubenswrapper[4885]: I1002 02:05:25.650447 4885 generic.go:334] "Generic (PLEG): container finished" podID="9df73c9d-22c1-47c0-a996-0983f206dad5" containerID="d8383b35ce2909f3835cd2327a42f284bd94f3880d123202563b36247af92b23" exitCode=2 Oct 02 02:05:25 crc kubenswrapper[4885]: I1002 02:05:25.650456 4885 generic.go:334] "Generic (PLEG): container finished" podID="9df73c9d-22c1-47c0-a996-0983f206dad5" containerID="9706f642020809b48c4b255916f1caf766eb34203bb17a12bfc4d4acb363e7d3" exitCode=0 Oct 02 02:05:25 crc kubenswrapper[4885]: I1002 02:05:25.650490 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9df73c9d-22c1-47c0-a996-0983f206dad5","Type":"ContainerDied","Data":"f2f63cc40789d147d4c97ba73e7b6f04c28d928c576d2b12411df129a24ece8f"} Oct 02 02:05:25 crc kubenswrapper[4885]: I1002 02:05:25.650511 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9df73c9d-22c1-47c0-a996-0983f206dad5","Type":"ContainerDied","Data":"d8383b35ce2909f3835cd2327a42f284bd94f3880d123202563b36247af92b23"} Oct 02 02:05:25 crc kubenswrapper[4885]: I1002 02:05:25.650520 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9df73c9d-22c1-47c0-a996-0983f206dad5","Type":"ContainerDied","Data":"9706f642020809b48c4b255916f1caf766eb34203bb17a12bfc4d4acb363e7d3"} Oct 02 02:05:25 crc kubenswrapper[4885]: I1002 02:05:25.652085 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"72053d12-6b1a-4df1-9fbe-ba2973d53dc3","Type":"ContainerStarted","Data":"1bab2764732e628230707bff05f072c449d8f783b77372dfedcf2d5c5acb4e10"} Oct 02 02:05:25 crc kubenswrapper[4885]: I1002 02:05:25.654742 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"f985ab4a-2149-40c7-9af9-5ac242974f28","Type":"ContainerStarted","Data":"3dec1875c31681ffdcc4fa041fc6c6f1e2bf30faea41cc03cd5081ca94280caf"} Oct 02 02:05:25 crc kubenswrapper[4885]: I1002 02:05:25.654862 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="f985ab4a-2149-40c7-9af9-5ac242974f28" containerName="cinder-api-log" containerID="cri-o://2a5ee4d55456ceb8c5773da638d05738c8b83706c49ead5b8fabb99d0934b508" gracePeriod=30 Oct 02 02:05:25 crc kubenswrapper[4885]: I1002 02:05:25.655139 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="f985ab4a-2149-40c7-9af9-5ac242974f28" containerName="cinder-api" containerID="cri-o://3dec1875c31681ffdcc4fa041fc6c6f1e2bf30faea41cc03cd5081ca94280caf" gracePeriod=30 Oct 02 02:05:25 crc kubenswrapper[4885]: I1002 02:05:25.657132 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 02 02:05:25 crc kubenswrapper[4885]: I1002 02:05:25.685497 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=6.685480535 podStartE2EDuration="6.685480535s" podCreationTimestamp="2025-10-02 02:05:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:05:25.681884176 +0000 UTC m=+1114.493631575" watchObservedRunningTime="2025-10-02 02:05:25.685480535 +0000 UTC m=+1114.497227934" Oct 02 02:05:26 crc kubenswrapper[4885]: I1002 02:05:26.284402 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 02 02:05:26 crc kubenswrapper[4885]: I1002 02:05:26.381121 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f985ab4a-2149-40c7-9af9-5ac242974f28-etc-machine-id\") pod \"f985ab4a-2149-40c7-9af9-5ac242974f28\" (UID: \"f985ab4a-2149-40c7-9af9-5ac242974f28\") " Oct 02 02:05:26 crc kubenswrapper[4885]: I1002 02:05:26.381184 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hc8gw\" (UniqueName: \"kubernetes.io/projected/f985ab4a-2149-40c7-9af9-5ac242974f28-kube-api-access-hc8gw\") pod \"f985ab4a-2149-40c7-9af9-5ac242974f28\" (UID: \"f985ab4a-2149-40c7-9af9-5ac242974f28\") " Oct 02 02:05:26 crc kubenswrapper[4885]: I1002 02:05:26.381234 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f985ab4a-2149-40c7-9af9-5ac242974f28-config-data\") pod \"f985ab4a-2149-40c7-9af9-5ac242974f28\" (UID: \"f985ab4a-2149-40c7-9af9-5ac242974f28\") " Oct 02 02:05:26 crc kubenswrapper[4885]: I1002 02:05:26.381329 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f985ab4a-2149-40c7-9af9-5ac242974f28-scripts\") pod \"f985ab4a-2149-40c7-9af9-5ac242974f28\" (UID: \"f985ab4a-2149-40c7-9af9-5ac242974f28\") " Oct 02 02:05:26 crc kubenswrapper[4885]: I1002 02:05:26.381351 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f985ab4a-2149-40c7-9af9-5ac242974f28-config-data-custom\") pod \"f985ab4a-2149-40c7-9af9-5ac242974f28\" (UID: \"f985ab4a-2149-40c7-9af9-5ac242974f28\") " Oct 02 02:05:26 crc kubenswrapper[4885]: I1002 02:05:26.381372 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f985ab4a-2149-40c7-9af9-5ac242974f28-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "f985ab4a-2149-40c7-9af9-5ac242974f28" (UID: "f985ab4a-2149-40c7-9af9-5ac242974f28"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 02:05:26 crc kubenswrapper[4885]: I1002 02:05:26.381462 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f985ab4a-2149-40c7-9af9-5ac242974f28-logs\") pod \"f985ab4a-2149-40c7-9af9-5ac242974f28\" (UID: \"f985ab4a-2149-40c7-9af9-5ac242974f28\") " Oct 02 02:05:26 crc kubenswrapper[4885]: I1002 02:05:26.381534 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f985ab4a-2149-40c7-9af9-5ac242974f28-combined-ca-bundle\") pod \"f985ab4a-2149-40c7-9af9-5ac242974f28\" (UID: \"f985ab4a-2149-40c7-9af9-5ac242974f28\") " Oct 02 02:05:26 crc kubenswrapper[4885]: I1002 02:05:26.381920 4885 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f985ab4a-2149-40c7-9af9-5ac242974f28-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:26 crc kubenswrapper[4885]: I1002 02:05:26.386045 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f985ab4a-2149-40c7-9af9-5ac242974f28-logs" (OuterVolumeSpecName: "logs") pod "f985ab4a-2149-40c7-9af9-5ac242974f28" (UID: "f985ab4a-2149-40c7-9af9-5ac242974f28"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:05:26 crc kubenswrapper[4885]: I1002 02:05:26.397599 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f985ab4a-2149-40c7-9af9-5ac242974f28-scripts" (OuterVolumeSpecName: "scripts") pod "f985ab4a-2149-40c7-9af9-5ac242974f28" (UID: "f985ab4a-2149-40c7-9af9-5ac242974f28"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:05:26 crc kubenswrapper[4885]: I1002 02:05:26.398048 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f985ab4a-2149-40c7-9af9-5ac242974f28-kube-api-access-hc8gw" (OuterVolumeSpecName: "kube-api-access-hc8gw") pod "f985ab4a-2149-40c7-9af9-5ac242974f28" (UID: "f985ab4a-2149-40c7-9af9-5ac242974f28"). InnerVolumeSpecName "kube-api-access-hc8gw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:05:26 crc kubenswrapper[4885]: I1002 02:05:26.419130 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f985ab4a-2149-40c7-9af9-5ac242974f28-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "f985ab4a-2149-40c7-9af9-5ac242974f28" (UID: "f985ab4a-2149-40c7-9af9-5ac242974f28"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:05:26 crc kubenswrapper[4885]: I1002 02:05:26.443463 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f985ab4a-2149-40c7-9af9-5ac242974f28-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f985ab4a-2149-40c7-9af9-5ac242974f28" (UID: "f985ab4a-2149-40c7-9af9-5ac242974f28"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:05:26 crc kubenswrapper[4885]: I1002 02:05:26.484362 4885 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f985ab4a-2149-40c7-9af9-5ac242974f28-logs\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:26 crc kubenswrapper[4885]: I1002 02:05:26.484597 4885 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f985ab4a-2149-40c7-9af9-5ac242974f28-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:26 crc kubenswrapper[4885]: I1002 02:05:26.484609 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hc8gw\" (UniqueName: \"kubernetes.io/projected/f985ab4a-2149-40c7-9af9-5ac242974f28-kube-api-access-hc8gw\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:26 crc kubenswrapper[4885]: I1002 02:05:26.484618 4885 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f985ab4a-2149-40c7-9af9-5ac242974f28-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:26 crc kubenswrapper[4885]: I1002 02:05:26.484626 4885 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f985ab4a-2149-40c7-9af9-5ac242974f28-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:26 crc kubenswrapper[4885]: I1002 02:05:26.486760 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f985ab4a-2149-40c7-9af9-5ac242974f28-config-data" (OuterVolumeSpecName: "config-data") pod "f985ab4a-2149-40c7-9af9-5ac242974f28" (UID: "f985ab4a-2149-40c7-9af9-5ac242974f28"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:05:26 crc kubenswrapper[4885]: I1002 02:05:26.586127 4885 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f985ab4a-2149-40c7-9af9-5ac242974f28-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:26 crc kubenswrapper[4885]: I1002 02:05:26.666097 4885 generic.go:334] "Generic (PLEG): container finished" podID="f985ab4a-2149-40c7-9af9-5ac242974f28" containerID="3dec1875c31681ffdcc4fa041fc6c6f1e2bf30faea41cc03cd5081ca94280caf" exitCode=0 Oct 02 02:05:26 crc kubenswrapper[4885]: I1002 02:05:26.666127 4885 generic.go:334] "Generic (PLEG): container finished" podID="f985ab4a-2149-40c7-9af9-5ac242974f28" containerID="2a5ee4d55456ceb8c5773da638d05738c8b83706c49ead5b8fabb99d0934b508" exitCode=143 Oct 02 02:05:26 crc kubenswrapper[4885]: I1002 02:05:26.666214 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 02 02:05:26 crc kubenswrapper[4885]: I1002 02:05:26.666255 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"f985ab4a-2149-40c7-9af9-5ac242974f28","Type":"ContainerDied","Data":"3dec1875c31681ffdcc4fa041fc6c6f1e2bf30faea41cc03cd5081ca94280caf"} Oct 02 02:05:26 crc kubenswrapper[4885]: I1002 02:05:26.666295 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"f985ab4a-2149-40c7-9af9-5ac242974f28","Type":"ContainerDied","Data":"2a5ee4d55456ceb8c5773da638d05738c8b83706c49ead5b8fabb99d0934b508"} Oct 02 02:05:26 crc kubenswrapper[4885]: I1002 02:05:26.666309 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"f985ab4a-2149-40c7-9af9-5ac242974f28","Type":"ContainerDied","Data":"8cd87ad3eebf28cb59f2be5ee00434a3d27fa1575d07f887add30bdbd99425d1"} Oct 02 02:05:26 crc kubenswrapper[4885]: I1002 02:05:26.666327 4885 scope.go:117] "RemoveContainer" containerID="3dec1875c31681ffdcc4fa041fc6c6f1e2bf30faea41cc03cd5081ca94280caf" Oct 02 02:05:26 crc kubenswrapper[4885]: I1002 02:05:26.670106 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d6e0fd15-e949-444a-86e2-894963c1267d","Type":"ContainerStarted","Data":"3a64f7db6cf158663b66ae424e25b0d451d9b110f226dc5d80b3cf158a8911e3"} Oct 02 02:05:26 crc kubenswrapper[4885]: I1002 02:05:26.670138 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d6e0fd15-e949-444a-86e2-894963c1267d","Type":"ContainerStarted","Data":"2ce59d2e6421a3b44f1fe2ca9dd46df9bd45c89fdc8b07e29f75f4cbd53bf58d"} Oct 02 02:05:26 crc kubenswrapper[4885]: I1002 02:05:26.678674 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"511f1564-a39d-48be-9c35-ed3343e4650c","Type":"ContainerStarted","Data":"db6b531a75ff55e9ef5f688cf6863825f24dd4ac3e12328717eb6a21136e59ea"} Oct 02 02:05:26 crc kubenswrapper[4885]: I1002 02:05:26.684850 4885 generic.go:334] "Generic (PLEG): container finished" podID="9df73c9d-22c1-47c0-a996-0983f206dad5" containerID="597d57d9acdd3cbb0ea1e6a0716e324596b7fdad2025c8334be56d83e0eb7b4a" exitCode=0 Oct 02 02:05:26 crc kubenswrapper[4885]: I1002 02:05:26.684901 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9df73c9d-22c1-47c0-a996-0983f206dad5","Type":"ContainerDied","Data":"597d57d9acdd3cbb0ea1e6a0716e324596b7fdad2025c8334be56d83e0eb7b4a"} Oct 02 02:05:26 crc kubenswrapper[4885]: I1002 02:05:26.693468 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"72053d12-6b1a-4df1-9fbe-ba2973d53dc3","Type":"ContainerStarted","Data":"fa295d66b1785b87309be69f6e91a2a2adb0fd6b866e9c326e188f45b8d98337"} Oct 02 02:05:26 crc kubenswrapper[4885]: I1002 02:05:26.695175 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.69516502 podStartE2EDuration="3.69516502s" podCreationTimestamp="2025-10-02 02:05:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:05:26.689674064 +0000 UTC m=+1115.501421463" watchObservedRunningTime="2025-10-02 02:05:26.69516502 +0000 UTC m=+1115.506912419" Oct 02 02:05:26 crc kubenswrapper[4885]: I1002 02:05:26.715494 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.715478138 podStartE2EDuration="4.715478138s" podCreationTimestamp="2025-10-02 02:05:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:05:26.710313072 +0000 UTC m=+1115.522060471" watchObservedRunningTime="2025-10-02 02:05:26.715478138 +0000 UTC m=+1115.527225537" Oct 02 02:05:26 crc kubenswrapper[4885]: I1002 02:05:26.716338 4885 scope.go:117] "RemoveContainer" containerID="2a5ee4d55456ceb8c5773da638d05738c8b83706c49ead5b8fabb99d0934b508" Oct 02 02:05:26 crc kubenswrapper[4885]: I1002 02:05:26.730620 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 02 02:05:26 crc kubenswrapper[4885]: I1002 02:05:26.743517 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Oct 02 02:05:26 crc kubenswrapper[4885]: I1002 02:05:26.753504 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 02 02:05:26 crc kubenswrapper[4885]: E1002 02:05:26.753953 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f985ab4a-2149-40c7-9af9-5ac242974f28" containerName="cinder-api-log" Oct 02 02:05:26 crc kubenswrapper[4885]: I1002 02:05:26.753967 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="f985ab4a-2149-40c7-9af9-5ac242974f28" containerName="cinder-api-log" Oct 02 02:05:26 crc kubenswrapper[4885]: E1002 02:05:26.754001 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f985ab4a-2149-40c7-9af9-5ac242974f28" containerName="cinder-api" Oct 02 02:05:26 crc kubenswrapper[4885]: I1002 02:05:26.754007 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="f985ab4a-2149-40c7-9af9-5ac242974f28" containerName="cinder-api" Oct 02 02:05:26 crc kubenswrapper[4885]: I1002 02:05:26.754185 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="f985ab4a-2149-40c7-9af9-5ac242974f28" containerName="cinder-api-log" Oct 02 02:05:26 crc kubenswrapper[4885]: I1002 02:05:26.754206 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="f985ab4a-2149-40c7-9af9-5ac242974f28" containerName="cinder-api" Oct 02 02:05:26 crc kubenswrapper[4885]: I1002 02:05:26.755130 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 02 02:05:26 crc kubenswrapper[4885]: I1002 02:05:26.758902 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Oct 02 02:05:26 crc kubenswrapper[4885]: I1002 02:05:26.759083 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 02 02:05:26 crc kubenswrapper[4885]: I1002 02:05:26.759160 4885 scope.go:117] "RemoveContainer" containerID="3dec1875c31681ffdcc4fa041fc6c6f1e2bf30faea41cc03cd5081ca94280caf" Oct 02 02:05:26 crc kubenswrapper[4885]: I1002 02:05:26.759185 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Oct 02 02:05:26 crc kubenswrapper[4885]: E1002 02:05:26.761031 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3dec1875c31681ffdcc4fa041fc6c6f1e2bf30faea41cc03cd5081ca94280caf\": container with ID starting with 3dec1875c31681ffdcc4fa041fc6c6f1e2bf30faea41cc03cd5081ca94280caf not found: ID does not exist" containerID="3dec1875c31681ffdcc4fa041fc6c6f1e2bf30faea41cc03cd5081ca94280caf" Oct 02 02:05:26 crc kubenswrapper[4885]: I1002 02:05:26.761063 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3dec1875c31681ffdcc4fa041fc6c6f1e2bf30faea41cc03cd5081ca94280caf"} err="failed to get container status \"3dec1875c31681ffdcc4fa041fc6c6f1e2bf30faea41cc03cd5081ca94280caf\": rpc error: code = NotFound desc = could not find container \"3dec1875c31681ffdcc4fa041fc6c6f1e2bf30faea41cc03cd5081ca94280caf\": container with ID starting with 3dec1875c31681ffdcc4fa041fc6c6f1e2bf30faea41cc03cd5081ca94280caf not found: ID does not exist" Oct 02 02:05:26 crc kubenswrapper[4885]: I1002 02:05:26.761084 4885 scope.go:117] "RemoveContainer" containerID="2a5ee4d55456ceb8c5773da638d05738c8b83706c49ead5b8fabb99d0934b508" Oct 02 02:05:26 crc kubenswrapper[4885]: I1002 02:05:26.762032 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=6.492752926 podStartE2EDuration="7.762014413s" podCreationTimestamp="2025-10-02 02:05:19 +0000 UTC" firstStartedPulling="2025-10-02 02:05:22.400088835 +0000 UTC m=+1111.211836234" lastFinishedPulling="2025-10-02 02:05:23.669350322 +0000 UTC m=+1112.481097721" observedRunningTime="2025-10-02 02:05:26.744696936 +0000 UTC m=+1115.556444335" watchObservedRunningTime="2025-10-02 02:05:26.762014413 +0000 UTC m=+1115.573761812" Oct 02 02:05:26 crc kubenswrapper[4885]: E1002 02:05:26.764591 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2a5ee4d55456ceb8c5773da638d05738c8b83706c49ead5b8fabb99d0934b508\": container with ID starting with 2a5ee4d55456ceb8c5773da638d05738c8b83706c49ead5b8fabb99d0934b508 not found: ID does not exist" containerID="2a5ee4d55456ceb8c5773da638d05738c8b83706c49ead5b8fabb99d0934b508" Oct 02 02:05:26 crc kubenswrapper[4885]: I1002 02:05:26.764618 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2a5ee4d55456ceb8c5773da638d05738c8b83706c49ead5b8fabb99d0934b508"} err="failed to get container status \"2a5ee4d55456ceb8c5773da638d05738c8b83706c49ead5b8fabb99d0934b508\": rpc error: code = NotFound desc = could not find container \"2a5ee4d55456ceb8c5773da638d05738c8b83706c49ead5b8fabb99d0934b508\": container with ID starting with 2a5ee4d55456ceb8c5773da638d05738c8b83706c49ead5b8fabb99d0934b508 not found: ID does not exist" Oct 02 02:05:26 crc kubenswrapper[4885]: I1002 02:05:26.764635 4885 scope.go:117] "RemoveContainer" containerID="3dec1875c31681ffdcc4fa041fc6c6f1e2bf30faea41cc03cd5081ca94280caf" Oct 02 02:05:26 crc kubenswrapper[4885]: I1002 02:05:26.769403 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3dec1875c31681ffdcc4fa041fc6c6f1e2bf30faea41cc03cd5081ca94280caf"} err="failed to get container status \"3dec1875c31681ffdcc4fa041fc6c6f1e2bf30faea41cc03cd5081ca94280caf\": rpc error: code = NotFound desc = could not find container \"3dec1875c31681ffdcc4fa041fc6c6f1e2bf30faea41cc03cd5081ca94280caf\": container with ID starting with 3dec1875c31681ffdcc4fa041fc6c6f1e2bf30faea41cc03cd5081ca94280caf not found: ID does not exist" Oct 02 02:05:26 crc kubenswrapper[4885]: I1002 02:05:26.769433 4885 scope.go:117] "RemoveContainer" containerID="2a5ee4d55456ceb8c5773da638d05738c8b83706c49ead5b8fabb99d0934b508" Oct 02 02:05:26 crc kubenswrapper[4885]: I1002 02:05:26.772638 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2a5ee4d55456ceb8c5773da638d05738c8b83706c49ead5b8fabb99d0934b508"} err="failed to get container status \"2a5ee4d55456ceb8c5773da638d05738c8b83706c49ead5b8fabb99d0934b508\": rpc error: code = NotFound desc = could not find container \"2a5ee4d55456ceb8c5773da638d05738c8b83706c49ead5b8fabb99d0934b508\": container with ID starting with 2a5ee4d55456ceb8c5773da638d05738c8b83706c49ead5b8fabb99d0934b508 not found: ID does not exist" Oct 02 02:05:26 crc kubenswrapper[4885]: I1002 02:05:26.780712 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 02 02:05:26 crc kubenswrapper[4885]: I1002 02:05:26.826213 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 02:05:26 crc kubenswrapper[4885]: I1002 02:05:26.895954 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9df73c9d-22c1-47c0-a996-0983f206dad5-combined-ca-bundle\") pod \"9df73c9d-22c1-47c0-a996-0983f206dad5\" (UID: \"9df73c9d-22c1-47c0-a996-0983f206dad5\") " Oct 02 02:05:26 crc kubenswrapper[4885]: I1002 02:05:26.896103 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9df73c9d-22c1-47c0-a996-0983f206dad5-log-httpd\") pod \"9df73c9d-22c1-47c0-a996-0983f206dad5\" (UID: \"9df73c9d-22c1-47c0-a996-0983f206dad5\") " Oct 02 02:05:26 crc kubenswrapper[4885]: I1002 02:05:26.896143 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9df73c9d-22c1-47c0-a996-0983f206dad5-config-data\") pod \"9df73c9d-22c1-47c0-a996-0983f206dad5\" (UID: \"9df73c9d-22c1-47c0-a996-0983f206dad5\") " Oct 02 02:05:26 crc kubenswrapper[4885]: I1002 02:05:26.896185 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9df73c9d-22c1-47c0-a996-0983f206dad5-run-httpd\") pod \"9df73c9d-22c1-47c0-a996-0983f206dad5\" (UID: \"9df73c9d-22c1-47c0-a996-0983f206dad5\") " Oct 02 02:05:26 crc kubenswrapper[4885]: I1002 02:05:26.896228 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6jr2d\" (UniqueName: \"kubernetes.io/projected/9df73c9d-22c1-47c0-a996-0983f206dad5-kube-api-access-6jr2d\") pod \"9df73c9d-22c1-47c0-a996-0983f206dad5\" (UID: \"9df73c9d-22c1-47c0-a996-0983f206dad5\") " Oct 02 02:05:26 crc kubenswrapper[4885]: I1002 02:05:26.896270 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9df73c9d-22c1-47c0-a996-0983f206dad5-sg-core-conf-yaml\") pod \"9df73c9d-22c1-47c0-a996-0983f206dad5\" (UID: \"9df73c9d-22c1-47c0-a996-0983f206dad5\") " Oct 02 02:05:26 crc kubenswrapper[4885]: I1002 02:05:26.896311 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9df73c9d-22c1-47c0-a996-0983f206dad5-scripts\") pod \"9df73c9d-22c1-47c0-a996-0983f206dad5\" (UID: \"9df73c9d-22c1-47c0-a996-0983f206dad5\") " Oct 02 02:05:26 crc kubenswrapper[4885]: I1002 02:05:26.897197 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9df73c9d-22c1-47c0-a996-0983f206dad5-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "9df73c9d-22c1-47c0-a996-0983f206dad5" (UID: "9df73c9d-22c1-47c0-a996-0983f206dad5"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:05:26 crc kubenswrapper[4885]: I1002 02:05:26.897185 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9df73c9d-22c1-47c0-a996-0983f206dad5-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "9df73c9d-22c1-47c0-a996-0983f206dad5" (UID: "9df73c9d-22c1-47c0-a996-0983f206dad5"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:05:26 crc kubenswrapper[4885]: I1002 02:05:26.897392 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ca40a36d-76b5-4cfe-8f8c-ccb4625d6f56-etc-machine-id\") pod \"cinder-api-0\" (UID: \"ca40a36d-76b5-4cfe-8f8c-ccb4625d6f56\") " pod="openstack/cinder-api-0" Oct 02 02:05:26 crc kubenswrapper[4885]: I1002 02:05:26.897466 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ca40a36d-76b5-4cfe-8f8c-ccb4625d6f56-config-data-custom\") pod \"cinder-api-0\" (UID: \"ca40a36d-76b5-4cfe-8f8c-ccb4625d6f56\") " pod="openstack/cinder-api-0" Oct 02 02:05:26 crc kubenswrapper[4885]: I1002 02:05:26.897491 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-98lvd\" (UniqueName: \"kubernetes.io/projected/ca40a36d-76b5-4cfe-8f8c-ccb4625d6f56-kube-api-access-98lvd\") pod \"cinder-api-0\" (UID: \"ca40a36d-76b5-4cfe-8f8c-ccb4625d6f56\") " pod="openstack/cinder-api-0" Oct 02 02:05:26 crc kubenswrapper[4885]: I1002 02:05:26.897563 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca40a36d-76b5-4cfe-8f8c-ccb4625d6f56-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"ca40a36d-76b5-4cfe-8f8c-ccb4625d6f56\") " pod="openstack/cinder-api-0" Oct 02 02:05:26 crc kubenswrapper[4885]: I1002 02:05:26.897638 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca40a36d-76b5-4cfe-8f8c-ccb4625d6f56-public-tls-certs\") pod \"cinder-api-0\" (UID: \"ca40a36d-76b5-4cfe-8f8c-ccb4625d6f56\") " pod="openstack/cinder-api-0" Oct 02 02:05:26 crc kubenswrapper[4885]: I1002 02:05:26.897653 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca40a36d-76b5-4cfe-8f8c-ccb4625d6f56-config-data\") pod \"cinder-api-0\" (UID: \"ca40a36d-76b5-4cfe-8f8c-ccb4625d6f56\") " pod="openstack/cinder-api-0" Oct 02 02:05:26 crc kubenswrapper[4885]: I1002 02:05:26.897708 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca40a36d-76b5-4cfe-8f8c-ccb4625d6f56-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"ca40a36d-76b5-4cfe-8f8c-ccb4625d6f56\") " pod="openstack/cinder-api-0" Oct 02 02:05:26 crc kubenswrapper[4885]: I1002 02:05:26.897723 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ca40a36d-76b5-4cfe-8f8c-ccb4625d6f56-scripts\") pod \"cinder-api-0\" (UID: \"ca40a36d-76b5-4cfe-8f8c-ccb4625d6f56\") " pod="openstack/cinder-api-0" Oct 02 02:05:26 crc kubenswrapper[4885]: I1002 02:05:26.897764 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ca40a36d-76b5-4cfe-8f8c-ccb4625d6f56-logs\") pod \"cinder-api-0\" (UID: \"ca40a36d-76b5-4cfe-8f8c-ccb4625d6f56\") " pod="openstack/cinder-api-0" Oct 02 02:05:26 crc kubenswrapper[4885]: I1002 02:05:26.899401 4885 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9df73c9d-22c1-47c0-a996-0983f206dad5-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:26 crc kubenswrapper[4885]: I1002 02:05:26.899427 4885 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9df73c9d-22c1-47c0-a996-0983f206dad5-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:26 crc kubenswrapper[4885]: I1002 02:05:26.901036 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9df73c9d-22c1-47c0-a996-0983f206dad5-kube-api-access-6jr2d" (OuterVolumeSpecName: "kube-api-access-6jr2d") pod "9df73c9d-22c1-47c0-a996-0983f206dad5" (UID: "9df73c9d-22c1-47c0-a996-0983f206dad5"). InnerVolumeSpecName "kube-api-access-6jr2d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:05:26 crc kubenswrapper[4885]: I1002 02:05:26.901109 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9df73c9d-22c1-47c0-a996-0983f206dad5-scripts" (OuterVolumeSpecName: "scripts") pod "9df73c9d-22c1-47c0-a996-0983f206dad5" (UID: "9df73c9d-22c1-47c0-a996-0983f206dad5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:05:26 crc kubenswrapper[4885]: I1002 02:05:26.919153 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9df73c9d-22c1-47c0-a996-0983f206dad5-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "9df73c9d-22c1-47c0-a996-0983f206dad5" (UID: "9df73c9d-22c1-47c0-a996-0983f206dad5"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:05:26 crc kubenswrapper[4885]: I1002 02:05:26.968640 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9df73c9d-22c1-47c0-a996-0983f206dad5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9df73c9d-22c1-47c0-a996-0983f206dad5" (UID: "9df73c9d-22c1-47c0-a996-0983f206dad5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:05:26 crc kubenswrapper[4885]: I1002 02:05:26.996388 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9df73c9d-22c1-47c0-a996-0983f206dad5-config-data" (OuterVolumeSpecName: "config-data") pod "9df73c9d-22c1-47c0-a996-0983f206dad5" (UID: "9df73c9d-22c1-47c0-a996-0983f206dad5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:05:27 crc kubenswrapper[4885]: I1002 02:05:27.001160 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca40a36d-76b5-4cfe-8f8c-ccb4625d6f56-public-tls-certs\") pod \"cinder-api-0\" (UID: \"ca40a36d-76b5-4cfe-8f8c-ccb4625d6f56\") " pod="openstack/cinder-api-0" Oct 02 02:05:27 crc kubenswrapper[4885]: I1002 02:05:27.001192 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca40a36d-76b5-4cfe-8f8c-ccb4625d6f56-config-data\") pod \"cinder-api-0\" (UID: \"ca40a36d-76b5-4cfe-8f8c-ccb4625d6f56\") " pod="openstack/cinder-api-0" Oct 02 02:05:27 crc kubenswrapper[4885]: I1002 02:05:27.001235 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ca40a36d-76b5-4cfe-8f8c-ccb4625d6f56-scripts\") pod \"cinder-api-0\" (UID: \"ca40a36d-76b5-4cfe-8f8c-ccb4625d6f56\") " pod="openstack/cinder-api-0" Oct 02 02:05:27 crc kubenswrapper[4885]: I1002 02:05:27.001267 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca40a36d-76b5-4cfe-8f8c-ccb4625d6f56-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"ca40a36d-76b5-4cfe-8f8c-ccb4625d6f56\") " pod="openstack/cinder-api-0" Oct 02 02:05:27 crc kubenswrapper[4885]: I1002 02:05:27.001294 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ca40a36d-76b5-4cfe-8f8c-ccb4625d6f56-logs\") pod \"cinder-api-0\" (UID: \"ca40a36d-76b5-4cfe-8f8c-ccb4625d6f56\") " pod="openstack/cinder-api-0" Oct 02 02:05:27 crc kubenswrapper[4885]: I1002 02:05:27.001328 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ca40a36d-76b5-4cfe-8f8c-ccb4625d6f56-etc-machine-id\") pod \"cinder-api-0\" (UID: \"ca40a36d-76b5-4cfe-8f8c-ccb4625d6f56\") " pod="openstack/cinder-api-0" Oct 02 02:05:27 crc kubenswrapper[4885]: I1002 02:05:27.001354 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ca40a36d-76b5-4cfe-8f8c-ccb4625d6f56-config-data-custom\") pod \"cinder-api-0\" (UID: \"ca40a36d-76b5-4cfe-8f8c-ccb4625d6f56\") " pod="openstack/cinder-api-0" Oct 02 02:05:27 crc kubenswrapper[4885]: I1002 02:05:27.001373 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-98lvd\" (UniqueName: \"kubernetes.io/projected/ca40a36d-76b5-4cfe-8f8c-ccb4625d6f56-kube-api-access-98lvd\") pod \"cinder-api-0\" (UID: \"ca40a36d-76b5-4cfe-8f8c-ccb4625d6f56\") " pod="openstack/cinder-api-0" Oct 02 02:05:27 crc kubenswrapper[4885]: I1002 02:05:27.001428 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca40a36d-76b5-4cfe-8f8c-ccb4625d6f56-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"ca40a36d-76b5-4cfe-8f8c-ccb4625d6f56\") " pod="openstack/cinder-api-0" Oct 02 02:05:27 crc kubenswrapper[4885]: I1002 02:05:27.001479 4885 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9df73c9d-22c1-47c0-a996-0983f206dad5-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:27 crc kubenswrapper[4885]: I1002 02:05:27.001490 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6jr2d\" (UniqueName: \"kubernetes.io/projected/9df73c9d-22c1-47c0-a996-0983f206dad5-kube-api-access-6jr2d\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:27 crc kubenswrapper[4885]: I1002 02:05:27.001499 4885 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9df73c9d-22c1-47c0-a996-0983f206dad5-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:27 crc kubenswrapper[4885]: I1002 02:05:27.001507 4885 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9df73c9d-22c1-47c0-a996-0983f206dad5-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:27 crc kubenswrapper[4885]: I1002 02:05:27.001517 4885 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9df73c9d-22c1-47c0-a996-0983f206dad5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:27 crc kubenswrapper[4885]: I1002 02:05:27.001984 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ca40a36d-76b5-4cfe-8f8c-ccb4625d6f56-logs\") pod \"cinder-api-0\" (UID: \"ca40a36d-76b5-4cfe-8f8c-ccb4625d6f56\") " pod="openstack/cinder-api-0" Oct 02 02:05:27 crc kubenswrapper[4885]: I1002 02:05:27.002028 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ca40a36d-76b5-4cfe-8f8c-ccb4625d6f56-etc-machine-id\") pod \"cinder-api-0\" (UID: \"ca40a36d-76b5-4cfe-8f8c-ccb4625d6f56\") " pod="openstack/cinder-api-0" Oct 02 02:05:27 crc kubenswrapper[4885]: I1002 02:05:27.005214 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ca40a36d-76b5-4cfe-8f8c-ccb4625d6f56-config-data-custom\") pod \"cinder-api-0\" (UID: \"ca40a36d-76b5-4cfe-8f8c-ccb4625d6f56\") " pod="openstack/cinder-api-0" Oct 02 02:05:27 crc kubenswrapper[4885]: I1002 02:05:27.005648 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca40a36d-76b5-4cfe-8f8c-ccb4625d6f56-config-data\") pod \"cinder-api-0\" (UID: \"ca40a36d-76b5-4cfe-8f8c-ccb4625d6f56\") " pod="openstack/cinder-api-0" Oct 02 02:05:27 crc kubenswrapper[4885]: I1002 02:05:27.007625 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca40a36d-76b5-4cfe-8f8c-ccb4625d6f56-public-tls-certs\") pod \"cinder-api-0\" (UID: \"ca40a36d-76b5-4cfe-8f8c-ccb4625d6f56\") " pod="openstack/cinder-api-0" Oct 02 02:05:27 crc kubenswrapper[4885]: I1002 02:05:27.009563 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca40a36d-76b5-4cfe-8f8c-ccb4625d6f56-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"ca40a36d-76b5-4cfe-8f8c-ccb4625d6f56\") " pod="openstack/cinder-api-0" Oct 02 02:05:27 crc kubenswrapper[4885]: I1002 02:05:27.010747 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ca40a36d-76b5-4cfe-8f8c-ccb4625d6f56-scripts\") pod \"cinder-api-0\" (UID: \"ca40a36d-76b5-4cfe-8f8c-ccb4625d6f56\") " pod="openstack/cinder-api-0" Oct 02 02:05:27 crc kubenswrapper[4885]: I1002 02:05:27.011869 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca40a36d-76b5-4cfe-8f8c-ccb4625d6f56-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"ca40a36d-76b5-4cfe-8f8c-ccb4625d6f56\") " pod="openstack/cinder-api-0" Oct 02 02:05:27 crc kubenswrapper[4885]: I1002 02:05:27.020383 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-98lvd\" (UniqueName: \"kubernetes.io/projected/ca40a36d-76b5-4cfe-8f8c-ccb4625d6f56-kube-api-access-98lvd\") pod \"cinder-api-0\" (UID: \"ca40a36d-76b5-4cfe-8f8c-ccb4625d6f56\") " pod="openstack/cinder-api-0" Oct 02 02:05:27 crc kubenswrapper[4885]: I1002 02:05:27.122863 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 02 02:05:27 crc kubenswrapper[4885]: I1002 02:05:27.565902 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 02 02:05:27 crc kubenswrapper[4885]: W1002 02:05:27.569695 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podca40a36d_76b5_4cfe_8f8c_ccb4625d6f56.slice/crio-6622cbf3a5ea27ae7f50ebf1db54a4cec49899ea6c11e2e72b444916fd4a691a WatchSource:0}: Error finding container 6622cbf3a5ea27ae7f50ebf1db54a4cec49899ea6c11e2e72b444916fd4a691a: Status 404 returned error can't find the container with id 6622cbf3a5ea27ae7f50ebf1db54a4cec49899ea6c11e2e72b444916fd4a691a Oct 02 02:05:27 crc kubenswrapper[4885]: I1002 02:05:27.710404 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"ca40a36d-76b5-4cfe-8f8c-ccb4625d6f56","Type":"ContainerStarted","Data":"6622cbf3a5ea27ae7f50ebf1db54a4cec49899ea6c11e2e72b444916fd4a691a"} Oct 02 02:05:27 crc kubenswrapper[4885]: I1002 02:05:27.713493 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9df73c9d-22c1-47c0-a996-0983f206dad5","Type":"ContainerDied","Data":"7afdecbaf1bfab89170f887053ae340028a863e6810764f52b01aebac9f7ee08"} Oct 02 02:05:27 crc kubenswrapper[4885]: I1002 02:05:27.713537 4885 scope.go:117] "RemoveContainer" containerID="f2f63cc40789d147d4c97ba73e7b6f04c28d928c576d2b12411df129a24ece8f" Oct 02 02:05:27 crc kubenswrapper[4885]: I1002 02:05:27.713569 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 02:05:27 crc kubenswrapper[4885]: I1002 02:05:27.766581 4885 scope.go:117] "RemoveContainer" containerID="d8383b35ce2909f3835cd2327a42f284bd94f3880d123202563b36247af92b23" Oct 02 02:05:27 crc kubenswrapper[4885]: I1002 02:05:27.783455 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 02:05:27 crc kubenswrapper[4885]: I1002 02:05:27.796183 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 02 02:05:27 crc kubenswrapper[4885]: I1002 02:05:27.815057 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 02 02:05:27 crc kubenswrapper[4885]: E1002 02:05:27.815557 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9df73c9d-22c1-47c0-a996-0983f206dad5" containerName="proxy-httpd" Oct 02 02:05:27 crc kubenswrapper[4885]: I1002 02:05:27.815575 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="9df73c9d-22c1-47c0-a996-0983f206dad5" containerName="proxy-httpd" Oct 02 02:05:27 crc kubenswrapper[4885]: E1002 02:05:27.815599 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9df73c9d-22c1-47c0-a996-0983f206dad5" containerName="sg-core" Oct 02 02:05:27 crc kubenswrapper[4885]: I1002 02:05:27.815607 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="9df73c9d-22c1-47c0-a996-0983f206dad5" containerName="sg-core" Oct 02 02:05:27 crc kubenswrapper[4885]: E1002 02:05:27.815618 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9df73c9d-22c1-47c0-a996-0983f206dad5" containerName="ceilometer-notification-agent" Oct 02 02:05:27 crc kubenswrapper[4885]: I1002 02:05:27.815625 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="9df73c9d-22c1-47c0-a996-0983f206dad5" containerName="ceilometer-notification-agent" Oct 02 02:05:27 crc kubenswrapper[4885]: E1002 02:05:27.815647 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9df73c9d-22c1-47c0-a996-0983f206dad5" containerName="ceilometer-central-agent" Oct 02 02:05:27 crc kubenswrapper[4885]: I1002 02:05:27.815653 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="9df73c9d-22c1-47c0-a996-0983f206dad5" containerName="ceilometer-central-agent" Oct 02 02:05:27 crc kubenswrapper[4885]: I1002 02:05:27.815898 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="9df73c9d-22c1-47c0-a996-0983f206dad5" containerName="sg-core" Oct 02 02:05:27 crc kubenswrapper[4885]: I1002 02:05:27.815917 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="9df73c9d-22c1-47c0-a996-0983f206dad5" containerName="ceilometer-notification-agent" Oct 02 02:05:27 crc kubenswrapper[4885]: I1002 02:05:27.815934 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="9df73c9d-22c1-47c0-a996-0983f206dad5" containerName="proxy-httpd" Oct 02 02:05:27 crc kubenswrapper[4885]: I1002 02:05:27.815947 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="9df73c9d-22c1-47c0-a996-0983f206dad5" containerName="ceilometer-central-agent" Oct 02 02:05:27 crc kubenswrapper[4885]: I1002 02:05:27.817761 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 02:05:27 crc kubenswrapper[4885]: I1002 02:05:27.820477 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 02 02:05:27 crc kubenswrapper[4885]: I1002 02:05:27.820532 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 02 02:05:27 crc kubenswrapper[4885]: I1002 02:05:27.836911 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 02:05:27 crc kubenswrapper[4885]: I1002 02:05:27.854801 4885 scope.go:117] "RemoveContainer" containerID="597d57d9acdd3cbb0ea1e6a0716e324596b7fdad2025c8334be56d83e0eb7b4a" Oct 02 02:05:27 crc kubenswrapper[4885]: I1002 02:05:27.873585 4885 scope.go:117] "RemoveContainer" containerID="9706f642020809b48c4b255916f1caf766eb34203bb17a12bfc4d4acb363e7d3" Oct 02 02:05:27 crc kubenswrapper[4885]: I1002 02:05:27.919484 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cg6n5\" (UniqueName: \"kubernetes.io/projected/097aaa2c-cd9c-4a5c-b04b-110fac788275-kube-api-access-cg6n5\") pod \"ceilometer-0\" (UID: \"097aaa2c-cd9c-4a5c-b04b-110fac788275\") " pod="openstack/ceilometer-0" Oct 02 02:05:27 crc kubenswrapper[4885]: I1002 02:05:27.919764 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/097aaa2c-cd9c-4a5c-b04b-110fac788275-config-data\") pod \"ceilometer-0\" (UID: \"097aaa2c-cd9c-4a5c-b04b-110fac788275\") " pod="openstack/ceilometer-0" Oct 02 02:05:27 crc kubenswrapper[4885]: I1002 02:05:27.919839 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/097aaa2c-cd9c-4a5c-b04b-110fac788275-run-httpd\") pod \"ceilometer-0\" (UID: \"097aaa2c-cd9c-4a5c-b04b-110fac788275\") " pod="openstack/ceilometer-0" Oct 02 02:05:27 crc kubenswrapper[4885]: I1002 02:05:27.920018 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/097aaa2c-cd9c-4a5c-b04b-110fac788275-log-httpd\") pod \"ceilometer-0\" (UID: \"097aaa2c-cd9c-4a5c-b04b-110fac788275\") " pod="openstack/ceilometer-0" Oct 02 02:05:27 crc kubenswrapper[4885]: I1002 02:05:27.920056 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/097aaa2c-cd9c-4a5c-b04b-110fac788275-scripts\") pod \"ceilometer-0\" (UID: \"097aaa2c-cd9c-4a5c-b04b-110fac788275\") " pod="openstack/ceilometer-0" Oct 02 02:05:27 crc kubenswrapper[4885]: I1002 02:05:27.920074 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/097aaa2c-cd9c-4a5c-b04b-110fac788275-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"097aaa2c-cd9c-4a5c-b04b-110fac788275\") " pod="openstack/ceilometer-0" Oct 02 02:05:27 crc kubenswrapper[4885]: I1002 02:05:27.920102 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/097aaa2c-cd9c-4a5c-b04b-110fac788275-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"097aaa2c-cd9c-4a5c-b04b-110fac788275\") " pod="openstack/ceilometer-0" Oct 02 02:05:28 crc kubenswrapper[4885]: I1002 02:05:28.022121 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/097aaa2c-cd9c-4a5c-b04b-110fac788275-config-data\") pod \"ceilometer-0\" (UID: \"097aaa2c-cd9c-4a5c-b04b-110fac788275\") " pod="openstack/ceilometer-0" Oct 02 02:05:28 crc kubenswrapper[4885]: I1002 02:05:28.022186 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/097aaa2c-cd9c-4a5c-b04b-110fac788275-run-httpd\") pod \"ceilometer-0\" (UID: \"097aaa2c-cd9c-4a5c-b04b-110fac788275\") " pod="openstack/ceilometer-0" Oct 02 02:05:28 crc kubenswrapper[4885]: I1002 02:05:28.022242 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/097aaa2c-cd9c-4a5c-b04b-110fac788275-log-httpd\") pod \"ceilometer-0\" (UID: \"097aaa2c-cd9c-4a5c-b04b-110fac788275\") " pod="openstack/ceilometer-0" Oct 02 02:05:28 crc kubenswrapper[4885]: I1002 02:05:28.022290 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/097aaa2c-cd9c-4a5c-b04b-110fac788275-scripts\") pod \"ceilometer-0\" (UID: \"097aaa2c-cd9c-4a5c-b04b-110fac788275\") " pod="openstack/ceilometer-0" Oct 02 02:05:28 crc kubenswrapper[4885]: I1002 02:05:28.022312 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/097aaa2c-cd9c-4a5c-b04b-110fac788275-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"097aaa2c-cd9c-4a5c-b04b-110fac788275\") " pod="openstack/ceilometer-0" Oct 02 02:05:28 crc kubenswrapper[4885]: I1002 02:05:28.022336 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/097aaa2c-cd9c-4a5c-b04b-110fac788275-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"097aaa2c-cd9c-4a5c-b04b-110fac788275\") " pod="openstack/ceilometer-0" Oct 02 02:05:28 crc kubenswrapper[4885]: I1002 02:05:28.022424 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cg6n5\" (UniqueName: \"kubernetes.io/projected/097aaa2c-cd9c-4a5c-b04b-110fac788275-kube-api-access-cg6n5\") pod \"ceilometer-0\" (UID: \"097aaa2c-cd9c-4a5c-b04b-110fac788275\") " pod="openstack/ceilometer-0" Oct 02 02:05:28 crc kubenswrapper[4885]: I1002 02:05:28.022737 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/097aaa2c-cd9c-4a5c-b04b-110fac788275-run-httpd\") pod \"ceilometer-0\" (UID: \"097aaa2c-cd9c-4a5c-b04b-110fac788275\") " pod="openstack/ceilometer-0" Oct 02 02:05:28 crc kubenswrapper[4885]: I1002 02:05:28.022802 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/097aaa2c-cd9c-4a5c-b04b-110fac788275-log-httpd\") pod \"ceilometer-0\" (UID: \"097aaa2c-cd9c-4a5c-b04b-110fac788275\") " pod="openstack/ceilometer-0" Oct 02 02:05:28 crc kubenswrapper[4885]: I1002 02:05:28.030760 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/097aaa2c-cd9c-4a5c-b04b-110fac788275-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"097aaa2c-cd9c-4a5c-b04b-110fac788275\") " pod="openstack/ceilometer-0" Oct 02 02:05:28 crc kubenswrapper[4885]: I1002 02:05:28.031088 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/097aaa2c-cd9c-4a5c-b04b-110fac788275-config-data\") pod \"ceilometer-0\" (UID: \"097aaa2c-cd9c-4a5c-b04b-110fac788275\") " pod="openstack/ceilometer-0" Oct 02 02:05:28 crc kubenswrapper[4885]: I1002 02:05:28.031113 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/097aaa2c-cd9c-4a5c-b04b-110fac788275-scripts\") pod \"ceilometer-0\" (UID: \"097aaa2c-cd9c-4a5c-b04b-110fac788275\") " pod="openstack/ceilometer-0" Oct 02 02:05:28 crc kubenswrapper[4885]: I1002 02:05:28.033006 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/097aaa2c-cd9c-4a5c-b04b-110fac788275-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"097aaa2c-cd9c-4a5c-b04b-110fac788275\") " pod="openstack/ceilometer-0" Oct 02 02:05:28 crc kubenswrapper[4885]: I1002 02:05:28.039095 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cg6n5\" (UniqueName: \"kubernetes.io/projected/097aaa2c-cd9c-4a5c-b04b-110fac788275-kube-api-access-cg6n5\") pod \"ceilometer-0\" (UID: \"097aaa2c-cd9c-4a5c-b04b-110fac788275\") " pod="openstack/ceilometer-0" Oct 02 02:05:28 crc kubenswrapper[4885]: I1002 02:05:28.058030 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9df73c9d-22c1-47c0-a996-0983f206dad5" path="/var/lib/kubelet/pods/9df73c9d-22c1-47c0-a996-0983f206dad5/volumes" Oct 02 02:05:28 crc kubenswrapper[4885]: I1002 02:05:28.059426 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f985ab4a-2149-40c7-9af9-5ac242974f28" path="/var/lib/kubelet/pods/f985ab4a-2149-40c7-9af9-5ac242974f28/volumes" Oct 02 02:05:28 crc kubenswrapper[4885]: I1002 02:05:28.136129 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 02:05:28 crc kubenswrapper[4885]: I1002 02:05:28.646449 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 02:05:28 crc kubenswrapper[4885]: I1002 02:05:28.735101 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"ca40a36d-76b5-4cfe-8f8c-ccb4625d6f56","Type":"ContainerStarted","Data":"19e3d1b5f6cc869c65535732f4cd5489d4de649c49d3d1d031b0440209ba486e"} Oct 02 02:05:28 crc kubenswrapper[4885]: I1002 02:05:28.739673 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"097aaa2c-cd9c-4a5c-b04b-110fac788275","Type":"ContainerStarted","Data":"3d7b30474d6c2da89c60385c1e31c04baebe3cb12fa7bc8206eef86ec9707b86"} Oct 02 02:05:29 crc kubenswrapper[4885]: I1002 02:05:29.569528 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5784cf869f-85ntb" Oct 02 02:05:29 crc kubenswrapper[4885]: I1002 02:05:29.671294 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-h9tlf"] Oct 02 02:05:29 crc kubenswrapper[4885]: I1002 02:05:29.671685 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-84b966f6c9-h9tlf" podUID="00531889-6c5d-444b-bfeb-4bb628d7ef94" containerName="dnsmasq-dns" containerID="cri-o://712c4978547cdf1d6a36eaa4a251de9c427d874458af321b65b0c84c2922da84" gracePeriod=10 Oct 02 02:05:29 crc kubenswrapper[4885]: I1002 02:05:29.700896 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 02 02:05:29 crc kubenswrapper[4885]: I1002 02:05:29.751719 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"ca40a36d-76b5-4cfe-8f8c-ccb4625d6f56","Type":"ContainerStarted","Data":"db4f8a292df4416f62cfba825e61c6f3bc73650f4704f647bbba69ddb2425f6e"} Oct 02 02:05:29 crc kubenswrapper[4885]: I1002 02:05:29.752014 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 02 02:05:29 crc kubenswrapper[4885]: I1002 02:05:29.754497 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"097aaa2c-cd9c-4a5c-b04b-110fac788275","Type":"ContainerStarted","Data":"ba193af1f4d20a5098cd947ade8f513cd3a54b42c1babee5847299d20278bb85"} Oct 02 02:05:29 crc kubenswrapper[4885]: I1002 02:05:29.776817 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.776794636 podStartE2EDuration="3.776794636s" podCreationTimestamp="2025-10-02 02:05:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:05:29.771634258 +0000 UTC m=+1118.583381677" watchObservedRunningTime="2025-10-02 02:05:29.776794636 +0000 UTC m=+1118.588542035" Oct 02 02:05:29 crc kubenswrapper[4885]: I1002 02:05:29.977502 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 02 02:05:30 crc kubenswrapper[4885]: I1002 02:05:30.039178 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 02:05:30 crc kubenswrapper[4885]: I1002 02:05:30.191793 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84b966f6c9-h9tlf" Oct 02 02:05:30 crc kubenswrapper[4885]: I1002 02:05:30.267666 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/00531889-6c5d-444b-bfeb-4bb628d7ef94-ovsdbserver-sb\") pod \"00531889-6c5d-444b-bfeb-4bb628d7ef94\" (UID: \"00531889-6c5d-444b-bfeb-4bb628d7ef94\") " Oct 02 02:05:30 crc kubenswrapper[4885]: I1002 02:05:30.267792 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v5nt5\" (UniqueName: \"kubernetes.io/projected/00531889-6c5d-444b-bfeb-4bb628d7ef94-kube-api-access-v5nt5\") pod \"00531889-6c5d-444b-bfeb-4bb628d7ef94\" (UID: \"00531889-6c5d-444b-bfeb-4bb628d7ef94\") " Oct 02 02:05:30 crc kubenswrapper[4885]: I1002 02:05:30.267837 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/00531889-6c5d-444b-bfeb-4bb628d7ef94-dns-swift-storage-0\") pod \"00531889-6c5d-444b-bfeb-4bb628d7ef94\" (UID: \"00531889-6c5d-444b-bfeb-4bb628d7ef94\") " Oct 02 02:05:30 crc kubenswrapper[4885]: I1002 02:05:30.267887 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/00531889-6c5d-444b-bfeb-4bb628d7ef94-dns-svc\") pod \"00531889-6c5d-444b-bfeb-4bb628d7ef94\" (UID: \"00531889-6c5d-444b-bfeb-4bb628d7ef94\") " Oct 02 02:05:30 crc kubenswrapper[4885]: I1002 02:05:30.267916 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00531889-6c5d-444b-bfeb-4bb628d7ef94-config\") pod \"00531889-6c5d-444b-bfeb-4bb628d7ef94\" (UID: \"00531889-6c5d-444b-bfeb-4bb628d7ef94\") " Oct 02 02:05:30 crc kubenswrapper[4885]: I1002 02:05:30.267948 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/00531889-6c5d-444b-bfeb-4bb628d7ef94-ovsdbserver-nb\") pod \"00531889-6c5d-444b-bfeb-4bb628d7ef94\" (UID: \"00531889-6c5d-444b-bfeb-4bb628d7ef94\") " Oct 02 02:05:30 crc kubenswrapper[4885]: I1002 02:05:30.291443 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/00531889-6c5d-444b-bfeb-4bb628d7ef94-kube-api-access-v5nt5" (OuterVolumeSpecName: "kube-api-access-v5nt5") pod "00531889-6c5d-444b-bfeb-4bb628d7ef94" (UID: "00531889-6c5d-444b-bfeb-4bb628d7ef94"). InnerVolumeSpecName "kube-api-access-v5nt5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:05:30 crc kubenswrapper[4885]: I1002 02:05:30.338941 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/00531889-6c5d-444b-bfeb-4bb628d7ef94-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "00531889-6c5d-444b-bfeb-4bb628d7ef94" (UID: "00531889-6c5d-444b-bfeb-4bb628d7ef94"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:05:30 crc kubenswrapper[4885]: I1002 02:05:30.354219 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/00531889-6c5d-444b-bfeb-4bb628d7ef94-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "00531889-6c5d-444b-bfeb-4bb628d7ef94" (UID: "00531889-6c5d-444b-bfeb-4bb628d7ef94"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:05:30 crc kubenswrapper[4885]: I1002 02:05:30.370717 4885 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/00531889-6c5d-444b-bfeb-4bb628d7ef94-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:30 crc kubenswrapper[4885]: I1002 02:05:30.370763 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v5nt5\" (UniqueName: \"kubernetes.io/projected/00531889-6c5d-444b-bfeb-4bb628d7ef94-kube-api-access-v5nt5\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:30 crc kubenswrapper[4885]: I1002 02:05:30.370773 4885 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/00531889-6c5d-444b-bfeb-4bb628d7ef94-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:30 crc kubenswrapper[4885]: I1002 02:05:30.396087 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/00531889-6c5d-444b-bfeb-4bb628d7ef94-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "00531889-6c5d-444b-bfeb-4bb628d7ef94" (UID: "00531889-6c5d-444b-bfeb-4bb628d7ef94"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:05:30 crc kubenswrapper[4885]: I1002 02:05:30.396210 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/00531889-6c5d-444b-bfeb-4bb628d7ef94-config" (OuterVolumeSpecName: "config") pod "00531889-6c5d-444b-bfeb-4bb628d7ef94" (UID: "00531889-6c5d-444b-bfeb-4bb628d7ef94"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:05:30 crc kubenswrapper[4885]: I1002 02:05:30.462028 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/00531889-6c5d-444b-bfeb-4bb628d7ef94-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "00531889-6c5d-444b-bfeb-4bb628d7ef94" (UID: "00531889-6c5d-444b-bfeb-4bb628d7ef94"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:05:30 crc kubenswrapper[4885]: I1002 02:05:30.475354 4885 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/00531889-6c5d-444b-bfeb-4bb628d7ef94-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:30 crc kubenswrapper[4885]: I1002 02:05:30.475387 4885 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/00531889-6c5d-444b-bfeb-4bb628d7ef94-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:30 crc kubenswrapper[4885]: I1002 02:05:30.475396 4885 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00531889-6c5d-444b-bfeb-4bb628d7ef94-config\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:30 crc kubenswrapper[4885]: I1002 02:05:30.765720 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"097aaa2c-cd9c-4a5c-b04b-110fac788275","Type":"ContainerStarted","Data":"47702dc0fdfc3bfcb1c07064c38f6df754b7f6ca18c464e77a494e84badde2bb"} Oct 02 02:05:30 crc kubenswrapper[4885]: I1002 02:05:30.766049 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"097aaa2c-cd9c-4a5c-b04b-110fac788275","Type":"ContainerStarted","Data":"5c5f0ddd01d35c10dcf352859dd51106a1ac9e6f582a2e00233251cce937ecf8"} Oct 02 02:05:30 crc kubenswrapper[4885]: I1002 02:05:30.770904 4885 generic.go:334] "Generic (PLEG): container finished" podID="00531889-6c5d-444b-bfeb-4bb628d7ef94" containerID="712c4978547cdf1d6a36eaa4a251de9c427d874458af321b65b0c84c2922da84" exitCode=0 Oct 02 02:05:30 crc kubenswrapper[4885]: I1002 02:05:30.771116 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="72053d12-6b1a-4df1-9fbe-ba2973d53dc3" containerName="cinder-scheduler" containerID="cri-o://1bab2764732e628230707bff05f072c449d8f783b77372dfedcf2d5c5acb4e10" gracePeriod=30 Oct 02 02:05:30 crc kubenswrapper[4885]: I1002 02:05:30.771430 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84b966f6c9-h9tlf" Oct 02 02:05:30 crc kubenswrapper[4885]: I1002 02:05:30.774364 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b966f6c9-h9tlf" event={"ID":"00531889-6c5d-444b-bfeb-4bb628d7ef94","Type":"ContainerDied","Data":"712c4978547cdf1d6a36eaa4a251de9c427d874458af321b65b0c84c2922da84"} Oct 02 02:05:30 crc kubenswrapper[4885]: I1002 02:05:30.774379 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="72053d12-6b1a-4df1-9fbe-ba2973d53dc3" containerName="probe" containerID="cri-o://fa295d66b1785b87309be69f6e91a2a2adb0fd6b866e9c326e188f45b8d98337" gracePeriod=30 Oct 02 02:05:30 crc kubenswrapper[4885]: I1002 02:05:30.774430 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b966f6c9-h9tlf" event={"ID":"00531889-6c5d-444b-bfeb-4bb628d7ef94","Type":"ContainerDied","Data":"0e0d258454fe494f02c5999bf34c7c24e261eaab00218f001ef3b7cc432d49c3"} Oct 02 02:05:30 crc kubenswrapper[4885]: I1002 02:05:30.774451 4885 scope.go:117] "RemoveContainer" containerID="712c4978547cdf1d6a36eaa4a251de9c427d874458af321b65b0c84c2922da84" Oct 02 02:05:30 crc kubenswrapper[4885]: I1002 02:05:30.794421 4885 scope.go:117] "RemoveContainer" containerID="8dc74af4b51491aba9029712d26b660cffe33813b769beea80ccb9b654f48b56" Oct 02 02:05:30 crc kubenswrapper[4885]: I1002 02:05:30.813310 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-h9tlf"] Oct 02 02:05:30 crc kubenswrapper[4885]: I1002 02:05:30.818888 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-h9tlf"] Oct 02 02:05:30 crc kubenswrapper[4885]: I1002 02:05:30.827409 4885 scope.go:117] "RemoveContainer" containerID="712c4978547cdf1d6a36eaa4a251de9c427d874458af321b65b0c84c2922da84" Oct 02 02:05:30 crc kubenswrapper[4885]: E1002 02:05:30.828891 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"712c4978547cdf1d6a36eaa4a251de9c427d874458af321b65b0c84c2922da84\": container with ID starting with 712c4978547cdf1d6a36eaa4a251de9c427d874458af321b65b0c84c2922da84 not found: ID does not exist" containerID="712c4978547cdf1d6a36eaa4a251de9c427d874458af321b65b0c84c2922da84" Oct 02 02:05:30 crc kubenswrapper[4885]: I1002 02:05:30.829068 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"712c4978547cdf1d6a36eaa4a251de9c427d874458af321b65b0c84c2922da84"} err="failed to get container status \"712c4978547cdf1d6a36eaa4a251de9c427d874458af321b65b0c84c2922da84\": rpc error: code = NotFound desc = could not find container \"712c4978547cdf1d6a36eaa4a251de9c427d874458af321b65b0c84c2922da84\": container with ID starting with 712c4978547cdf1d6a36eaa4a251de9c427d874458af321b65b0c84c2922da84 not found: ID does not exist" Oct 02 02:05:30 crc kubenswrapper[4885]: I1002 02:05:30.829151 4885 scope.go:117] "RemoveContainer" containerID="8dc74af4b51491aba9029712d26b660cffe33813b769beea80ccb9b654f48b56" Oct 02 02:05:30 crc kubenswrapper[4885]: E1002 02:05:30.829491 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8dc74af4b51491aba9029712d26b660cffe33813b769beea80ccb9b654f48b56\": container with ID starting with 8dc74af4b51491aba9029712d26b660cffe33813b769beea80ccb9b654f48b56 not found: ID does not exist" containerID="8dc74af4b51491aba9029712d26b660cffe33813b769beea80ccb9b654f48b56" Oct 02 02:05:30 crc kubenswrapper[4885]: I1002 02:05:30.829532 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8dc74af4b51491aba9029712d26b660cffe33813b769beea80ccb9b654f48b56"} err="failed to get container status \"8dc74af4b51491aba9029712d26b660cffe33813b769beea80ccb9b654f48b56\": rpc error: code = NotFound desc = could not find container \"8dc74af4b51491aba9029712d26b660cffe33813b769beea80ccb9b654f48b56\": container with ID starting with 8dc74af4b51491aba9029712d26b660cffe33813b769beea80ccb9b654f48b56 not found: ID does not exist" Oct 02 02:05:31 crc kubenswrapper[4885]: I1002 02:05:31.626285 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6b5d8cdf54-gr28p" Oct 02 02:05:31 crc kubenswrapper[4885]: I1002 02:05:31.660314 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6b5d8cdf54-gr28p" Oct 02 02:05:31 crc kubenswrapper[4885]: I1002 02:05:31.745617 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-6c87b5899d-lvdbp"] Oct 02 02:05:31 crc kubenswrapper[4885]: I1002 02:05:31.745861 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-6c87b5899d-lvdbp" podUID="bb9e27f4-9f80-40ff-af3e-fe5aca9f5755" containerName="barbican-api-log" containerID="cri-o://6126dae4aa6846b8af80f7984b55f1fce7b7ce164a49363c7ae5e9d582b813a2" gracePeriod=30 Oct 02 02:05:31 crc kubenswrapper[4885]: I1002 02:05:31.746010 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-6c87b5899d-lvdbp" podUID="bb9e27f4-9f80-40ff-af3e-fe5aca9f5755" containerName="barbican-api" containerID="cri-o://2a03a6f00575b3ffc6251f3cfb33c8f81551cffb35a3821957a66e0609a9d69d" gracePeriod=30 Oct 02 02:05:31 crc kubenswrapper[4885]: I1002 02:05:31.787415 4885 generic.go:334] "Generic (PLEG): container finished" podID="72053d12-6b1a-4df1-9fbe-ba2973d53dc3" containerID="fa295d66b1785b87309be69f6e91a2a2adb0fd6b866e9c326e188f45b8d98337" exitCode=0 Oct 02 02:05:31 crc kubenswrapper[4885]: I1002 02:05:31.787476 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"72053d12-6b1a-4df1-9fbe-ba2973d53dc3","Type":"ContainerDied","Data":"fa295d66b1785b87309be69f6e91a2a2adb0fd6b866e9c326e188f45b8d98337"} Oct 02 02:05:32 crc kubenswrapper[4885]: I1002 02:05:32.058892 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="00531889-6c5d-444b-bfeb-4bb628d7ef94" path="/var/lib/kubelet/pods/00531889-6c5d-444b-bfeb-4bb628d7ef94/volumes" Oct 02 02:05:32 crc kubenswrapper[4885]: I1002 02:05:32.657401 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 02 02:05:32 crc kubenswrapper[4885]: I1002 02:05:32.718385 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72053d12-6b1a-4df1-9fbe-ba2973d53dc3-config-data\") pod \"72053d12-6b1a-4df1-9fbe-ba2973d53dc3\" (UID: \"72053d12-6b1a-4df1-9fbe-ba2973d53dc3\") " Oct 02 02:05:32 crc kubenswrapper[4885]: I1002 02:05:32.718490 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/72053d12-6b1a-4df1-9fbe-ba2973d53dc3-config-data-custom\") pod \"72053d12-6b1a-4df1-9fbe-ba2973d53dc3\" (UID: \"72053d12-6b1a-4df1-9fbe-ba2973d53dc3\") " Oct 02 02:05:32 crc kubenswrapper[4885]: I1002 02:05:32.718549 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/72053d12-6b1a-4df1-9fbe-ba2973d53dc3-scripts\") pod \"72053d12-6b1a-4df1-9fbe-ba2973d53dc3\" (UID: \"72053d12-6b1a-4df1-9fbe-ba2973d53dc3\") " Oct 02 02:05:32 crc kubenswrapper[4885]: I1002 02:05:32.718589 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wk4tq\" (UniqueName: \"kubernetes.io/projected/72053d12-6b1a-4df1-9fbe-ba2973d53dc3-kube-api-access-wk4tq\") pod \"72053d12-6b1a-4df1-9fbe-ba2973d53dc3\" (UID: \"72053d12-6b1a-4df1-9fbe-ba2973d53dc3\") " Oct 02 02:05:32 crc kubenswrapper[4885]: I1002 02:05:32.718651 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72053d12-6b1a-4df1-9fbe-ba2973d53dc3-combined-ca-bundle\") pod \"72053d12-6b1a-4df1-9fbe-ba2973d53dc3\" (UID: \"72053d12-6b1a-4df1-9fbe-ba2973d53dc3\") " Oct 02 02:05:32 crc kubenswrapper[4885]: I1002 02:05:32.718729 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/72053d12-6b1a-4df1-9fbe-ba2973d53dc3-etc-machine-id\") pod \"72053d12-6b1a-4df1-9fbe-ba2973d53dc3\" (UID: \"72053d12-6b1a-4df1-9fbe-ba2973d53dc3\") " Oct 02 02:05:32 crc kubenswrapper[4885]: I1002 02:05:32.719242 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/72053d12-6b1a-4df1-9fbe-ba2973d53dc3-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "72053d12-6b1a-4df1-9fbe-ba2973d53dc3" (UID: "72053d12-6b1a-4df1-9fbe-ba2973d53dc3"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 02:05:32 crc kubenswrapper[4885]: I1002 02:05:32.725430 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72053d12-6b1a-4df1-9fbe-ba2973d53dc3-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "72053d12-6b1a-4df1-9fbe-ba2973d53dc3" (UID: "72053d12-6b1a-4df1-9fbe-ba2973d53dc3"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:05:32 crc kubenswrapper[4885]: I1002 02:05:32.727943 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/72053d12-6b1a-4df1-9fbe-ba2973d53dc3-kube-api-access-wk4tq" (OuterVolumeSpecName: "kube-api-access-wk4tq") pod "72053d12-6b1a-4df1-9fbe-ba2973d53dc3" (UID: "72053d12-6b1a-4df1-9fbe-ba2973d53dc3"). InnerVolumeSpecName "kube-api-access-wk4tq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:05:32 crc kubenswrapper[4885]: I1002 02:05:32.734437 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72053d12-6b1a-4df1-9fbe-ba2973d53dc3-scripts" (OuterVolumeSpecName: "scripts") pod "72053d12-6b1a-4df1-9fbe-ba2973d53dc3" (UID: "72053d12-6b1a-4df1-9fbe-ba2973d53dc3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:05:32 crc kubenswrapper[4885]: I1002 02:05:32.771787 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72053d12-6b1a-4df1-9fbe-ba2973d53dc3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "72053d12-6b1a-4df1-9fbe-ba2973d53dc3" (UID: "72053d12-6b1a-4df1-9fbe-ba2973d53dc3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:05:32 crc kubenswrapper[4885]: I1002 02:05:32.814564 4885 generic.go:334] "Generic (PLEG): container finished" podID="bb9e27f4-9f80-40ff-af3e-fe5aca9f5755" containerID="6126dae4aa6846b8af80f7984b55f1fce7b7ce164a49363c7ae5e9d582b813a2" exitCode=143 Oct 02 02:05:32 crc kubenswrapper[4885]: I1002 02:05:32.814622 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6c87b5899d-lvdbp" event={"ID":"bb9e27f4-9f80-40ff-af3e-fe5aca9f5755","Type":"ContainerDied","Data":"6126dae4aa6846b8af80f7984b55f1fce7b7ce164a49363c7ae5e9d582b813a2"} Oct 02 02:05:32 crc kubenswrapper[4885]: I1002 02:05:32.815239 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72053d12-6b1a-4df1-9fbe-ba2973d53dc3-config-data" (OuterVolumeSpecName: "config-data") pod "72053d12-6b1a-4df1-9fbe-ba2973d53dc3" (UID: "72053d12-6b1a-4df1-9fbe-ba2973d53dc3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:05:32 crc kubenswrapper[4885]: I1002 02:05:32.817427 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"097aaa2c-cd9c-4a5c-b04b-110fac788275","Type":"ContainerStarted","Data":"ede4bb977749881ab907bad8372399181c033936d9b20b8bc307e6f7f8fff7e5"} Oct 02 02:05:32 crc kubenswrapper[4885]: I1002 02:05:32.817733 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 02 02:05:32 crc kubenswrapper[4885]: I1002 02:05:32.819567 4885 generic.go:334] "Generic (PLEG): container finished" podID="72053d12-6b1a-4df1-9fbe-ba2973d53dc3" containerID="1bab2764732e628230707bff05f072c449d8f783b77372dfedcf2d5c5acb4e10" exitCode=0 Oct 02 02:05:32 crc kubenswrapper[4885]: I1002 02:05:32.819599 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"72053d12-6b1a-4df1-9fbe-ba2973d53dc3","Type":"ContainerDied","Data":"1bab2764732e628230707bff05f072c449d8f783b77372dfedcf2d5c5acb4e10"} Oct 02 02:05:32 crc kubenswrapper[4885]: I1002 02:05:32.819616 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"72053d12-6b1a-4df1-9fbe-ba2973d53dc3","Type":"ContainerDied","Data":"978695589c6389422d521cee5fa7ae0b367fda9b2ccfb5c01a5c38aeebd5308c"} Oct 02 02:05:32 crc kubenswrapper[4885]: I1002 02:05:32.819631 4885 scope.go:117] "RemoveContainer" containerID="fa295d66b1785b87309be69f6e91a2a2adb0fd6b866e9c326e188f45b8d98337" Oct 02 02:05:32 crc kubenswrapper[4885]: I1002 02:05:32.819699 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 02 02:05:32 crc kubenswrapper[4885]: I1002 02:05:32.821361 4885 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72053d12-6b1a-4df1-9fbe-ba2973d53dc3-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:32 crc kubenswrapper[4885]: I1002 02:05:32.821388 4885 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/72053d12-6b1a-4df1-9fbe-ba2973d53dc3-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:32 crc kubenswrapper[4885]: I1002 02:05:32.821399 4885 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/72053d12-6b1a-4df1-9fbe-ba2973d53dc3-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:32 crc kubenswrapper[4885]: I1002 02:05:32.821408 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wk4tq\" (UniqueName: \"kubernetes.io/projected/72053d12-6b1a-4df1-9fbe-ba2973d53dc3-kube-api-access-wk4tq\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:32 crc kubenswrapper[4885]: I1002 02:05:32.821417 4885 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72053d12-6b1a-4df1-9fbe-ba2973d53dc3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:32 crc kubenswrapper[4885]: I1002 02:05:32.821425 4885 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/72053d12-6b1a-4df1-9fbe-ba2973d53dc3-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:32 crc kubenswrapper[4885]: I1002 02:05:32.843473 4885 scope.go:117] "RemoveContainer" containerID="1bab2764732e628230707bff05f072c449d8f783b77372dfedcf2d5c5acb4e10" Oct 02 02:05:32 crc kubenswrapper[4885]: I1002 02:05:32.843681 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.025902478 podStartE2EDuration="5.843655691s" podCreationTimestamp="2025-10-02 02:05:27 +0000 UTC" firstStartedPulling="2025-10-02 02:05:28.655094145 +0000 UTC m=+1117.466841544" lastFinishedPulling="2025-10-02 02:05:32.472847358 +0000 UTC m=+1121.284594757" observedRunningTime="2025-10-02 02:05:32.836658288 +0000 UTC m=+1121.648405687" watchObservedRunningTime="2025-10-02 02:05:32.843655691 +0000 UTC m=+1121.655403100" Oct 02 02:05:32 crc kubenswrapper[4885]: I1002 02:05:32.864586 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 02:05:32 crc kubenswrapper[4885]: I1002 02:05:32.871469 4885 scope.go:117] "RemoveContainer" containerID="fa295d66b1785b87309be69f6e91a2a2adb0fd6b866e9c326e188f45b8d98337" Oct 02 02:05:32 crc kubenswrapper[4885]: E1002 02:05:32.875366 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fa295d66b1785b87309be69f6e91a2a2adb0fd6b866e9c326e188f45b8d98337\": container with ID starting with fa295d66b1785b87309be69f6e91a2a2adb0fd6b866e9c326e188f45b8d98337 not found: ID does not exist" containerID="fa295d66b1785b87309be69f6e91a2a2adb0fd6b866e9c326e188f45b8d98337" Oct 02 02:05:32 crc kubenswrapper[4885]: I1002 02:05:32.875536 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fa295d66b1785b87309be69f6e91a2a2adb0fd6b866e9c326e188f45b8d98337"} err="failed to get container status \"fa295d66b1785b87309be69f6e91a2a2adb0fd6b866e9c326e188f45b8d98337\": rpc error: code = NotFound desc = could not find container \"fa295d66b1785b87309be69f6e91a2a2adb0fd6b866e9c326e188f45b8d98337\": container with ID starting with fa295d66b1785b87309be69f6e91a2a2adb0fd6b866e9c326e188f45b8d98337 not found: ID does not exist" Oct 02 02:05:32 crc kubenswrapper[4885]: I1002 02:05:32.875613 4885 scope.go:117] "RemoveContainer" containerID="1bab2764732e628230707bff05f072c449d8f783b77372dfedcf2d5c5acb4e10" Oct 02 02:05:32 crc kubenswrapper[4885]: I1002 02:05:32.875797 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 02:05:32 crc kubenswrapper[4885]: E1002 02:05:32.878030 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1bab2764732e628230707bff05f072c449d8f783b77372dfedcf2d5c5acb4e10\": container with ID starting with 1bab2764732e628230707bff05f072c449d8f783b77372dfedcf2d5c5acb4e10 not found: ID does not exist" containerID="1bab2764732e628230707bff05f072c449d8f783b77372dfedcf2d5c5acb4e10" Oct 02 02:05:32 crc kubenswrapper[4885]: I1002 02:05:32.878140 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1bab2764732e628230707bff05f072c449d8f783b77372dfedcf2d5c5acb4e10"} err="failed to get container status \"1bab2764732e628230707bff05f072c449d8f783b77372dfedcf2d5c5acb4e10\": rpc error: code = NotFound desc = could not find container \"1bab2764732e628230707bff05f072c449d8f783b77372dfedcf2d5c5acb4e10\": container with ID starting with 1bab2764732e628230707bff05f072c449d8f783b77372dfedcf2d5c5acb4e10 not found: ID does not exist" Oct 02 02:05:32 crc kubenswrapper[4885]: I1002 02:05:32.890104 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 02:05:32 crc kubenswrapper[4885]: E1002 02:05:32.890904 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00531889-6c5d-444b-bfeb-4bb628d7ef94" containerName="init" Oct 02 02:05:32 crc kubenswrapper[4885]: I1002 02:05:32.890921 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="00531889-6c5d-444b-bfeb-4bb628d7ef94" containerName="init" Oct 02 02:05:32 crc kubenswrapper[4885]: E1002 02:05:32.890963 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00531889-6c5d-444b-bfeb-4bb628d7ef94" containerName="dnsmasq-dns" Oct 02 02:05:32 crc kubenswrapper[4885]: I1002 02:05:32.890971 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="00531889-6c5d-444b-bfeb-4bb628d7ef94" containerName="dnsmasq-dns" Oct 02 02:05:32 crc kubenswrapper[4885]: E1002 02:05:32.890990 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72053d12-6b1a-4df1-9fbe-ba2973d53dc3" containerName="probe" Oct 02 02:05:32 crc kubenswrapper[4885]: I1002 02:05:32.890997 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="72053d12-6b1a-4df1-9fbe-ba2973d53dc3" containerName="probe" Oct 02 02:05:32 crc kubenswrapper[4885]: E1002 02:05:32.891012 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72053d12-6b1a-4df1-9fbe-ba2973d53dc3" containerName="cinder-scheduler" Oct 02 02:05:32 crc kubenswrapper[4885]: I1002 02:05:32.891018 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="72053d12-6b1a-4df1-9fbe-ba2973d53dc3" containerName="cinder-scheduler" Oct 02 02:05:32 crc kubenswrapper[4885]: I1002 02:05:32.891688 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="72053d12-6b1a-4df1-9fbe-ba2973d53dc3" containerName="probe" Oct 02 02:05:32 crc kubenswrapper[4885]: I1002 02:05:32.891715 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="72053d12-6b1a-4df1-9fbe-ba2973d53dc3" containerName="cinder-scheduler" Oct 02 02:05:32 crc kubenswrapper[4885]: I1002 02:05:32.891756 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="00531889-6c5d-444b-bfeb-4bb628d7ef94" containerName="dnsmasq-dns" Oct 02 02:05:32 crc kubenswrapper[4885]: I1002 02:05:32.908047 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 02 02:05:32 crc kubenswrapper[4885]: I1002 02:05:32.913410 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 02 02:05:32 crc kubenswrapper[4885]: I1002 02:05:32.926466 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 02:05:32 crc kubenswrapper[4885]: I1002 02:05:32.946394 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 02 02:05:32 crc kubenswrapper[4885]: I1002 02:05:32.946443 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 02 02:05:32 crc kubenswrapper[4885]: I1002 02:05:32.975771 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 02 02:05:32 crc kubenswrapper[4885]: I1002 02:05:32.990548 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 02 02:05:33 crc kubenswrapper[4885]: I1002 02:05:33.024764 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0609d820-e495-4845-a50f-166717c55f93-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"0609d820-e495-4845-a50f-166717c55f93\") " pod="openstack/cinder-scheduler-0" Oct 02 02:05:33 crc kubenswrapper[4885]: I1002 02:05:33.025059 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0609d820-e495-4845-a50f-166717c55f93-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"0609d820-e495-4845-a50f-166717c55f93\") " pod="openstack/cinder-scheduler-0" Oct 02 02:05:33 crc kubenswrapper[4885]: I1002 02:05:33.025164 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0609d820-e495-4845-a50f-166717c55f93-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"0609d820-e495-4845-a50f-166717c55f93\") " pod="openstack/cinder-scheduler-0" Oct 02 02:05:33 crc kubenswrapper[4885]: I1002 02:05:33.025293 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0609d820-e495-4845-a50f-166717c55f93-scripts\") pod \"cinder-scheduler-0\" (UID: \"0609d820-e495-4845-a50f-166717c55f93\") " pod="openstack/cinder-scheduler-0" Oct 02 02:05:33 crc kubenswrapper[4885]: I1002 02:05:33.025385 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0609d820-e495-4845-a50f-166717c55f93-config-data\") pod \"cinder-scheduler-0\" (UID: \"0609d820-e495-4845-a50f-166717c55f93\") " pod="openstack/cinder-scheduler-0" Oct 02 02:05:33 crc kubenswrapper[4885]: I1002 02:05:33.025971 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6l56v\" (UniqueName: \"kubernetes.io/projected/0609d820-e495-4845-a50f-166717c55f93-kube-api-access-6l56v\") pod \"cinder-scheduler-0\" (UID: \"0609d820-e495-4845-a50f-166717c55f93\") " pod="openstack/cinder-scheduler-0" Oct 02 02:05:33 crc kubenswrapper[4885]: I1002 02:05:33.126978 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6l56v\" (UniqueName: \"kubernetes.io/projected/0609d820-e495-4845-a50f-166717c55f93-kube-api-access-6l56v\") pod \"cinder-scheduler-0\" (UID: \"0609d820-e495-4845-a50f-166717c55f93\") " pod="openstack/cinder-scheduler-0" Oct 02 02:05:33 crc kubenswrapper[4885]: I1002 02:05:33.127302 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0609d820-e495-4845-a50f-166717c55f93-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"0609d820-e495-4845-a50f-166717c55f93\") " pod="openstack/cinder-scheduler-0" Oct 02 02:05:33 crc kubenswrapper[4885]: I1002 02:05:33.127328 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0609d820-e495-4845-a50f-166717c55f93-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"0609d820-e495-4845-a50f-166717c55f93\") " pod="openstack/cinder-scheduler-0" Oct 02 02:05:33 crc kubenswrapper[4885]: I1002 02:05:33.127368 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0609d820-e495-4845-a50f-166717c55f93-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"0609d820-e495-4845-a50f-166717c55f93\") " pod="openstack/cinder-scheduler-0" Oct 02 02:05:33 crc kubenswrapper[4885]: I1002 02:05:33.127428 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0609d820-e495-4845-a50f-166717c55f93-scripts\") pod \"cinder-scheduler-0\" (UID: \"0609d820-e495-4845-a50f-166717c55f93\") " pod="openstack/cinder-scheduler-0" Oct 02 02:05:33 crc kubenswrapper[4885]: I1002 02:05:33.127453 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0609d820-e495-4845-a50f-166717c55f93-config-data\") pod \"cinder-scheduler-0\" (UID: \"0609d820-e495-4845-a50f-166717c55f93\") " pod="openstack/cinder-scheduler-0" Oct 02 02:05:33 crc kubenswrapper[4885]: I1002 02:05:33.127666 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0609d820-e495-4845-a50f-166717c55f93-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"0609d820-e495-4845-a50f-166717c55f93\") " pod="openstack/cinder-scheduler-0" Oct 02 02:05:33 crc kubenswrapper[4885]: I1002 02:05:33.132017 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0609d820-e495-4845-a50f-166717c55f93-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"0609d820-e495-4845-a50f-166717c55f93\") " pod="openstack/cinder-scheduler-0" Oct 02 02:05:33 crc kubenswrapper[4885]: I1002 02:05:33.132610 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0609d820-e495-4845-a50f-166717c55f93-scripts\") pod \"cinder-scheduler-0\" (UID: \"0609d820-e495-4845-a50f-166717c55f93\") " pod="openstack/cinder-scheduler-0" Oct 02 02:05:33 crc kubenswrapper[4885]: I1002 02:05:33.133127 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0609d820-e495-4845-a50f-166717c55f93-config-data\") pod \"cinder-scheduler-0\" (UID: \"0609d820-e495-4845-a50f-166717c55f93\") " pod="openstack/cinder-scheduler-0" Oct 02 02:05:33 crc kubenswrapper[4885]: I1002 02:05:33.140749 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0609d820-e495-4845-a50f-166717c55f93-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"0609d820-e495-4845-a50f-166717c55f93\") " pod="openstack/cinder-scheduler-0" Oct 02 02:05:33 crc kubenswrapper[4885]: I1002 02:05:33.147832 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6l56v\" (UniqueName: \"kubernetes.io/projected/0609d820-e495-4845-a50f-166717c55f93-kube-api-access-6l56v\") pod \"cinder-scheduler-0\" (UID: \"0609d820-e495-4845-a50f-166717c55f93\") " pod="openstack/cinder-scheduler-0" Oct 02 02:05:33 crc kubenswrapper[4885]: I1002 02:05:33.230052 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 02 02:05:33 crc kubenswrapper[4885]: I1002 02:05:33.678845 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 02:05:33 crc kubenswrapper[4885]: I1002 02:05:33.831078 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"0609d820-e495-4845-a50f-166717c55f93","Type":"ContainerStarted","Data":"277538b17907688bd40679f9f1a6b88001f5f6066612aed02576da0bdbc30d6c"} Oct 02 02:05:33 crc kubenswrapper[4885]: I1002 02:05:33.833406 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 02 02:05:33 crc kubenswrapper[4885]: I1002 02:05:33.833524 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 02 02:05:33 crc kubenswrapper[4885]: I1002 02:05:33.970631 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-5d65d69454-8bzbn" Oct 02 02:05:34 crc kubenswrapper[4885]: I1002 02:05:34.081901 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="72053d12-6b1a-4df1-9fbe-ba2973d53dc3" path="/var/lib/kubelet/pods/72053d12-6b1a-4df1-9fbe-ba2973d53dc3/volumes" Oct 02 02:05:34 crc kubenswrapper[4885]: I1002 02:05:34.082605 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 02 02:05:34 crc kubenswrapper[4885]: I1002 02:05:34.082632 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 02 02:05:34 crc kubenswrapper[4885]: I1002 02:05:34.107565 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 02 02:05:34 crc kubenswrapper[4885]: I1002 02:05:34.125410 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 02 02:05:34 crc kubenswrapper[4885]: I1002 02:05:34.843744 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"0609d820-e495-4845-a50f-166717c55f93","Type":"ContainerStarted","Data":"16d7c6ad98e6aa969f24a2a00854d22e0f9b1da881a148951ec014bba0cc4128"} Oct 02 02:05:34 crc kubenswrapper[4885]: I1002 02:05:34.844769 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 02 02:05:34 crc kubenswrapper[4885]: I1002 02:05:34.844788 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 02 02:05:34 crc kubenswrapper[4885]: I1002 02:05:34.959974 4885 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6c87b5899d-lvdbp" podUID="bb9e27f4-9f80-40ff-af3e-fe5aca9f5755" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.166:9311/healthcheck\": read tcp 10.217.0.2:43196->10.217.0.166:9311: read: connection reset by peer" Oct 02 02:05:34 crc kubenswrapper[4885]: I1002 02:05:34.959974 4885 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6c87b5899d-lvdbp" podUID="bb9e27f4-9f80-40ff-af3e-fe5aca9f5755" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.166:9311/healthcheck\": read tcp 10.217.0.2:43180->10.217.0.166:9311: read: connection reset by peer" Oct 02 02:05:35 crc kubenswrapper[4885]: I1002 02:05:35.479545 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6c87b5899d-lvdbp" Oct 02 02:05:35 crc kubenswrapper[4885]: I1002 02:05:35.579875 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb9e27f4-9f80-40ff-af3e-fe5aca9f5755-config-data\") pod \"bb9e27f4-9f80-40ff-af3e-fe5aca9f5755\" (UID: \"bb9e27f4-9f80-40ff-af3e-fe5aca9f5755\") " Oct 02 02:05:35 crc kubenswrapper[4885]: I1002 02:05:35.580136 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bb9e27f4-9f80-40ff-af3e-fe5aca9f5755-logs\") pod \"bb9e27f4-9f80-40ff-af3e-fe5aca9f5755\" (UID: \"bb9e27f4-9f80-40ff-af3e-fe5aca9f5755\") " Oct 02 02:05:35 crc kubenswrapper[4885]: I1002 02:05:35.580157 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bb9e27f4-9f80-40ff-af3e-fe5aca9f5755-config-data-custom\") pod \"bb9e27f4-9f80-40ff-af3e-fe5aca9f5755\" (UID: \"bb9e27f4-9f80-40ff-af3e-fe5aca9f5755\") " Oct 02 02:05:35 crc kubenswrapper[4885]: I1002 02:05:35.580206 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-72p9b\" (UniqueName: \"kubernetes.io/projected/bb9e27f4-9f80-40ff-af3e-fe5aca9f5755-kube-api-access-72p9b\") pod \"bb9e27f4-9f80-40ff-af3e-fe5aca9f5755\" (UID: \"bb9e27f4-9f80-40ff-af3e-fe5aca9f5755\") " Oct 02 02:05:35 crc kubenswrapper[4885]: I1002 02:05:35.580288 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb9e27f4-9f80-40ff-af3e-fe5aca9f5755-combined-ca-bundle\") pod \"bb9e27f4-9f80-40ff-af3e-fe5aca9f5755\" (UID: \"bb9e27f4-9f80-40ff-af3e-fe5aca9f5755\") " Oct 02 02:05:35 crc kubenswrapper[4885]: I1002 02:05:35.581697 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bb9e27f4-9f80-40ff-af3e-fe5aca9f5755-logs" (OuterVolumeSpecName: "logs") pod "bb9e27f4-9f80-40ff-af3e-fe5aca9f5755" (UID: "bb9e27f4-9f80-40ff-af3e-fe5aca9f5755"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:05:35 crc kubenswrapper[4885]: I1002 02:05:35.606440 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb9e27f4-9f80-40ff-af3e-fe5aca9f5755-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "bb9e27f4-9f80-40ff-af3e-fe5aca9f5755" (UID: "bb9e27f4-9f80-40ff-af3e-fe5aca9f5755"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:05:35 crc kubenswrapper[4885]: I1002 02:05:35.619280 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb9e27f4-9f80-40ff-af3e-fe5aca9f5755-kube-api-access-72p9b" (OuterVolumeSpecName: "kube-api-access-72p9b") pod "bb9e27f4-9f80-40ff-af3e-fe5aca9f5755" (UID: "bb9e27f4-9f80-40ff-af3e-fe5aca9f5755"). InnerVolumeSpecName "kube-api-access-72p9b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:05:35 crc kubenswrapper[4885]: I1002 02:05:35.628496 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb9e27f4-9f80-40ff-af3e-fe5aca9f5755-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bb9e27f4-9f80-40ff-af3e-fe5aca9f5755" (UID: "bb9e27f4-9f80-40ff-af3e-fe5aca9f5755"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:05:35 crc kubenswrapper[4885]: I1002 02:05:35.652484 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb9e27f4-9f80-40ff-af3e-fe5aca9f5755-config-data" (OuterVolumeSpecName: "config-data") pod "bb9e27f4-9f80-40ff-af3e-fe5aca9f5755" (UID: "bb9e27f4-9f80-40ff-af3e-fe5aca9f5755"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:05:35 crc kubenswrapper[4885]: I1002 02:05:35.682799 4885 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb9e27f4-9f80-40ff-af3e-fe5aca9f5755-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:35 crc kubenswrapper[4885]: I1002 02:05:35.682846 4885 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bb9e27f4-9f80-40ff-af3e-fe5aca9f5755-logs\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:35 crc kubenswrapper[4885]: I1002 02:05:35.682856 4885 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bb9e27f4-9f80-40ff-af3e-fe5aca9f5755-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:35 crc kubenswrapper[4885]: I1002 02:05:35.682867 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-72p9b\" (UniqueName: \"kubernetes.io/projected/bb9e27f4-9f80-40ff-af3e-fe5aca9f5755-kube-api-access-72p9b\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:35 crc kubenswrapper[4885]: I1002 02:05:35.682877 4885 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb9e27f4-9f80-40ff-af3e-fe5aca9f5755-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:35 crc kubenswrapper[4885]: I1002 02:05:35.854425 4885 generic.go:334] "Generic (PLEG): container finished" podID="bb9e27f4-9f80-40ff-af3e-fe5aca9f5755" containerID="2a03a6f00575b3ffc6251f3cfb33c8f81551cffb35a3821957a66e0609a9d69d" exitCode=0 Oct 02 02:05:35 crc kubenswrapper[4885]: I1002 02:05:35.854511 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6c87b5899d-lvdbp" Oct 02 02:05:35 crc kubenswrapper[4885]: I1002 02:05:35.854509 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6c87b5899d-lvdbp" event={"ID":"bb9e27f4-9f80-40ff-af3e-fe5aca9f5755","Type":"ContainerDied","Data":"2a03a6f00575b3ffc6251f3cfb33c8f81551cffb35a3821957a66e0609a9d69d"} Oct 02 02:05:35 crc kubenswrapper[4885]: I1002 02:05:35.854630 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6c87b5899d-lvdbp" event={"ID":"bb9e27f4-9f80-40ff-af3e-fe5aca9f5755","Type":"ContainerDied","Data":"67f3f877dee69316bb6d1e486462adc80083a21a3f72f2b5be1c92c40c746e3c"} Oct 02 02:05:35 crc kubenswrapper[4885]: I1002 02:05:35.854648 4885 scope.go:117] "RemoveContainer" containerID="2a03a6f00575b3ffc6251f3cfb33c8f81551cffb35a3821957a66e0609a9d69d" Oct 02 02:05:35 crc kubenswrapper[4885]: I1002 02:05:35.863609 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"0609d820-e495-4845-a50f-166717c55f93","Type":"ContainerStarted","Data":"fe0216bf1f9dc4c2728ff5c0fc9723d3414e9b4f534c104ba6024f21fdde11e3"} Oct 02 02:05:35 crc kubenswrapper[4885]: I1002 02:05:35.864407 4885 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 02 02:05:35 crc kubenswrapper[4885]: I1002 02:05:35.864437 4885 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 02 02:05:35 crc kubenswrapper[4885]: I1002 02:05:35.880456 4885 scope.go:117] "RemoveContainer" containerID="6126dae4aa6846b8af80f7984b55f1fce7b7ce164a49363c7ae5e9d582b813a2" Oct 02 02:05:35 crc kubenswrapper[4885]: I1002 02:05:35.892194 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.892175189 podStartE2EDuration="3.892175189s" podCreationTimestamp="2025-10-02 02:05:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:05:35.88763642 +0000 UTC m=+1124.699383819" watchObservedRunningTime="2025-10-02 02:05:35.892175189 +0000 UTC m=+1124.703922588" Oct 02 02:05:35 crc kubenswrapper[4885]: I1002 02:05:35.905230 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-6c87b5899d-lvdbp"] Oct 02 02:05:35 crc kubenswrapper[4885]: I1002 02:05:35.908542 4885 scope.go:117] "RemoveContainer" containerID="2a03a6f00575b3ffc6251f3cfb33c8f81551cffb35a3821957a66e0609a9d69d" Oct 02 02:05:35 crc kubenswrapper[4885]: E1002 02:05:35.909143 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2a03a6f00575b3ffc6251f3cfb33c8f81551cffb35a3821957a66e0609a9d69d\": container with ID starting with 2a03a6f00575b3ffc6251f3cfb33c8f81551cffb35a3821957a66e0609a9d69d not found: ID does not exist" containerID="2a03a6f00575b3ffc6251f3cfb33c8f81551cffb35a3821957a66e0609a9d69d" Oct 02 02:05:35 crc kubenswrapper[4885]: I1002 02:05:35.909213 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2a03a6f00575b3ffc6251f3cfb33c8f81551cffb35a3821957a66e0609a9d69d"} err="failed to get container status \"2a03a6f00575b3ffc6251f3cfb33c8f81551cffb35a3821957a66e0609a9d69d\": rpc error: code = NotFound desc = could not find container \"2a03a6f00575b3ffc6251f3cfb33c8f81551cffb35a3821957a66e0609a9d69d\": container with ID starting with 2a03a6f00575b3ffc6251f3cfb33c8f81551cffb35a3821957a66e0609a9d69d not found: ID does not exist" Oct 02 02:05:35 crc kubenswrapper[4885]: I1002 02:05:35.909245 4885 scope.go:117] "RemoveContainer" containerID="6126dae4aa6846b8af80f7984b55f1fce7b7ce164a49363c7ae5e9d582b813a2" Oct 02 02:05:35 crc kubenswrapper[4885]: E1002 02:05:35.909607 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6126dae4aa6846b8af80f7984b55f1fce7b7ce164a49363c7ae5e9d582b813a2\": container with ID starting with 6126dae4aa6846b8af80f7984b55f1fce7b7ce164a49363c7ae5e9d582b813a2 not found: ID does not exist" containerID="6126dae4aa6846b8af80f7984b55f1fce7b7ce164a49363c7ae5e9d582b813a2" Oct 02 02:05:35 crc kubenswrapper[4885]: I1002 02:05:35.909657 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6126dae4aa6846b8af80f7984b55f1fce7b7ce164a49363c7ae5e9d582b813a2"} err="failed to get container status \"6126dae4aa6846b8af80f7984b55f1fce7b7ce164a49363c7ae5e9d582b813a2\": rpc error: code = NotFound desc = could not find container \"6126dae4aa6846b8af80f7984b55f1fce7b7ce164a49363c7ae5e9d582b813a2\": container with ID starting with 6126dae4aa6846b8af80f7984b55f1fce7b7ce164a49363c7ae5e9d582b813a2 not found: ID does not exist" Oct 02 02:05:35 crc kubenswrapper[4885]: I1002 02:05:35.917076 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-6c87b5899d-lvdbp"] Oct 02 02:05:36 crc kubenswrapper[4885]: I1002 02:05:36.057713 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb9e27f4-9f80-40ff-af3e-fe5aca9f5755" path="/var/lib/kubelet/pods/bb9e27f4-9f80-40ff-af3e-fe5aca9f5755/volumes" Oct 02 02:05:36 crc kubenswrapper[4885]: I1002 02:05:36.317410 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 02 02:05:36 crc kubenswrapper[4885]: I1002 02:05:36.320045 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 02 02:05:36 crc kubenswrapper[4885]: I1002 02:05:36.760805 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 02 02:05:36 crc kubenswrapper[4885]: I1002 02:05:36.762657 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 02 02:05:37 crc kubenswrapper[4885]: I1002 02:05:37.052536 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-fjgfv"] Oct 02 02:05:37 crc kubenswrapper[4885]: E1002 02:05:37.052871 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb9e27f4-9f80-40ff-af3e-fe5aca9f5755" containerName="barbican-api" Oct 02 02:05:37 crc kubenswrapper[4885]: I1002 02:05:37.052883 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb9e27f4-9f80-40ff-af3e-fe5aca9f5755" containerName="barbican-api" Oct 02 02:05:37 crc kubenswrapper[4885]: E1002 02:05:37.052914 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb9e27f4-9f80-40ff-af3e-fe5aca9f5755" containerName="barbican-api-log" Oct 02 02:05:37 crc kubenswrapper[4885]: I1002 02:05:37.052920 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb9e27f4-9f80-40ff-af3e-fe5aca9f5755" containerName="barbican-api-log" Oct 02 02:05:37 crc kubenswrapper[4885]: I1002 02:05:37.053089 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb9e27f4-9f80-40ff-af3e-fe5aca9f5755" containerName="barbican-api" Oct 02 02:05:37 crc kubenswrapper[4885]: I1002 02:05:37.053104 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb9e27f4-9f80-40ff-af3e-fe5aca9f5755" containerName="barbican-api-log" Oct 02 02:05:37 crc kubenswrapper[4885]: I1002 02:05:37.053737 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-fjgfv" Oct 02 02:05:37 crc kubenswrapper[4885]: I1002 02:05:37.063334 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-fjgfv"] Oct 02 02:05:37 crc kubenswrapper[4885]: I1002 02:05:37.111544 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pwm7t\" (UniqueName: \"kubernetes.io/projected/925401e1-aa66-44fc-a81b-80d48c105c9e-kube-api-access-pwm7t\") pod \"nova-api-db-create-fjgfv\" (UID: \"925401e1-aa66-44fc-a81b-80d48c105c9e\") " pod="openstack/nova-api-db-create-fjgfv" Oct 02 02:05:37 crc kubenswrapper[4885]: I1002 02:05:37.162203 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-vl726"] Oct 02 02:05:37 crc kubenswrapper[4885]: I1002 02:05:37.163621 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-vl726" Oct 02 02:05:37 crc kubenswrapper[4885]: I1002 02:05:37.168217 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-vl726"] Oct 02 02:05:37 crc kubenswrapper[4885]: I1002 02:05:37.213464 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2tqnl\" (UniqueName: \"kubernetes.io/projected/93487043-76c8-4e9b-a290-8a8411ccda2c-kube-api-access-2tqnl\") pod \"nova-cell0-db-create-vl726\" (UID: \"93487043-76c8-4e9b-a290-8a8411ccda2c\") " pod="openstack/nova-cell0-db-create-vl726" Oct 02 02:05:37 crc kubenswrapper[4885]: I1002 02:05:37.213513 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pwm7t\" (UniqueName: \"kubernetes.io/projected/925401e1-aa66-44fc-a81b-80d48c105c9e-kube-api-access-pwm7t\") pod \"nova-api-db-create-fjgfv\" (UID: \"925401e1-aa66-44fc-a81b-80d48c105c9e\") " pod="openstack/nova-api-db-create-fjgfv" Oct 02 02:05:37 crc kubenswrapper[4885]: I1002 02:05:37.247921 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pwm7t\" (UniqueName: \"kubernetes.io/projected/925401e1-aa66-44fc-a81b-80d48c105c9e-kube-api-access-pwm7t\") pod \"nova-api-db-create-fjgfv\" (UID: \"925401e1-aa66-44fc-a81b-80d48c105c9e\") " pod="openstack/nova-api-db-create-fjgfv" Oct 02 02:05:37 crc kubenswrapper[4885]: I1002 02:05:37.315932 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2tqnl\" (UniqueName: \"kubernetes.io/projected/93487043-76c8-4e9b-a290-8a8411ccda2c-kube-api-access-2tqnl\") pod \"nova-cell0-db-create-vl726\" (UID: \"93487043-76c8-4e9b-a290-8a8411ccda2c\") " pod="openstack/nova-cell0-db-create-vl726" Oct 02 02:05:37 crc kubenswrapper[4885]: I1002 02:05:37.346916 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2tqnl\" (UniqueName: \"kubernetes.io/projected/93487043-76c8-4e9b-a290-8a8411ccda2c-kube-api-access-2tqnl\") pod \"nova-cell0-db-create-vl726\" (UID: \"93487043-76c8-4e9b-a290-8a8411ccda2c\") " pod="openstack/nova-cell0-db-create-vl726" Oct 02 02:05:37 crc kubenswrapper[4885]: I1002 02:05:37.363477 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-2tfgv"] Oct 02 02:05:37 crc kubenswrapper[4885]: I1002 02:05:37.366577 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-2tfgv" Oct 02 02:05:37 crc kubenswrapper[4885]: I1002 02:05:37.372698 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-2tfgv"] Oct 02 02:05:37 crc kubenswrapper[4885]: I1002 02:05:37.372866 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-fjgfv" Oct 02 02:05:37 crc kubenswrapper[4885]: I1002 02:05:37.420285 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tjn2r\" (UniqueName: \"kubernetes.io/projected/9ffa5a6b-478d-4f4d-967d-61b6b79a8097-kube-api-access-tjn2r\") pod \"nova-cell1-db-create-2tfgv\" (UID: \"9ffa5a6b-478d-4f4d-967d-61b6b79a8097\") " pod="openstack/nova-cell1-db-create-2tfgv" Oct 02 02:05:37 crc kubenswrapper[4885]: I1002 02:05:37.480655 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-vl726" Oct 02 02:05:37 crc kubenswrapper[4885]: I1002 02:05:37.523548 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tjn2r\" (UniqueName: \"kubernetes.io/projected/9ffa5a6b-478d-4f4d-967d-61b6b79a8097-kube-api-access-tjn2r\") pod \"nova-cell1-db-create-2tfgv\" (UID: \"9ffa5a6b-478d-4f4d-967d-61b6b79a8097\") " pod="openstack/nova-cell1-db-create-2tfgv" Oct 02 02:05:37 crc kubenswrapper[4885]: I1002 02:05:37.547039 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tjn2r\" (UniqueName: \"kubernetes.io/projected/9ffa5a6b-478d-4f4d-967d-61b6b79a8097-kube-api-access-tjn2r\") pod \"nova-cell1-db-create-2tfgv\" (UID: \"9ffa5a6b-478d-4f4d-967d-61b6b79a8097\") " pod="openstack/nova-cell1-db-create-2tfgv" Oct 02 02:05:37 crc kubenswrapper[4885]: I1002 02:05:37.701646 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-2tfgv" Oct 02 02:05:37 crc kubenswrapper[4885]: I1002 02:05:37.878569 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-fjgfv"] Oct 02 02:05:37 crc kubenswrapper[4885]: W1002 02:05:37.884219 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod925401e1_aa66_44fc_a81b_80d48c105c9e.slice/crio-28fa6f6049b8d79a371ef723fdfb4f98732623522358bc97d19e41021dea10b7 WatchSource:0}: Error finding container 28fa6f6049b8d79a371ef723fdfb4f98732623522358bc97d19e41021dea10b7: Status 404 returned error can't find the container with id 28fa6f6049b8d79a371ef723fdfb4f98732623522358bc97d19e41021dea10b7 Oct 02 02:05:38 crc kubenswrapper[4885]: I1002 02:05:38.012903 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-vl726"] Oct 02 02:05:38 crc kubenswrapper[4885]: I1002 02:05:38.200055 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-2tfgv"] Oct 02 02:05:38 crc kubenswrapper[4885]: W1002 02:05:38.218096 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9ffa5a6b_478d_4f4d_967d_61b6b79a8097.slice/crio-e35d2b0659f76151e18313f87ac4b7b18f1a2ecbbab6eb2bb3e690fad1ab2899 WatchSource:0}: Error finding container e35d2b0659f76151e18313f87ac4b7b18f1a2ecbbab6eb2bb3e690fad1ab2899: Status 404 returned error can't find the container with id e35d2b0659f76151e18313f87ac4b7b18f1a2ecbbab6eb2bb3e690fad1ab2899 Oct 02 02:05:38 crc kubenswrapper[4885]: I1002 02:05:38.231186 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 02 02:05:38 crc kubenswrapper[4885]: I1002 02:05:38.907691 4885 generic.go:334] "Generic (PLEG): container finished" podID="93487043-76c8-4e9b-a290-8a8411ccda2c" containerID="d1c077f945323e970ddc5489d47a203f6ed334052d474e1176c845e8518d0596" exitCode=0 Oct 02 02:05:38 crc kubenswrapper[4885]: I1002 02:05:38.907802 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-vl726" event={"ID":"93487043-76c8-4e9b-a290-8a8411ccda2c","Type":"ContainerDied","Data":"d1c077f945323e970ddc5489d47a203f6ed334052d474e1176c845e8518d0596"} Oct 02 02:05:38 crc kubenswrapper[4885]: I1002 02:05:38.907955 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-vl726" event={"ID":"93487043-76c8-4e9b-a290-8a8411ccda2c","Type":"ContainerStarted","Data":"175eb10b5b398828eeac4011dab5e6afec07784e19e635c486ec27164aa4e32a"} Oct 02 02:05:38 crc kubenswrapper[4885]: I1002 02:05:38.911417 4885 generic.go:334] "Generic (PLEG): container finished" podID="9ffa5a6b-478d-4f4d-967d-61b6b79a8097" containerID="f2a80ac179f355c797c3ac3052f0e7b4127c9bc773ee633cb5fd43ac644ecf3a" exitCode=0 Oct 02 02:05:38 crc kubenswrapper[4885]: I1002 02:05:38.911505 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-2tfgv" event={"ID":"9ffa5a6b-478d-4f4d-967d-61b6b79a8097","Type":"ContainerDied","Data":"f2a80ac179f355c797c3ac3052f0e7b4127c9bc773ee633cb5fd43ac644ecf3a"} Oct 02 02:05:38 crc kubenswrapper[4885]: I1002 02:05:38.911538 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-2tfgv" event={"ID":"9ffa5a6b-478d-4f4d-967d-61b6b79a8097","Type":"ContainerStarted","Data":"e35d2b0659f76151e18313f87ac4b7b18f1a2ecbbab6eb2bb3e690fad1ab2899"} Oct 02 02:05:38 crc kubenswrapper[4885]: I1002 02:05:38.912668 4885 generic.go:334] "Generic (PLEG): container finished" podID="925401e1-aa66-44fc-a81b-80d48c105c9e" containerID="d6c5dcd3f24c7e7a3ba090cbc513323434befe12c944ae98c3d55b394329fd49" exitCode=0 Oct 02 02:05:38 crc kubenswrapper[4885]: I1002 02:05:38.912711 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-fjgfv" event={"ID":"925401e1-aa66-44fc-a81b-80d48c105c9e","Type":"ContainerDied","Data":"d6c5dcd3f24c7e7a3ba090cbc513323434befe12c944ae98c3d55b394329fd49"} Oct 02 02:05:38 crc kubenswrapper[4885]: I1002 02:05:38.912738 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-fjgfv" event={"ID":"925401e1-aa66-44fc-a81b-80d48c105c9e","Type":"ContainerStarted","Data":"28fa6f6049b8d79a371ef723fdfb4f98732623522358bc97d19e41021dea10b7"} Oct 02 02:05:39 crc kubenswrapper[4885]: I1002 02:05:39.360949 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Oct 02 02:05:39 crc kubenswrapper[4885]: I1002 02:05:39.564568 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 02:05:39 crc kubenswrapper[4885]: I1002 02:05:39.565024 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="097aaa2c-cd9c-4a5c-b04b-110fac788275" containerName="ceilometer-central-agent" containerID="cri-o://ba193af1f4d20a5098cd947ade8f513cd3a54b42c1babee5847299d20278bb85" gracePeriod=30 Oct 02 02:05:39 crc kubenswrapper[4885]: I1002 02:05:39.565143 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="097aaa2c-cd9c-4a5c-b04b-110fac788275" containerName="proxy-httpd" containerID="cri-o://ede4bb977749881ab907bad8372399181c033936d9b20b8bc307e6f7f8fff7e5" gracePeriod=30 Oct 02 02:05:39 crc kubenswrapper[4885]: I1002 02:05:39.565088 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="097aaa2c-cd9c-4a5c-b04b-110fac788275" containerName="ceilometer-notification-agent" containerID="cri-o://5c5f0ddd01d35c10dcf352859dd51106a1ac9e6f582a2e00233251cce937ecf8" gracePeriod=30 Oct 02 02:05:39 crc kubenswrapper[4885]: I1002 02:05:39.565073 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="097aaa2c-cd9c-4a5c-b04b-110fac788275" containerName="sg-core" containerID="cri-o://47702dc0fdfc3bfcb1c07064c38f6df754b7f6ca18c464e77a494e84badde2bb" gracePeriod=30 Oct 02 02:05:39 crc kubenswrapper[4885]: I1002 02:05:39.923288 4885 generic.go:334] "Generic (PLEG): container finished" podID="097aaa2c-cd9c-4a5c-b04b-110fac788275" containerID="ede4bb977749881ab907bad8372399181c033936d9b20b8bc307e6f7f8fff7e5" exitCode=0 Oct 02 02:05:39 crc kubenswrapper[4885]: I1002 02:05:39.923525 4885 generic.go:334] "Generic (PLEG): container finished" podID="097aaa2c-cd9c-4a5c-b04b-110fac788275" containerID="47702dc0fdfc3bfcb1c07064c38f6df754b7f6ca18c464e77a494e84badde2bb" exitCode=2 Oct 02 02:05:39 crc kubenswrapper[4885]: I1002 02:05:39.923582 4885 generic.go:334] "Generic (PLEG): container finished" podID="097aaa2c-cd9c-4a5c-b04b-110fac788275" containerID="ba193af1f4d20a5098cd947ade8f513cd3a54b42c1babee5847299d20278bb85" exitCode=0 Oct 02 02:05:39 crc kubenswrapper[4885]: I1002 02:05:39.923472 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"097aaa2c-cd9c-4a5c-b04b-110fac788275","Type":"ContainerDied","Data":"ede4bb977749881ab907bad8372399181c033936d9b20b8bc307e6f7f8fff7e5"} Oct 02 02:05:39 crc kubenswrapper[4885]: I1002 02:05:39.923833 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"097aaa2c-cd9c-4a5c-b04b-110fac788275","Type":"ContainerDied","Data":"47702dc0fdfc3bfcb1c07064c38f6df754b7f6ca18c464e77a494e84badde2bb"} Oct 02 02:05:39 crc kubenswrapper[4885]: I1002 02:05:39.923897 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"097aaa2c-cd9c-4a5c-b04b-110fac788275","Type":"ContainerDied","Data":"ba193af1f4d20a5098cd947ade8f513cd3a54b42c1babee5847299d20278bb85"} Oct 02 02:05:40 crc kubenswrapper[4885]: I1002 02:05:40.292335 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-fjgfv" Oct 02 02:05:40 crc kubenswrapper[4885]: I1002 02:05:40.374247 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pwm7t\" (UniqueName: \"kubernetes.io/projected/925401e1-aa66-44fc-a81b-80d48c105c9e-kube-api-access-pwm7t\") pod \"925401e1-aa66-44fc-a81b-80d48c105c9e\" (UID: \"925401e1-aa66-44fc-a81b-80d48c105c9e\") " Oct 02 02:05:40 crc kubenswrapper[4885]: I1002 02:05:40.379033 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925401e1-aa66-44fc-a81b-80d48c105c9e-kube-api-access-pwm7t" (OuterVolumeSpecName: "kube-api-access-pwm7t") pod "925401e1-aa66-44fc-a81b-80d48c105c9e" (UID: "925401e1-aa66-44fc-a81b-80d48c105c9e"). InnerVolumeSpecName "kube-api-access-pwm7t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:05:40 crc kubenswrapper[4885]: I1002 02:05:40.382385 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-vl726" Oct 02 02:05:40 crc kubenswrapper[4885]: I1002 02:05:40.443214 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-2tfgv" Oct 02 02:05:40 crc kubenswrapper[4885]: I1002 02:05:40.477042 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tjn2r\" (UniqueName: \"kubernetes.io/projected/9ffa5a6b-478d-4f4d-967d-61b6b79a8097-kube-api-access-tjn2r\") pod \"9ffa5a6b-478d-4f4d-967d-61b6b79a8097\" (UID: \"9ffa5a6b-478d-4f4d-967d-61b6b79a8097\") " Oct 02 02:05:40 crc kubenswrapper[4885]: I1002 02:05:40.477216 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2tqnl\" (UniqueName: \"kubernetes.io/projected/93487043-76c8-4e9b-a290-8a8411ccda2c-kube-api-access-2tqnl\") pod \"93487043-76c8-4e9b-a290-8a8411ccda2c\" (UID: \"93487043-76c8-4e9b-a290-8a8411ccda2c\") " Oct 02 02:05:40 crc kubenswrapper[4885]: I1002 02:05:40.477602 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pwm7t\" (UniqueName: \"kubernetes.io/projected/925401e1-aa66-44fc-a81b-80d48c105c9e-kube-api-access-pwm7t\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:40 crc kubenswrapper[4885]: I1002 02:05:40.480483 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93487043-76c8-4e9b-a290-8a8411ccda2c-kube-api-access-2tqnl" (OuterVolumeSpecName: "kube-api-access-2tqnl") pod "93487043-76c8-4e9b-a290-8a8411ccda2c" (UID: "93487043-76c8-4e9b-a290-8a8411ccda2c"). InnerVolumeSpecName "kube-api-access-2tqnl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:05:40 crc kubenswrapper[4885]: I1002 02:05:40.480549 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ffa5a6b-478d-4f4d-967d-61b6b79a8097-kube-api-access-tjn2r" (OuterVolumeSpecName: "kube-api-access-tjn2r") pod "9ffa5a6b-478d-4f4d-967d-61b6b79a8097" (UID: "9ffa5a6b-478d-4f4d-967d-61b6b79a8097"). InnerVolumeSpecName "kube-api-access-tjn2r". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:05:40 crc kubenswrapper[4885]: I1002 02:05:40.579896 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tjn2r\" (UniqueName: \"kubernetes.io/projected/9ffa5a6b-478d-4f4d-967d-61b6b79a8097-kube-api-access-tjn2r\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:40 crc kubenswrapper[4885]: I1002 02:05:40.579938 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2tqnl\" (UniqueName: \"kubernetes.io/projected/93487043-76c8-4e9b-a290-8a8411ccda2c-kube-api-access-2tqnl\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:40 crc kubenswrapper[4885]: I1002 02:05:40.778723 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 02:05:40 crc kubenswrapper[4885]: I1002 02:05:40.883028 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/097aaa2c-cd9c-4a5c-b04b-110fac788275-combined-ca-bundle\") pod \"097aaa2c-cd9c-4a5c-b04b-110fac788275\" (UID: \"097aaa2c-cd9c-4a5c-b04b-110fac788275\") " Oct 02 02:05:40 crc kubenswrapper[4885]: I1002 02:05:40.883090 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/097aaa2c-cd9c-4a5c-b04b-110fac788275-sg-core-conf-yaml\") pod \"097aaa2c-cd9c-4a5c-b04b-110fac788275\" (UID: \"097aaa2c-cd9c-4a5c-b04b-110fac788275\") " Oct 02 02:05:40 crc kubenswrapper[4885]: I1002 02:05:40.883166 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/097aaa2c-cd9c-4a5c-b04b-110fac788275-run-httpd\") pod \"097aaa2c-cd9c-4a5c-b04b-110fac788275\" (UID: \"097aaa2c-cd9c-4a5c-b04b-110fac788275\") " Oct 02 02:05:40 crc kubenswrapper[4885]: I1002 02:05:40.883205 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/097aaa2c-cd9c-4a5c-b04b-110fac788275-scripts\") pod \"097aaa2c-cd9c-4a5c-b04b-110fac788275\" (UID: \"097aaa2c-cd9c-4a5c-b04b-110fac788275\") " Oct 02 02:05:40 crc kubenswrapper[4885]: I1002 02:05:40.883297 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/097aaa2c-cd9c-4a5c-b04b-110fac788275-config-data\") pod \"097aaa2c-cd9c-4a5c-b04b-110fac788275\" (UID: \"097aaa2c-cd9c-4a5c-b04b-110fac788275\") " Oct 02 02:05:40 crc kubenswrapper[4885]: I1002 02:05:40.883357 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cg6n5\" (UniqueName: \"kubernetes.io/projected/097aaa2c-cd9c-4a5c-b04b-110fac788275-kube-api-access-cg6n5\") pod \"097aaa2c-cd9c-4a5c-b04b-110fac788275\" (UID: \"097aaa2c-cd9c-4a5c-b04b-110fac788275\") " Oct 02 02:05:40 crc kubenswrapper[4885]: I1002 02:05:40.883374 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/097aaa2c-cd9c-4a5c-b04b-110fac788275-log-httpd\") pod \"097aaa2c-cd9c-4a5c-b04b-110fac788275\" (UID: \"097aaa2c-cd9c-4a5c-b04b-110fac788275\") " Oct 02 02:05:40 crc kubenswrapper[4885]: I1002 02:05:40.883921 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/097aaa2c-cd9c-4a5c-b04b-110fac788275-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "097aaa2c-cd9c-4a5c-b04b-110fac788275" (UID: "097aaa2c-cd9c-4a5c-b04b-110fac788275"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:05:40 crc kubenswrapper[4885]: I1002 02:05:40.884154 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/097aaa2c-cd9c-4a5c-b04b-110fac788275-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "097aaa2c-cd9c-4a5c-b04b-110fac788275" (UID: "097aaa2c-cd9c-4a5c-b04b-110fac788275"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:05:40 crc kubenswrapper[4885]: I1002 02:05:40.884286 4885 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/097aaa2c-cd9c-4a5c-b04b-110fac788275-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:40 crc kubenswrapper[4885]: I1002 02:05:40.884570 4885 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/097aaa2c-cd9c-4a5c-b04b-110fac788275-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:40 crc kubenswrapper[4885]: I1002 02:05:40.888518 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/097aaa2c-cd9c-4a5c-b04b-110fac788275-scripts" (OuterVolumeSpecName: "scripts") pod "097aaa2c-cd9c-4a5c-b04b-110fac788275" (UID: "097aaa2c-cd9c-4a5c-b04b-110fac788275"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:05:40 crc kubenswrapper[4885]: I1002 02:05:40.890406 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/097aaa2c-cd9c-4a5c-b04b-110fac788275-kube-api-access-cg6n5" (OuterVolumeSpecName: "kube-api-access-cg6n5") pod "097aaa2c-cd9c-4a5c-b04b-110fac788275" (UID: "097aaa2c-cd9c-4a5c-b04b-110fac788275"). InnerVolumeSpecName "kube-api-access-cg6n5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:05:40 crc kubenswrapper[4885]: I1002 02:05:40.919050 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/097aaa2c-cd9c-4a5c-b04b-110fac788275-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "097aaa2c-cd9c-4a5c-b04b-110fac788275" (UID: "097aaa2c-cd9c-4a5c-b04b-110fac788275"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:05:40 crc kubenswrapper[4885]: I1002 02:05:40.932357 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-fjgfv" event={"ID":"925401e1-aa66-44fc-a81b-80d48c105c9e","Type":"ContainerDied","Data":"28fa6f6049b8d79a371ef723fdfb4f98732623522358bc97d19e41021dea10b7"} Oct 02 02:05:40 crc kubenswrapper[4885]: I1002 02:05:40.932389 4885 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="28fa6f6049b8d79a371ef723fdfb4f98732623522358bc97d19e41021dea10b7" Oct 02 02:05:40 crc kubenswrapper[4885]: I1002 02:05:40.932434 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-fjgfv" Oct 02 02:05:40 crc kubenswrapper[4885]: I1002 02:05:40.938392 4885 generic.go:334] "Generic (PLEG): container finished" podID="097aaa2c-cd9c-4a5c-b04b-110fac788275" containerID="5c5f0ddd01d35c10dcf352859dd51106a1ac9e6f582a2e00233251cce937ecf8" exitCode=0 Oct 02 02:05:40 crc kubenswrapper[4885]: I1002 02:05:40.938461 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"097aaa2c-cd9c-4a5c-b04b-110fac788275","Type":"ContainerDied","Data":"5c5f0ddd01d35c10dcf352859dd51106a1ac9e6f582a2e00233251cce937ecf8"} Oct 02 02:05:40 crc kubenswrapper[4885]: I1002 02:05:40.938485 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"097aaa2c-cd9c-4a5c-b04b-110fac788275","Type":"ContainerDied","Data":"3d7b30474d6c2da89c60385c1e31c04baebe3cb12fa7bc8206eef86ec9707b86"} Oct 02 02:05:40 crc kubenswrapper[4885]: I1002 02:05:40.938511 4885 scope.go:117] "RemoveContainer" containerID="ede4bb977749881ab907bad8372399181c033936d9b20b8bc307e6f7f8fff7e5" Oct 02 02:05:40 crc kubenswrapper[4885]: I1002 02:05:40.938761 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 02:05:40 crc kubenswrapper[4885]: I1002 02:05:40.946347 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-vl726" event={"ID":"93487043-76c8-4e9b-a290-8a8411ccda2c","Type":"ContainerDied","Data":"175eb10b5b398828eeac4011dab5e6afec07784e19e635c486ec27164aa4e32a"} Oct 02 02:05:40 crc kubenswrapper[4885]: I1002 02:05:40.946368 4885 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="175eb10b5b398828eeac4011dab5e6afec07784e19e635c486ec27164aa4e32a" Oct 02 02:05:40 crc kubenswrapper[4885]: I1002 02:05:40.946404 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-vl726" Oct 02 02:05:40 crc kubenswrapper[4885]: I1002 02:05:40.948172 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-2tfgv" event={"ID":"9ffa5a6b-478d-4f4d-967d-61b6b79a8097","Type":"ContainerDied","Data":"e35d2b0659f76151e18313f87ac4b7b18f1a2ecbbab6eb2bb3e690fad1ab2899"} Oct 02 02:05:40 crc kubenswrapper[4885]: I1002 02:05:40.948219 4885 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e35d2b0659f76151e18313f87ac4b7b18f1a2ecbbab6eb2bb3e690fad1ab2899" Oct 02 02:05:40 crc kubenswrapper[4885]: I1002 02:05:40.948326 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-2tfgv" Oct 02 02:05:40 crc kubenswrapper[4885]: I1002 02:05:40.958462 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/097aaa2c-cd9c-4a5c-b04b-110fac788275-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "097aaa2c-cd9c-4a5c-b04b-110fac788275" (UID: "097aaa2c-cd9c-4a5c-b04b-110fac788275"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:05:40 crc kubenswrapper[4885]: I1002 02:05:40.986258 4885 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/097aaa2c-cd9c-4a5c-b04b-110fac788275-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:40 crc kubenswrapper[4885]: I1002 02:05:40.986298 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cg6n5\" (UniqueName: \"kubernetes.io/projected/097aaa2c-cd9c-4a5c-b04b-110fac788275-kube-api-access-cg6n5\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:40 crc kubenswrapper[4885]: I1002 02:05:40.986310 4885 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/097aaa2c-cd9c-4a5c-b04b-110fac788275-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:40 crc kubenswrapper[4885]: I1002 02:05:40.986320 4885 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/097aaa2c-cd9c-4a5c-b04b-110fac788275-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:40 crc kubenswrapper[4885]: I1002 02:05:40.989210 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/097aaa2c-cd9c-4a5c-b04b-110fac788275-config-data" (OuterVolumeSpecName: "config-data") pod "097aaa2c-cd9c-4a5c-b04b-110fac788275" (UID: "097aaa2c-cd9c-4a5c-b04b-110fac788275"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:05:41 crc kubenswrapper[4885]: I1002 02:05:41.068222 4885 scope.go:117] "RemoveContainer" containerID="47702dc0fdfc3bfcb1c07064c38f6df754b7f6ca18c464e77a494e84badde2bb" Oct 02 02:05:41 crc kubenswrapper[4885]: I1002 02:05:41.088014 4885 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/097aaa2c-cd9c-4a5c-b04b-110fac788275-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:41 crc kubenswrapper[4885]: I1002 02:05:41.095098 4885 scope.go:117] "RemoveContainer" containerID="5c5f0ddd01d35c10dcf352859dd51106a1ac9e6f582a2e00233251cce937ecf8" Oct 02 02:05:41 crc kubenswrapper[4885]: I1002 02:05:41.122798 4885 scope.go:117] "RemoveContainer" containerID="ba193af1f4d20a5098cd947ade8f513cd3a54b42c1babee5847299d20278bb85" Oct 02 02:05:41 crc kubenswrapper[4885]: I1002 02:05:41.141518 4885 scope.go:117] "RemoveContainer" containerID="ede4bb977749881ab907bad8372399181c033936d9b20b8bc307e6f7f8fff7e5" Oct 02 02:05:41 crc kubenswrapper[4885]: E1002 02:05:41.141998 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ede4bb977749881ab907bad8372399181c033936d9b20b8bc307e6f7f8fff7e5\": container with ID starting with ede4bb977749881ab907bad8372399181c033936d9b20b8bc307e6f7f8fff7e5 not found: ID does not exist" containerID="ede4bb977749881ab907bad8372399181c033936d9b20b8bc307e6f7f8fff7e5" Oct 02 02:05:41 crc kubenswrapper[4885]: I1002 02:05:41.142028 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ede4bb977749881ab907bad8372399181c033936d9b20b8bc307e6f7f8fff7e5"} err="failed to get container status \"ede4bb977749881ab907bad8372399181c033936d9b20b8bc307e6f7f8fff7e5\": rpc error: code = NotFound desc = could not find container \"ede4bb977749881ab907bad8372399181c033936d9b20b8bc307e6f7f8fff7e5\": container with ID starting with ede4bb977749881ab907bad8372399181c033936d9b20b8bc307e6f7f8fff7e5 not found: ID does not exist" Oct 02 02:05:41 crc kubenswrapper[4885]: I1002 02:05:41.142049 4885 scope.go:117] "RemoveContainer" containerID="47702dc0fdfc3bfcb1c07064c38f6df754b7f6ca18c464e77a494e84badde2bb" Oct 02 02:05:41 crc kubenswrapper[4885]: E1002 02:05:41.142259 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"47702dc0fdfc3bfcb1c07064c38f6df754b7f6ca18c464e77a494e84badde2bb\": container with ID starting with 47702dc0fdfc3bfcb1c07064c38f6df754b7f6ca18c464e77a494e84badde2bb not found: ID does not exist" containerID="47702dc0fdfc3bfcb1c07064c38f6df754b7f6ca18c464e77a494e84badde2bb" Oct 02 02:05:41 crc kubenswrapper[4885]: I1002 02:05:41.142293 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"47702dc0fdfc3bfcb1c07064c38f6df754b7f6ca18c464e77a494e84badde2bb"} err="failed to get container status \"47702dc0fdfc3bfcb1c07064c38f6df754b7f6ca18c464e77a494e84badde2bb\": rpc error: code = NotFound desc = could not find container \"47702dc0fdfc3bfcb1c07064c38f6df754b7f6ca18c464e77a494e84badde2bb\": container with ID starting with 47702dc0fdfc3bfcb1c07064c38f6df754b7f6ca18c464e77a494e84badde2bb not found: ID does not exist" Oct 02 02:05:41 crc kubenswrapper[4885]: I1002 02:05:41.142305 4885 scope.go:117] "RemoveContainer" containerID="5c5f0ddd01d35c10dcf352859dd51106a1ac9e6f582a2e00233251cce937ecf8" Oct 02 02:05:41 crc kubenswrapper[4885]: E1002 02:05:41.142537 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5c5f0ddd01d35c10dcf352859dd51106a1ac9e6f582a2e00233251cce937ecf8\": container with ID starting with 5c5f0ddd01d35c10dcf352859dd51106a1ac9e6f582a2e00233251cce937ecf8 not found: ID does not exist" containerID="5c5f0ddd01d35c10dcf352859dd51106a1ac9e6f582a2e00233251cce937ecf8" Oct 02 02:05:41 crc kubenswrapper[4885]: I1002 02:05:41.142555 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c5f0ddd01d35c10dcf352859dd51106a1ac9e6f582a2e00233251cce937ecf8"} err="failed to get container status \"5c5f0ddd01d35c10dcf352859dd51106a1ac9e6f582a2e00233251cce937ecf8\": rpc error: code = NotFound desc = could not find container \"5c5f0ddd01d35c10dcf352859dd51106a1ac9e6f582a2e00233251cce937ecf8\": container with ID starting with 5c5f0ddd01d35c10dcf352859dd51106a1ac9e6f582a2e00233251cce937ecf8 not found: ID does not exist" Oct 02 02:05:41 crc kubenswrapper[4885]: I1002 02:05:41.142566 4885 scope.go:117] "RemoveContainer" containerID="ba193af1f4d20a5098cd947ade8f513cd3a54b42c1babee5847299d20278bb85" Oct 02 02:05:41 crc kubenswrapper[4885]: E1002 02:05:41.142741 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ba193af1f4d20a5098cd947ade8f513cd3a54b42c1babee5847299d20278bb85\": container with ID starting with ba193af1f4d20a5098cd947ade8f513cd3a54b42c1babee5847299d20278bb85 not found: ID does not exist" containerID="ba193af1f4d20a5098cd947ade8f513cd3a54b42c1babee5847299d20278bb85" Oct 02 02:05:41 crc kubenswrapper[4885]: I1002 02:05:41.142764 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba193af1f4d20a5098cd947ade8f513cd3a54b42c1babee5847299d20278bb85"} err="failed to get container status \"ba193af1f4d20a5098cd947ade8f513cd3a54b42c1babee5847299d20278bb85\": rpc error: code = NotFound desc = could not find container \"ba193af1f4d20a5098cd947ade8f513cd3a54b42c1babee5847299d20278bb85\": container with ID starting with ba193af1f4d20a5098cd947ade8f513cd3a54b42c1babee5847299d20278bb85 not found: ID does not exist" Oct 02 02:05:41 crc kubenswrapper[4885]: I1002 02:05:41.278119 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 02:05:41 crc kubenswrapper[4885]: I1002 02:05:41.286838 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 02 02:05:41 crc kubenswrapper[4885]: I1002 02:05:41.307060 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 02 02:05:41 crc kubenswrapper[4885]: E1002 02:05:41.307420 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ffa5a6b-478d-4f4d-967d-61b6b79a8097" containerName="mariadb-database-create" Oct 02 02:05:41 crc kubenswrapper[4885]: I1002 02:05:41.307431 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ffa5a6b-478d-4f4d-967d-61b6b79a8097" containerName="mariadb-database-create" Oct 02 02:05:41 crc kubenswrapper[4885]: E1002 02:05:41.307452 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="925401e1-aa66-44fc-a81b-80d48c105c9e" containerName="mariadb-database-create" Oct 02 02:05:41 crc kubenswrapper[4885]: I1002 02:05:41.307458 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="925401e1-aa66-44fc-a81b-80d48c105c9e" containerName="mariadb-database-create" Oct 02 02:05:41 crc kubenswrapper[4885]: E1002 02:05:41.307467 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93487043-76c8-4e9b-a290-8a8411ccda2c" containerName="mariadb-database-create" Oct 02 02:05:41 crc kubenswrapper[4885]: I1002 02:05:41.307473 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="93487043-76c8-4e9b-a290-8a8411ccda2c" containerName="mariadb-database-create" Oct 02 02:05:41 crc kubenswrapper[4885]: E1002 02:05:41.307491 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="097aaa2c-cd9c-4a5c-b04b-110fac788275" containerName="ceilometer-notification-agent" Oct 02 02:05:41 crc kubenswrapper[4885]: I1002 02:05:41.307497 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="097aaa2c-cd9c-4a5c-b04b-110fac788275" containerName="ceilometer-notification-agent" Oct 02 02:05:41 crc kubenswrapper[4885]: E1002 02:05:41.307511 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="097aaa2c-cd9c-4a5c-b04b-110fac788275" containerName="proxy-httpd" Oct 02 02:05:41 crc kubenswrapper[4885]: I1002 02:05:41.307516 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="097aaa2c-cd9c-4a5c-b04b-110fac788275" containerName="proxy-httpd" Oct 02 02:05:41 crc kubenswrapper[4885]: E1002 02:05:41.307526 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="097aaa2c-cd9c-4a5c-b04b-110fac788275" containerName="ceilometer-central-agent" Oct 02 02:05:41 crc kubenswrapper[4885]: I1002 02:05:41.307532 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="097aaa2c-cd9c-4a5c-b04b-110fac788275" containerName="ceilometer-central-agent" Oct 02 02:05:41 crc kubenswrapper[4885]: E1002 02:05:41.307545 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="097aaa2c-cd9c-4a5c-b04b-110fac788275" containerName="sg-core" Oct 02 02:05:41 crc kubenswrapper[4885]: I1002 02:05:41.307550 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="097aaa2c-cd9c-4a5c-b04b-110fac788275" containerName="sg-core" Oct 02 02:05:41 crc kubenswrapper[4885]: I1002 02:05:41.307702 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="097aaa2c-cd9c-4a5c-b04b-110fac788275" containerName="ceilometer-central-agent" Oct 02 02:05:41 crc kubenswrapper[4885]: I1002 02:05:41.307715 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="925401e1-aa66-44fc-a81b-80d48c105c9e" containerName="mariadb-database-create" Oct 02 02:05:41 crc kubenswrapper[4885]: I1002 02:05:41.307730 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ffa5a6b-478d-4f4d-967d-61b6b79a8097" containerName="mariadb-database-create" Oct 02 02:05:41 crc kubenswrapper[4885]: I1002 02:05:41.307740 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="93487043-76c8-4e9b-a290-8a8411ccda2c" containerName="mariadb-database-create" Oct 02 02:05:41 crc kubenswrapper[4885]: I1002 02:05:41.307747 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="097aaa2c-cd9c-4a5c-b04b-110fac788275" containerName="sg-core" Oct 02 02:05:41 crc kubenswrapper[4885]: I1002 02:05:41.307755 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="097aaa2c-cd9c-4a5c-b04b-110fac788275" containerName="proxy-httpd" Oct 02 02:05:41 crc kubenswrapper[4885]: I1002 02:05:41.307767 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="097aaa2c-cd9c-4a5c-b04b-110fac788275" containerName="ceilometer-notification-agent" Oct 02 02:05:41 crc kubenswrapper[4885]: I1002 02:05:41.309321 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 02:05:41 crc kubenswrapper[4885]: I1002 02:05:41.311717 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 02 02:05:41 crc kubenswrapper[4885]: I1002 02:05:41.311844 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 02 02:05:41 crc kubenswrapper[4885]: I1002 02:05:41.322402 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 02:05:41 crc kubenswrapper[4885]: I1002 02:05:41.392692 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b616d391-356d-4e48-bf47-be58358aa15f-scripts\") pod \"ceilometer-0\" (UID: \"b616d391-356d-4e48-bf47-be58358aa15f\") " pod="openstack/ceilometer-0" Oct 02 02:05:41 crc kubenswrapper[4885]: I1002 02:05:41.392829 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b616d391-356d-4e48-bf47-be58358aa15f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b616d391-356d-4e48-bf47-be58358aa15f\") " pod="openstack/ceilometer-0" Oct 02 02:05:41 crc kubenswrapper[4885]: I1002 02:05:41.392997 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wgbkf\" (UniqueName: \"kubernetes.io/projected/b616d391-356d-4e48-bf47-be58358aa15f-kube-api-access-wgbkf\") pod \"ceilometer-0\" (UID: \"b616d391-356d-4e48-bf47-be58358aa15f\") " pod="openstack/ceilometer-0" Oct 02 02:05:41 crc kubenswrapper[4885]: I1002 02:05:41.393061 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b616d391-356d-4e48-bf47-be58358aa15f-run-httpd\") pod \"ceilometer-0\" (UID: \"b616d391-356d-4e48-bf47-be58358aa15f\") " pod="openstack/ceilometer-0" Oct 02 02:05:41 crc kubenswrapper[4885]: I1002 02:05:41.393232 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b616d391-356d-4e48-bf47-be58358aa15f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b616d391-356d-4e48-bf47-be58358aa15f\") " pod="openstack/ceilometer-0" Oct 02 02:05:41 crc kubenswrapper[4885]: I1002 02:05:41.393355 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b616d391-356d-4e48-bf47-be58358aa15f-log-httpd\") pod \"ceilometer-0\" (UID: \"b616d391-356d-4e48-bf47-be58358aa15f\") " pod="openstack/ceilometer-0" Oct 02 02:05:41 crc kubenswrapper[4885]: I1002 02:05:41.393394 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b616d391-356d-4e48-bf47-be58358aa15f-config-data\") pod \"ceilometer-0\" (UID: \"b616d391-356d-4e48-bf47-be58358aa15f\") " pod="openstack/ceilometer-0" Oct 02 02:05:41 crc kubenswrapper[4885]: I1002 02:05:41.495416 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b616d391-356d-4e48-bf47-be58358aa15f-scripts\") pod \"ceilometer-0\" (UID: \"b616d391-356d-4e48-bf47-be58358aa15f\") " pod="openstack/ceilometer-0" Oct 02 02:05:41 crc kubenswrapper[4885]: I1002 02:05:41.495508 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b616d391-356d-4e48-bf47-be58358aa15f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b616d391-356d-4e48-bf47-be58358aa15f\") " pod="openstack/ceilometer-0" Oct 02 02:05:41 crc kubenswrapper[4885]: I1002 02:05:41.495572 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wgbkf\" (UniqueName: \"kubernetes.io/projected/b616d391-356d-4e48-bf47-be58358aa15f-kube-api-access-wgbkf\") pod \"ceilometer-0\" (UID: \"b616d391-356d-4e48-bf47-be58358aa15f\") " pod="openstack/ceilometer-0" Oct 02 02:05:41 crc kubenswrapper[4885]: I1002 02:05:41.495608 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b616d391-356d-4e48-bf47-be58358aa15f-run-httpd\") pod \"ceilometer-0\" (UID: \"b616d391-356d-4e48-bf47-be58358aa15f\") " pod="openstack/ceilometer-0" Oct 02 02:05:41 crc kubenswrapper[4885]: I1002 02:05:41.495650 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b616d391-356d-4e48-bf47-be58358aa15f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b616d391-356d-4e48-bf47-be58358aa15f\") " pod="openstack/ceilometer-0" Oct 02 02:05:41 crc kubenswrapper[4885]: I1002 02:05:41.495680 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b616d391-356d-4e48-bf47-be58358aa15f-log-httpd\") pod \"ceilometer-0\" (UID: \"b616d391-356d-4e48-bf47-be58358aa15f\") " pod="openstack/ceilometer-0" Oct 02 02:05:41 crc kubenswrapper[4885]: I1002 02:05:41.495699 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b616d391-356d-4e48-bf47-be58358aa15f-config-data\") pod \"ceilometer-0\" (UID: \"b616d391-356d-4e48-bf47-be58358aa15f\") " pod="openstack/ceilometer-0" Oct 02 02:05:41 crc kubenswrapper[4885]: I1002 02:05:41.496437 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b616d391-356d-4e48-bf47-be58358aa15f-log-httpd\") pod \"ceilometer-0\" (UID: \"b616d391-356d-4e48-bf47-be58358aa15f\") " pod="openstack/ceilometer-0" Oct 02 02:05:41 crc kubenswrapper[4885]: I1002 02:05:41.496478 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b616d391-356d-4e48-bf47-be58358aa15f-run-httpd\") pod \"ceilometer-0\" (UID: \"b616d391-356d-4e48-bf47-be58358aa15f\") " pod="openstack/ceilometer-0" Oct 02 02:05:41 crc kubenswrapper[4885]: I1002 02:05:41.499849 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b616d391-356d-4e48-bf47-be58358aa15f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b616d391-356d-4e48-bf47-be58358aa15f\") " pod="openstack/ceilometer-0" Oct 02 02:05:41 crc kubenswrapper[4885]: I1002 02:05:41.500078 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b616d391-356d-4e48-bf47-be58358aa15f-scripts\") pod \"ceilometer-0\" (UID: \"b616d391-356d-4e48-bf47-be58358aa15f\") " pod="openstack/ceilometer-0" Oct 02 02:05:41 crc kubenswrapper[4885]: I1002 02:05:41.500660 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b616d391-356d-4e48-bf47-be58358aa15f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b616d391-356d-4e48-bf47-be58358aa15f\") " pod="openstack/ceilometer-0" Oct 02 02:05:41 crc kubenswrapper[4885]: I1002 02:05:41.500920 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b616d391-356d-4e48-bf47-be58358aa15f-config-data\") pod \"ceilometer-0\" (UID: \"b616d391-356d-4e48-bf47-be58358aa15f\") " pod="openstack/ceilometer-0" Oct 02 02:05:41 crc kubenswrapper[4885]: I1002 02:05:41.510526 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wgbkf\" (UniqueName: \"kubernetes.io/projected/b616d391-356d-4e48-bf47-be58358aa15f-kube-api-access-wgbkf\") pod \"ceilometer-0\" (UID: \"b616d391-356d-4e48-bf47-be58358aa15f\") " pod="openstack/ceilometer-0" Oct 02 02:05:41 crc kubenswrapper[4885]: I1002 02:05:41.629408 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 02:05:42 crc kubenswrapper[4885]: I1002 02:05:42.060004 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="097aaa2c-cd9c-4a5c-b04b-110fac788275" path="/var/lib/kubelet/pods/097aaa2c-cd9c-4a5c-b04b-110fac788275/volumes" Oct 02 02:05:42 crc kubenswrapper[4885]: I1002 02:05:42.117904 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 02:05:42 crc kubenswrapper[4885]: W1002 02:05:42.133507 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb616d391_356d_4e48_bf47_be58358aa15f.slice/crio-7ae32af7799c216fa031b9b6a5202294e2eb44b3fe6fe74c71bc0408f2e1e870 WatchSource:0}: Error finding container 7ae32af7799c216fa031b9b6a5202294e2eb44b3fe6fe74c71bc0408f2e1e870: Status 404 returned error can't find the container with id 7ae32af7799c216fa031b9b6a5202294e2eb44b3fe6fe74c71bc0408f2e1e870 Oct 02 02:05:42 crc kubenswrapper[4885]: I1002 02:05:42.971955 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b616d391-356d-4e48-bf47-be58358aa15f","Type":"ContainerStarted","Data":"ca16afc7b8004a174b9e14ba3f50325e85b8b8826a36d267971a64ae668ed1d5"} Oct 02 02:05:42 crc kubenswrapper[4885]: I1002 02:05:42.972303 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b616d391-356d-4e48-bf47-be58358aa15f","Type":"ContainerStarted","Data":"7ae32af7799c216fa031b9b6a5202294e2eb44b3fe6fe74c71bc0408f2e1e870"} Oct 02 02:05:43 crc kubenswrapper[4885]: I1002 02:05:43.266340 4885 patch_prober.go:28] interesting pod/machine-config-daemon-rttx8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 02:05:43 crc kubenswrapper[4885]: I1002 02:05:43.266779 4885 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 02:05:43 crc kubenswrapper[4885]: I1002 02:05:43.475526 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 02 02:05:43 crc kubenswrapper[4885]: I1002 02:05:43.983647 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b616d391-356d-4e48-bf47-be58358aa15f","Type":"ContainerStarted","Data":"4581427155c36a20f533490e126666c54ed86a65b174b7b13d1e54fc9afff5c5"} Oct 02 02:05:44 crc kubenswrapper[4885]: I1002 02:05:44.995217 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b616d391-356d-4e48-bf47-be58358aa15f","Type":"ContainerStarted","Data":"c8bdb8ca08431286ee72e245809e203a98e6fb829a3bbedfe6650e20522c545a"} Oct 02 02:05:47 crc kubenswrapper[4885]: I1002 02:05:47.021077 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b616d391-356d-4e48-bf47-be58358aa15f","Type":"ContainerStarted","Data":"3173daded4dc1af5a514fac97c9d595dbf68ceb3b7420cfb3d1f1a96772ac5fa"} Oct 02 02:05:47 crc kubenswrapper[4885]: I1002 02:05:47.021390 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 02 02:05:47 crc kubenswrapper[4885]: I1002 02:05:47.062098 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.044480654 podStartE2EDuration="6.062071993s" podCreationTimestamp="2025-10-02 02:05:41 +0000 UTC" firstStartedPulling="2025-10-02 02:05:42.135380647 +0000 UTC m=+1130.947128096" lastFinishedPulling="2025-10-02 02:05:46.152972036 +0000 UTC m=+1134.964719435" observedRunningTime="2025-10-02 02:05:47.052661127 +0000 UTC m=+1135.864408526" watchObservedRunningTime="2025-10-02 02:05:47.062071993 +0000 UTC m=+1135.873819432" Oct 02 02:05:47 crc kubenswrapper[4885]: I1002 02:05:47.216357 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-5b7d-account-create-r6fp9"] Oct 02 02:05:47 crc kubenswrapper[4885]: I1002 02:05:47.219073 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-5b7d-account-create-r6fp9" Oct 02 02:05:47 crc kubenswrapper[4885]: I1002 02:05:47.221387 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Oct 02 02:05:47 crc kubenswrapper[4885]: I1002 02:05:47.226143 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-5b7d-account-create-r6fp9"] Oct 02 02:05:47 crc kubenswrapper[4885]: I1002 02:05:47.302248 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k26lz\" (UniqueName: \"kubernetes.io/projected/5e1ee780-2c31-401f-9822-cc58ca573908-kube-api-access-k26lz\") pod \"nova-api-5b7d-account-create-r6fp9\" (UID: \"5e1ee780-2c31-401f-9822-cc58ca573908\") " pod="openstack/nova-api-5b7d-account-create-r6fp9" Oct 02 02:05:47 crc kubenswrapper[4885]: I1002 02:05:47.404573 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k26lz\" (UniqueName: \"kubernetes.io/projected/5e1ee780-2c31-401f-9822-cc58ca573908-kube-api-access-k26lz\") pod \"nova-api-5b7d-account-create-r6fp9\" (UID: \"5e1ee780-2c31-401f-9822-cc58ca573908\") " pod="openstack/nova-api-5b7d-account-create-r6fp9" Oct 02 02:05:47 crc kubenswrapper[4885]: I1002 02:05:47.410183 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-3f2e-account-create-lp5n9"] Oct 02 02:05:47 crc kubenswrapper[4885]: I1002 02:05:47.411464 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-3f2e-account-create-lp5n9" Oct 02 02:05:47 crc kubenswrapper[4885]: I1002 02:05:47.414820 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Oct 02 02:05:47 crc kubenswrapper[4885]: I1002 02:05:47.426214 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k26lz\" (UniqueName: \"kubernetes.io/projected/5e1ee780-2c31-401f-9822-cc58ca573908-kube-api-access-k26lz\") pod \"nova-api-5b7d-account-create-r6fp9\" (UID: \"5e1ee780-2c31-401f-9822-cc58ca573908\") " pod="openstack/nova-api-5b7d-account-create-r6fp9" Oct 02 02:05:47 crc kubenswrapper[4885]: I1002 02:05:47.426594 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-3f2e-account-create-lp5n9"] Oct 02 02:05:47 crc kubenswrapper[4885]: I1002 02:05:47.506667 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lshr2\" (UniqueName: \"kubernetes.io/projected/6d1c46db-7827-474e-8dec-7b3fb2070449-kube-api-access-lshr2\") pod \"nova-cell0-3f2e-account-create-lp5n9\" (UID: \"6d1c46db-7827-474e-8dec-7b3fb2070449\") " pod="openstack/nova-cell0-3f2e-account-create-lp5n9" Oct 02 02:05:47 crc kubenswrapper[4885]: I1002 02:05:47.543462 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-5b7d-account-create-r6fp9" Oct 02 02:05:47 crc kubenswrapper[4885]: I1002 02:05:47.607773 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lshr2\" (UniqueName: \"kubernetes.io/projected/6d1c46db-7827-474e-8dec-7b3fb2070449-kube-api-access-lshr2\") pod \"nova-cell0-3f2e-account-create-lp5n9\" (UID: \"6d1c46db-7827-474e-8dec-7b3fb2070449\") " pod="openstack/nova-cell0-3f2e-account-create-lp5n9" Oct 02 02:05:47 crc kubenswrapper[4885]: I1002 02:05:47.612511 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-b3c0-account-create-tcx7f"] Oct 02 02:05:47 crc kubenswrapper[4885]: I1002 02:05:47.620897 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-b3c0-account-create-tcx7f" Oct 02 02:05:47 crc kubenswrapper[4885]: I1002 02:05:47.627988 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Oct 02 02:05:47 crc kubenswrapper[4885]: I1002 02:05:47.661775 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lshr2\" (UniqueName: \"kubernetes.io/projected/6d1c46db-7827-474e-8dec-7b3fb2070449-kube-api-access-lshr2\") pod \"nova-cell0-3f2e-account-create-lp5n9\" (UID: \"6d1c46db-7827-474e-8dec-7b3fb2070449\") " pod="openstack/nova-cell0-3f2e-account-create-lp5n9" Oct 02 02:05:47 crc kubenswrapper[4885]: I1002 02:05:47.695511 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-b3c0-account-create-tcx7f"] Oct 02 02:05:47 crc kubenswrapper[4885]: I1002 02:05:47.709583 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z9kzb\" (UniqueName: \"kubernetes.io/projected/af2b6dad-dd76-4c03-9902-d190e8a5949e-kube-api-access-z9kzb\") pod \"nova-cell1-b3c0-account-create-tcx7f\" (UID: \"af2b6dad-dd76-4c03-9902-d190e8a5949e\") " pod="openstack/nova-cell1-b3c0-account-create-tcx7f" Oct 02 02:05:47 crc kubenswrapper[4885]: I1002 02:05:47.725400 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-3f2e-account-create-lp5n9" Oct 02 02:05:47 crc kubenswrapper[4885]: I1002 02:05:47.813246 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z9kzb\" (UniqueName: \"kubernetes.io/projected/af2b6dad-dd76-4c03-9902-d190e8a5949e-kube-api-access-z9kzb\") pod \"nova-cell1-b3c0-account-create-tcx7f\" (UID: \"af2b6dad-dd76-4c03-9902-d190e8a5949e\") " pod="openstack/nova-cell1-b3c0-account-create-tcx7f" Oct 02 02:05:47 crc kubenswrapper[4885]: I1002 02:05:47.840167 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z9kzb\" (UniqueName: \"kubernetes.io/projected/af2b6dad-dd76-4c03-9902-d190e8a5949e-kube-api-access-z9kzb\") pod \"nova-cell1-b3c0-account-create-tcx7f\" (UID: \"af2b6dad-dd76-4c03-9902-d190e8a5949e\") " pod="openstack/nova-cell1-b3c0-account-create-tcx7f" Oct 02 02:05:48 crc kubenswrapper[4885]: I1002 02:05:48.004454 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-b3c0-account-create-tcx7f" Oct 02 02:05:48 crc kubenswrapper[4885]: I1002 02:05:48.015820 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-5b7d-account-create-r6fp9"] Oct 02 02:05:48 crc kubenswrapper[4885]: I1002 02:05:48.209237 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-3f2e-account-create-lp5n9"] Oct 02 02:05:48 crc kubenswrapper[4885]: W1002 02:05:48.220545 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6d1c46db_7827_474e_8dec_7b3fb2070449.slice/crio-2e30841bacfacc9a7114a7fbf723e00c5f0dec2796f0c4aa5a8baf3683402ab9 WatchSource:0}: Error finding container 2e30841bacfacc9a7114a7fbf723e00c5f0dec2796f0c4aa5a8baf3683402ab9: Status 404 returned error can't find the container with id 2e30841bacfacc9a7114a7fbf723e00c5f0dec2796f0c4aa5a8baf3683402ab9 Oct 02 02:05:48 crc kubenswrapper[4885]: I1002 02:05:48.466972 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-b3c0-account-create-tcx7f"] Oct 02 02:05:48 crc kubenswrapper[4885]: W1002 02:05:48.469662 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaf2b6dad_dd76_4c03_9902_d190e8a5949e.slice/crio-5531ac703849d57149354a0e7c8142df4a99bd444e0bd9a58669cdc972b89575 WatchSource:0}: Error finding container 5531ac703849d57149354a0e7c8142df4a99bd444e0bd9a58669cdc972b89575: Status 404 returned error can't find the container with id 5531ac703849d57149354a0e7c8142df4a99bd444e0bd9a58669cdc972b89575 Oct 02 02:05:49 crc kubenswrapper[4885]: I1002 02:05:49.041744 4885 generic.go:334] "Generic (PLEG): container finished" podID="af2b6dad-dd76-4c03-9902-d190e8a5949e" containerID="38e9ae7aa4568465f545b227259addc8c26bbdb84e7f74a00eacfc79ca7d54b2" exitCode=0 Oct 02 02:05:49 crc kubenswrapper[4885]: I1002 02:05:49.041858 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-b3c0-account-create-tcx7f" event={"ID":"af2b6dad-dd76-4c03-9902-d190e8a5949e","Type":"ContainerDied","Data":"38e9ae7aa4568465f545b227259addc8c26bbdb84e7f74a00eacfc79ca7d54b2"} Oct 02 02:05:49 crc kubenswrapper[4885]: I1002 02:05:49.041884 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-b3c0-account-create-tcx7f" event={"ID":"af2b6dad-dd76-4c03-9902-d190e8a5949e","Type":"ContainerStarted","Data":"5531ac703849d57149354a0e7c8142df4a99bd444e0bd9a58669cdc972b89575"} Oct 02 02:05:49 crc kubenswrapper[4885]: I1002 02:05:49.045216 4885 generic.go:334] "Generic (PLEG): container finished" podID="5e1ee780-2c31-401f-9822-cc58ca573908" containerID="1492fc89ff995999425a0af3d182ae6feac4ff3c58a85684452fa3c1b7ea5fd4" exitCode=0 Oct 02 02:05:49 crc kubenswrapper[4885]: I1002 02:05:49.045327 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-5b7d-account-create-r6fp9" event={"ID":"5e1ee780-2c31-401f-9822-cc58ca573908","Type":"ContainerDied","Data":"1492fc89ff995999425a0af3d182ae6feac4ff3c58a85684452fa3c1b7ea5fd4"} Oct 02 02:05:49 crc kubenswrapper[4885]: I1002 02:05:49.045355 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-5b7d-account-create-r6fp9" event={"ID":"5e1ee780-2c31-401f-9822-cc58ca573908","Type":"ContainerStarted","Data":"cfe8f83fe49e43ad0f4b4d2892c10566444b8f45af592f961b950efdda987a3f"} Oct 02 02:05:49 crc kubenswrapper[4885]: I1002 02:05:49.047804 4885 generic.go:334] "Generic (PLEG): container finished" podID="6d1c46db-7827-474e-8dec-7b3fb2070449" containerID="f0f6b0f74803390b757a6e1c56567398bbc625990fe01c57c6c902f5bf5310c8" exitCode=0 Oct 02 02:05:49 crc kubenswrapper[4885]: I1002 02:05:49.047832 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-3f2e-account-create-lp5n9" event={"ID":"6d1c46db-7827-474e-8dec-7b3fb2070449","Type":"ContainerDied","Data":"f0f6b0f74803390b757a6e1c56567398bbc625990fe01c57c6c902f5bf5310c8"} Oct 02 02:05:49 crc kubenswrapper[4885]: I1002 02:05:49.047847 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-3f2e-account-create-lp5n9" event={"ID":"6d1c46db-7827-474e-8dec-7b3fb2070449","Type":"ContainerStarted","Data":"2e30841bacfacc9a7114a7fbf723e00c5f0dec2796f0c4aa5a8baf3683402ab9"} Oct 02 02:05:50 crc kubenswrapper[4885]: I1002 02:05:50.489558 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-3f2e-account-create-lp5n9" Oct 02 02:05:50 crc kubenswrapper[4885]: I1002 02:05:50.581959 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lshr2\" (UniqueName: \"kubernetes.io/projected/6d1c46db-7827-474e-8dec-7b3fb2070449-kube-api-access-lshr2\") pod \"6d1c46db-7827-474e-8dec-7b3fb2070449\" (UID: \"6d1c46db-7827-474e-8dec-7b3fb2070449\") " Oct 02 02:05:50 crc kubenswrapper[4885]: I1002 02:05:50.586908 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d1c46db-7827-474e-8dec-7b3fb2070449-kube-api-access-lshr2" (OuterVolumeSpecName: "kube-api-access-lshr2") pod "6d1c46db-7827-474e-8dec-7b3fb2070449" (UID: "6d1c46db-7827-474e-8dec-7b3fb2070449"). InnerVolumeSpecName "kube-api-access-lshr2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:05:50 crc kubenswrapper[4885]: I1002 02:05:50.589559 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-b3c0-account-create-tcx7f" Oct 02 02:05:50 crc kubenswrapper[4885]: I1002 02:05:50.629008 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-5b7d-account-create-r6fp9" Oct 02 02:05:50 crc kubenswrapper[4885]: I1002 02:05:50.683945 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k26lz\" (UniqueName: \"kubernetes.io/projected/5e1ee780-2c31-401f-9822-cc58ca573908-kube-api-access-k26lz\") pod \"5e1ee780-2c31-401f-9822-cc58ca573908\" (UID: \"5e1ee780-2c31-401f-9822-cc58ca573908\") " Oct 02 02:05:50 crc kubenswrapper[4885]: I1002 02:05:50.684143 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z9kzb\" (UniqueName: \"kubernetes.io/projected/af2b6dad-dd76-4c03-9902-d190e8a5949e-kube-api-access-z9kzb\") pod \"af2b6dad-dd76-4c03-9902-d190e8a5949e\" (UID: \"af2b6dad-dd76-4c03-9902-d190e8a5949e\") " Oct 02 02:05:50 crc kubenswrapper[4885]: I1002 02:05:50.684646 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lshr2\" (UniqueName: \"kubernetes.io/projected/6d1c46db-7827-474e-8dec-7b3fb2070449-kube-api-access-lshr2\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:50 crc kubenswrapper[4885]: I1002 02:05:50.687557 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e1ee780-2c31-401f-9822-cc58ca573908-kube-api-access-k26lz" (OuterVolumeSpecName: "kube-api-access-k26lz") pod "5e1ee780-2c31-401f-9822-cc58ca573908" (UID: "5e1ee780-2c31-401f-9822-cc58ca573908"). InnerVolumeSpecName "kube-api-access-k26lz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:05:50 crc kubenswrapper[4885]: I1002 02:05:50.688277 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af2b6dad-dd76-4c03-9902-d190e8a5949e-kube-api-access-z9kzb" (OuterVolumeSpecName: "kube-api-access-z9kzb") pod "af2b6dad-dd76-4c03-9902-d190e8a5949e" (UID: "af2b6dad-dd76-4c03-9902-d190e8a5949e"). InnerVolumeSpecName "kube-api-access-z9kzb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:05:50 crc kubenswrapper[4885]: I1002 02:05:50.786387 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k26lz\" (UniqueName: \"kubernetes.io/projected/5e1ee780-2c31-401f-9822-cc58ca573908-kube-api-access-k26lz\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:50 crc kubenswrapper[4885]: I1002 02:05:50.786662 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z9kzb\" (UniqueName: \"kubernetes.io/projected/af2b6dad-dd76-4c03-9902-d190e8a5949e-kube-api-access-z9kzb\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:51 crc kubenswrapper[4885]: I1002 02:05:51.068479 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-b3c0-account-create-tcx7f" Oct 02 02:05:51 crc kubenswrapper[4885]: I1002 02:05:51.068776 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-b3c0-account-create-tcx7f" event={"ID":"af2b6dad-dd76-4c03-9902-d190e8a5949e","Type":"ContainerDied","Data":"5531ac703849d57149354a0e7c8142df4a99bd444e0bd9a58669cdc972b89575"} Oct 02 02:05:51 crc kubenswrapper[4885]: I1002 02:05:51.068947 4885 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5531ac703849d57149354a0e7c8142df4a99bd444e0bd9a58669cdc972b89575" Oct 02 02:05:51 crc kubenswrapper[4885]: I1002 02:05:51.069894 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-5b7d-account-create-r6fp9" event={"ID":"5e1ee780-2c31-401f-9822-cc58ca573908","Type":"ContainerDied","Data":"cfe8f83fe49e43ad0f4b4d2892c10566444b8f45af592f961b950efdda987a3f"} Oct 02 02:05:51 crc kubenswrapper[4885]: I1002 02:05:51.069940 4885 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cfe8f83fe49e43ad0f4b4d2892c10566444b8f45af592f961b950efdda987a3f" Oct 02 02:05:51 crc kubenswrapper[4885]: I1002 02:05:51.070527 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-5b7d-account-create-r6fp9" Oct 02 02:05:51 crc kubenswrapper[4885]: I1002 02:05:51.071415 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-3f2e-account-create-lp5n9" event={"ID":"6d1c46db-7827-474e-8dec-7b3fb2070449","Type":"ContainerDied","Data":"2e30841bacfacc9a7114a7fbf723e00c5f0dec2796f0c4aa5a8baf3683402ab9"} Oct 02 02:05:51 crc kubenswrapper[4885]: I1002 02:05:51.071457 4885 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2e30841bacfacc9a7114a7fbf723e00c5f0dec2796f0c4aa5a8baf3683402ab9" Oct 02 02:05:51 crc kubenswrapper[4885]: I1002 02:05:51.071508 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-3f2e-account-create-lp5n9" Oct 02 02:05:51 crc kubenswrapper[4885]: E1002 02:05:51.328139 4885 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6d1c46db_7827_474e_8dec_7b3fb2070449.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5e1ee780_2c31_401f_9822_cc58ca573908.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5e1ee780_2c31_401f_9822_cc58ca573908.slice/crio-cfe8f83fe49e43ad0f4b4d2892c10566444b8f45af592f961b950efdda987a3f\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaf2b6dad_dd76_4c03_9902_d190e8a5949e.slice\": RecentStats: unable to find data in memory cache]" Oct 02 02:05:51 crc kubenswrapper[4885]: I1002 02:05:51.697534 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 02:05:51 crc kubenswrapper[4885]: I1002 02:05:51.698071 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b616d391-356d-4e48-bf47-be58358aa15f" containerName="ceilometer-central-agent" containerID="cri-o://ca16afc7b8004a174b9e14ba3f50325e85b8b8826a36d267971a64ae668ed1d5" gracePeriod=30 Oct 02 02:05:51 crc kubenswrapper[4885]: I1002 02:05:51.698798 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b616d391-356d-4e48-bf47-be58358aa15f" containerName="proxy-httpd" containerID="cri-o://3173daded4dc1af5a514fac97c9d595dbf68ceb3b7420cfb3d1f1a96772ac5fa" gracePeriod=30 Oct 02 02:05:51 crc kubenswrapper[4885]: I1002 02:05:51.698861 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b616d391-356d-4e48-bf47-be58358aa15f" containerName="sg-core" containerID="cri-o://c8bdb8ca08431286ee72e245809e203a98e6fb829a3bbedfe6650e20522c545a" gracePeriod=30 Oct 02 02:05:51 crc kubenswrapper[4885]: I1002 02:05:51.698901 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b616d391-356d-4e48-bf47-be58358aa15f" containerName="ceilometer-notification-agent" containerID="cri-o://4581427155c36a20f533490e126666c54ed86a65b174b7b13d1e54fc9afff5c5" gracePeriod=30 Oct 02 02:05:52 crc kubenswrapper[4885]: I1002 02:05:52.082766 4885 generic.go:334] "Generic (PLEG): container finished" podID="b616d391-356d-4e48-bf47-be58358aa15f" containerID="3173daded4dc1af5a514fac97c9d595dbf68ceb3b7420cfb3d1f1a96772ac5fa" exitCode=0 Oct 02 02:05:52 crc kubenswrapper[4885]: I1002 02:05:52.082797 4885 generic.go:334] "Generic (PLEG): container finished" podID="b616d391-356d-4e48-bf47-be58358aa15f" containerID="c8bdb8ca08431286ee72e245809e203a98e6fb829a3bbedfe6650e20522c545a" exitCode=2 Oct 02 02:05:52 crc kubenswrapper[4885]: I1002 02:05:52.082815 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b616d391-356d-4e48-bf47-be58358aa15f","Type":"ContainerDied","Data":"3173daded4dc1af5a514fac97c9d595dbf68ceb3b7420cfb3d1f1a96772ac5fa"} Oct 02 02:05:52 crc kubenswrapper[4885]: I1002 02:05:52.082838 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b616d391-356d-4e48-bf47-be58358aa15f","Type":"ContainerDied","Data":"c8bdb8ca08431286ee72e245809e203a98e6fb829a3bbedfe6650e20522c545a"} Oct 02 02:05:52 crc kubenswrapper[4885]: I1002 02:05:52.549959 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 02:05:52 crc kubenswrapper[4885]: I1002 02:05:52.619207 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b616d391-356d-4e48-bf47-be58358aa15f-log-httpd\") pod \"b616d391-356d-4e48-bf47-be58358aa15f\" (UID: \"b616d391-356d-4e48-bf47-be58358aa15f\") " Oct 02 02:05:52 crc kubenswrapper[4885]: I1002 02:05:52.619367 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wgbkf\" (UniqueName: \"kubernetes.io/projected/b616d391-356d-4e48-bf47-be58358aa15f-kube-api-access-wgbkf\") pod \"b616d391-356d-4e48-bf47-be58358aa15f\" (UID: \"b616d391-356d-4e48-bf47-be58358aa15f\") " Oct 02 02:05:52 crc kubenswrapper[4885]: I1002 02:05:52.619397 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b616d391-356d-4e48-bf47-be58358aa15f-sg-core-conf-yaml\") pod \"b616d391-356d-4e48-bf47-be58358aa15f\" (UID: \"b616d391-356d-4e48-bf47-be58358aa15f\") " Oct 02 02:05:52 crc kubenswrapper[4885]: I1002 02:05:52.619416 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b616d391-356d-4e48-bf47-be58358aa15f-config-data\") pod \"b616d391-356d-4e48-bf47-be58358aa15f\" (UID: \"b616d391-356d-4e48-bf47-be58358aa15f\") " Oct 02 02:05:52 crc kubenswrapper[4885]: I1002 02:05:52.619484 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b616d391-356d-4e48-bf47-be58358aa15f-run-httpd\") pod \"b616d391-356d-4e48-bf47-be58358aa15f\" (UID: \"b616d391-356d-4e48-bf47-be58358aa15f\") " Oct 02 02:05:52 crc kubenswrapper[4885]: I1002 02:05:52.619516 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b616d391-356d-4e48-bf47-be58358aa15f-scripts\") pod \"b616d391-356d-4e48-bf47-be58358aa15f\" (UID: \"b616d391-356d-4e48-bf47-be58358aa15f\") " Oct 02 02:05:52 crc kubenswrapper[4885]: I1002 02:05:52.619577 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b616d391-356d-4e48-bf47-be58358aa15f-combined-ca-bundle\") pod \"b616d391-356d-4e48-bf47-be58358aa15f\" (UID: \"b616d391-356d-4e48-bf47-be58358aa15f\") " Oct 02 02:05:52 crc kubenswrapper[4885]: I1002 02:05:52.620005 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b616d391-356d-4e48-bf47-be58358aa15f-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "b616d391-356d-4e48-bf47-be58358aa15f" (UID: "b616d391-356d-4e48-bf47-be58358aa15f"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:05:52 crc kubenswrapper[4885]: I1002 02:05:52.620028 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b616d391-356d-4e48-bf47-be58358aa15f-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "b616d391-356d-4e48-bf47-be58358aa15f" (UID: "b616d391-356d-4e48-bf47-be58358aa15f"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:05:52 crc kubenswrapper[4885]: I1002 02:05:52.637408 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b616d391-356d-4e48-bf47-be58358aa15f-scripts" (OuterVolumeSpecName: "scripts") pod "b616d391-356d-4e48-bf47-be58358aa15f" (UID: "b616d391-356d-4e48-bf47-be58358aa15f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:05:52 crc kubenswrapper[4885]: I1002 02:05:52.643122 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b616d391-356d-4e48-bf47-be58358aa15f-kube-api-access-wgbkf" (OuterVolumeSpecName: "kube-api-access-wgbkf") pod "b616d391-356d-4e48-bf47-be58358aa15f" (UID: "b616d391-356d-4e48-bf47-be58358aa15f"). InnerVolumeSpecName "kube-api-access-wgbkf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:05:52 crc kubenswrapper[4885]: I1002 02:05:52.673077 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-zcpc9"] Oct 02 02:05:52 crc kubenswrapper[4885]: E1002 02:05:52.674006 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af2b6dad-dd76-4c03-9902-d190e8a5949e" containerName="mariadb-account-create" Oct 02 02:05:52 crc kubenswrapper[4885]: I1002 02:05:52.674083 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="af2b6dad-dd76-4c03-9902-d190e8a5949e" containerName="mariadb-account-create" Oct 02 02:05:52 crc kubenswrapper[4885]: E1002 02:05:52.675515 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b616d391-356d-4e48-bf47-be58358aa15f" containerName="sg-core" Oct 02 02:05:52 crc kubenswrapper[4885]: I1002 02:05:52.675579 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="b616d391-356d-4e48-bf47-be58358aa15f" containerName="sg-core" Oct 02 02:05:52 crc kubenswrapper[4885]: E1002 02:05:52.675640 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d1c46db-7827-474e-8dec-7b3fb2070449" containerName="mariadb-account-create" Oct 02 02:05:52 crc kubenswrapper[4885]: I1002 02:05:52.675689 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d1c46db-7827-474e-8dec-7b3fb2070449" containerName="mariadb-account-create" Oct 02 02:05:52 crc kubenswrapper[4885]: E1002 02:05:52.675740 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b616d391-356d-4e48-bf47-be58358aa15f" containerName="ceilometer-central-agent" Oct 02 02:05:52 crc kubenswrapper[4885]: I1002 02:05:52.675800 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="b616d391-356d-4e48-bf47-be58358aa15f" containerName="ceilometer-central-agent" Oct 02 02:05:52 crc kubenswrapper[4885]: E1002 02:05:52.676522 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e1ee780-2c31-401f-9822-cc58ca573908" containerName="mariadb-account-create" Oct 02 02:05:52 crc kubenswrapper[4885]: I1002 02:05:52.676586 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e1ee780-2c31-401f-9822-cc58ca573908" containerName="mariadb-account-create" Oct 02 02:05:52 crc kubenswrapper[4885]: E1002 02:05:52.676651 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b616d391-356d-4e48-bf47-be58358aa15f" containerName="ceilometer-notification-agent" Oct 02 02:05:52 crc kubenswrapper[4885]: I1002 02:05:52.676699 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="b616d391-356d-4e48-bf47-be58358aa15f" containerName="ceilometer-notification-agent" Oct 02 02:05:52 crc kubenswrapper[4885]: E1002 02:05:52.676752 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b616d391-356d-4e48-bf47-be58358aa15f" containerName="proxy-httpd" Oct 02 02:05:52 crc kubenswrapper[4885]: I1002 02:05:52.676799 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="b616d391-356d-4e48-bf47-be58358aa15f" containerName="proxy-httpd" Oct 02 02:05:52 crc kubenswrapper[4885]: I1002 02:05:52.677286 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="b616d391-356d-4e48-bf47-be58358aa15f" containerName="proxy-httpd" Oct 02 02:05:52 crc kubenswrapper[4885]: I1002 02:05:52.677359 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="af2b6dad-dd76-4c03-9902-d190e8a5949e" containerName="mariadb-account-create" Oct 02 02:05:52 crc kubenswrapper[4885]: I1002 02:05:52.677425 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="b616d391-356d-4e48-bf47-be58358aa15f" containerName="ceilometer-notification-agent" Oct 02 02:05:52 crc kubenswrapper[4885]: I1002 02:05:52.677483 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d1c46db-7827-474e-8dec-7b3fb2070449" containerName="mariadb-account-create" Oct 02 02:05:52 crc kubenswrapper[4885]: I1002 02:05:52.677537 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="b616d391-356d-4e48-bf47-be58358aa15f" containerName="sg-core" Oct 02 02:05:52 crc kubenswrapper[4885]: I1002 02:05:52.677592 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e1ee780-2c31-401f-9822-cc58ca573908" containerName="mariadb-account-create" Oct 02 02:05:52 crc kubenswrapper[4885]: I1002 02:05:52.677662 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="b616d391-356d-4e48-bf47-be58358aa15f" containerName="ceilometer-central-agent" Oct 02 02:05:52 crc kubenswrapper[4885]: I1002 02:05:52.678301 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-zcpc9" Oct 02 02:05:52 crc kubenswrapper[4885]: I1002 02:05:52.679112 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-zcpc9"] Oct 02 02:05:52 crc kubenswrapper[4885]: I1002 02:05:52.682067 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 02 02:05:52 crc kubenswrapper[4885]: I1002 02:05:52.682185 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-fxj22" Oct 02 02:05:52 crc kubenswrapper[4885]: I1002 02:05:52.684053 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Oct 02 02:05:52 crc kubenswrapper[4885]: I1002 02:05:52.708127 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b616d391-356d-4e48-bf47-be58358aa15f-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "b616d391-356d-4e48-bf47-be58358aa15f" (UID: "b616d391-356d-4e48-bf47-be58358aa15f"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:05:52 crc kubenswrapper[4885]: I1002 02:05:52.725298 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb797f9e-26a8-42fb-bac6-f5c145103aca-config-data\") pod \"nova-cell0-conductor-db-sync-zcpc9\" (UID: \"cb797f9e-26a8-42fb-bac6-f5c145103aca\") " pod="openstack/nova-cell0-conductor-db-sync-zcpc9" Oct 02 02:05:52 crc kubenswrapper[4885]: I1002 02:05:52.725967 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cb797f9e-26a8-42fb-bac6-f5c145103aca-scripts\") pod \"nova-cell0-conductor-db-sync-zcpc9\" (UID: \"cb797f9e-26a8-42fb-bac6-f5c145103aca\") " pod="openstack/nova-cell0-conductor-db-sync-zcpc9" Oct 02 02:05:52 crc kubenswrapper[4885]: I1002 02:05:52.726569 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb797f9e-26a8-42fb-bac6-f5c145103aca-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-zcpc9\" (UID: \"cb797f9e-26a8-42fb-bac6-f5c145103aca\") " pod="openstack/nova-cell0-conductor-db-sync-zcpc9" Oct 02 02:05:52 crc kubenswrapper[4885]: I1002 02:05:52.726690 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jj59f\" (UniqueName: \"kubernetes.io/projected/cb797f9e-26a8-42fb-bac6-f5c145103aca-kube-api-access-jj59f\") pod \"nova-cell0-conductor-db-sync-zcpc9\" (UID: \"cb797f9e-26a8-42fb-bac6-f5c145103aca\") " pod="openstack/nova-cell0-conductor-db-sync-zcpc9" Oct 02 02:05:52 crc kubenswrapper[4885]: I1002 02:05:52.727054 4885 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b616d391-356d-4e48-bf47-be58358aa15f-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:52 crc kubenswrapper[4885]: I1002 02:05:52.727150 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wgbkf\" (UniqueName: \"kubernetes.io/projected/b616d391-356d-4e48-bf47-be58358aa15f-kube-api-access-wgbkf\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:52 crc kubenswrapper[4885]: I1002 02:05:52.727210 4885 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b616d391-356d-4e48-bf47-be58358aa15f-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:52 crc kubenswrapper[4885]: I1002 02:05:52.727280 4885 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b616d391-356d-4e48-bf47-be58358aa15f-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:52 crc kubenswrapper[4885]: I1002 02:05:52.727333 4885 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b616d391-356d-4e48-bf47-be58358aa15f-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:52 crc kubenswrapper[4885]: I1002 02:05:52.781473 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b616d391-356d-4e48-bf47-be58358aa15f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b616d391-356d-4e48-bf47-be58358aa15f" (UID: "b616d391-356d-4e48-bf47-be58358aa15f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:05:52 crc kubenswrapper[4885]: I1002 02:05:52.790426 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b616d391-356d-4e48-bf47-be58358aa15f-config-data" (OuterVolumeSpecName: "config-data") pod "b616d391-356d-4e48-bf47-be58358aa15f" (UID: "b616d391-356d-4e48-bf47-be58358aa15f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:05:52 crc kubenswrapper[4885]: I1002 02:05:52.829034 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jj59f\" (UniqueName: \"kubernetes.io/projected/cb797f9e-26a8-42fb-bac6-f5c145103aca-kube-api-access-jj59f\") pod \"nova-cell0-conductor-db-sync-zcpc9\" (UID: \"cb797f9e-26a8-42fb-bac6-f5c145103aca\") " pod="openstack/nova-cell0-conductor-db-sync-zcpc9" Oct 02 02:05:52 crc kubenswrapper[4885]: I1002 02:05:52.829483 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb797f9e-26a8-42fb-bac6-f5c145103aca-config-data\") pod \"nova-cell0-conductor-db-sync-zcpc9\" (UID: \"cb797f9e-26a8-42fb-bac6-f5c145103aca\") " pod="openstack/nova-cell0-conductor-db-sync-zcpc9" Oct 02 02:05:52 crc kubenswrapper[4885]: I1002 02:05:52.829679 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cb797f9e-26a8-42fb-bac6-f5c145103aca-scripts\") pod \"nova-cell0-conductor-db-sync-zcpc9\" (UID: \"cb797f9e-26a8-42fb-bac6-f5c145103aca\") " pod="openstack/nova-cell0-conductor-db-sync-zcpc9" Oct 02 02:05:52 crc kubenswrapper[4885]: I1002 02:05:52.830173 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb797f9e-26a8-42fb-bac6-f5c145103aca-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-zcpc9\" (UID: \"cb797f9e-26a8-42fb-bac6-f5c145103aca\") " pod="openstack/nova-cell0-conductor-db-sync-zcpc9" Oct 02 02:05:52 crc kubenswrapper[4885]: I1002 02:05:52.830421 4885 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b616d391-356d-4e48-bf47-be58358aa15f-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:52 crc kubenswrapper[4885]: I1002 02:05:52.830531 4885 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b616d391-356d-4e48-bf47-be58358aa15f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:52 crc kubenswrapper[4885]: I1002 02:05:52.832765 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cb797f9e-26a8-42fb-bac6-f5c145103aca-scripts\") pod \"nova-cell0-conductor-db-sync-zcpc9\" (UID: \"cb797f9e-26a8-42fb-bac6-f5c145103aca\") " pod="openstack/nova-cell0-conductor-db-sync-zcpc9" Oct 02 02:05:52 crc kubenswrapper[4885]: I1002 02:05:52.833728 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb797f9e-26a8-42fb-bac6-f5c145103aca-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-zcpc9\" (UID: \"cb797f9e-26a8-42fb-bac6-f5c145103aca\") " pod="openstack/nova-cell0-conductor-db-sync-zcpc9" Oct 02 02:05:52 crc kubenswrapper[4885]: I1002 02:05:52.835790 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb797f9e-26a8-42fb-bac6-f5c145103aca-config-data\") pod \"nova-cell0-conductor-db-sync-zcpc9\" (UID: \"cb797f9e-26a8-42fb-bac6-f5c145103aca\") " pod="openstack/nova-cell0-conductor-db-sync-zcpc9" Oct 02 02:05:52 crc kubenswrapper[4885]: I1002 02:05:52.846441 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jj59f\" (UniqueName: \"kubernetes.io/projected/cb797f9e-26a8-42fb-bac6-f5c145103aca-kube-api-access-jj59f\") pod \"nova-cell0-conductor-db-sync-zcpc9\" (UID: \"cb797f9e-26a8-42fb-bac6-f5c145103aca\") " pod="openstack/nova-cell0-conductor-db-sync-zcpc9" Oct 02 02:05:53 crc kubenswrapper[4885]: I1002 02:05:53.075520 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-zcpc9" Oct 02 02:05:53 crc kubenswrapper[4885]: I1002 02:05:53.110539 4885 generic.go:334] "Generic (PLEG): container finished" podID="b616d391-356d-4e48-bf47-be58358aa15f" containerID="4581427155c36a20f533490e126666c54ed86a65b174b7b13d1e54fc9afff5c5" exitCode=0 Oct 02 02:05:53 crc kubenswrapper[4885]: I1002 02:05:53.110575 4885 generic.go:334] "Generic (PLEG): container finished" podID="b616d391-356d-4e48-bf47-be58358aa15f" containerID="ca16afc7b8004a174b9e14ba3f50325e85b8b8826a36d267971a64ae668ed1d5" exitCode=0 Oct 02 02:05:53 crc kubenswrapper[4885]: I1002 02:05:53.110598 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b616d391-356d-4e48-bf47-be58358aa15f","Type":"ContainerDied","Data":"4581427155c36a20f533490e126666c54ed86a65b174b7b13d1e54fc9afff5c5"} Oct 02 02:05:53 crc kubenswrapper[4885]: I1002 02:05:53.110625 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b616d391-356d-4e48-bf47-be58358aa15f","Type":"ContainerDied","Data":"ca16afc7b8004a174b9e14ba3f50325e85b8b8826a36d267971a64ae668ed1d5"} Oct 02 02:05:53 crc kubenswrapper[4885]: I1002 02:05:53.110646 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b616d391-356d-4e48-bf47-be58358aa15f","Type":"ContainerDied","Data":"7ae32af7799c216fa031b9b6a5202294e2eb44b3fe6fe74c71bc0408f2e1e870"} Oct 02 02:05:53 crc kubenswrapper[4885]: I1002 02:05:53.110663 4885 scope.go:117] "RemoveContainer" containerID="3173daded4dc1af5a514fac97c9d595dbf68ceb3b7420cfb3d1f1a96772ac5fa" Oct 02 02:05:53 crc kubenswrapper[4885]: I1002 02:05:53.110695 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 02:05:53 crc kubenswrapper[4885]: I1002 02:05:53.150519 4885 scope.go:117] "RemoveContainer" containerID="c8bdb8ca08431286ee72e245809e203a98e6fb829a3bbedfe6650e20522c545a" Oct 02 02:05:53 crc kubenswrapper[4885]: I1002 02:05:53.182764 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 02:05:53 crc kubenswrapper[4885]: I1002 02:05:53.194470 4885 scope.go:117] "RemoveContainer" containerID="4581427155c36a20f533490e126666c54ed86a65b174b7b13d1e54fc9afff5c5" Oct 02 02:05:53 crc kubenswrapper[4885]: I1002 02:05:53.194628 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 02 02:05:53 crc kubenswrapper[4885]: I1002 02:05:53.201415 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 02 02:05:53 crc kubenswrapper[4885]: I1002 02:05:53.204360 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 02:05:53 crc kubenswrapper[4885]: I1002 02:05:53.206530 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 02:05:53 crc kubenswrapper[4885]: I1002 02:05:53.208288 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 02 02:05:53 crc kubenswrapper[4885]: I1002 02:05:53.208511 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 02 02:05:53 crc kubenswrapper[4885]: I1002 02:05:53.238790 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31\") " pod="openstack/ceilometer-0" Oct 02 02:05:53 crc kubenswrapper[4885]: I1002 02:05:53.238864 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sptk8\" (UniqueName: \"kubernetes.io/projected/3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31-kube-api-access-sptk8\") pod \"ceilometer-0\" (UID: \"3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31\") " pod="openstack/ceilometer-0" Oct 02 02:05:53 crc kubenswrapper[4885]: I1002 02:05:53.238882 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31-run-httpd\") pod \"ceilometer-0\" (UID: \"3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31\") " pod="openstack/ceilometer-0" Oct 02 02:05:53 crc kubenswrapper[4885]: I1002 02:05:53.238913 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31-config-data\") pod \"ceilometer-0\" (UID: \"3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31\") " pod="openstack/ceilometer-0" Oct 02 02:05:53 crc kubenswrapper[4885]: I1002 02:05:53.238934 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31-log-httpd\") pod \"ceilometer-0\" (UID: \"3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31\") " pod="openstack/ceilometer-0" Oct 02 02:05:53 crc kubenswrapper[4885]: I1002 02:05:53.238948 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31\") " pod="openstack/ceilometer-0" Oct 02 02:05:53 crc kubenswrapper[4885]: I1002 02:05:53.238970 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31-scripts\") pod \"ceilometer-0\" (UID: \"3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31\") " pod="openstack/ceilometer-0" Oct 02 02:05:53 crc kubenswrapper[4885]: I1002 02:05:53.260365 4885 scope.go:117] "RemoveContainer" containerID="ca16afc7b8004a174b9e14ba3f50325e85b8b8826a36d267971a64ae668ed1d5" Oct 02 02:05:53 crc kubenswrapper[4885]: I1002 02:05:53.305755 4885 scope.go:117] "RemoveContainer" containerID="3173daded4dc1af5a514fac97c9d595dbf68ceb3b7420cfb3d1f1a96772ac5fa" Oct 02 02:05:53 crc kubenswrapper[4885]: E1002 02:05:53.306418 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3173daded4dc1af5a514fac97c9d595dbf68ceb3b7420cfb3d1f1a96772ac5fa\": container with ID starting with 3173daded4dc1af5a514fac97c9d595dbf68ceb3b7420cfb3d1f1a96772ac5fa not found: ID does not exist" containerID="3173daded4dc1af5a514fac97c9d595dbf68ceb3b7420cfb3d1f1a96772ac5fa" Oct 02 02:05:53 crc kubenswrapper[4885]: I1002 02:05:53.306464 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3173daded4dc1af5a514fac97c9d595dbf68ceb3b7420cfb3d1f1a96772ac5fa"} err="failed to get container status \"3173daded4dc1af5a514fac97c9d595dbf68ceb3b7420cfb3d1f1a96772ac5fa\": rpc error: code = NotFound desc = could not find container \"3173daded4dc1af5a514fac97c9d595dbf68ceb3b7420cfb3d1f1a96772ac5fa\": container with ID starting with 3173daded4dc1af5a514fac97c9d595dbf68ceb3b7420cfb3d1f1a96772ac5fa not found: ID does not exist" Oct 02 02:05:53 crc kubenswrapper[4885]: I1002 02:05:53.306501 4885 scope.go:117] "RemoveContainer" containerID="c8bdb8ca08431286ee72e245809e203a98e6fb829a3bbedfe6650e20522c545a" Oct 02 02:05:53 crc kubenswrapper[4885]: E1002 02:05:53.308517 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c8bdb8ca08431286ee72e245809e203a98e6fb829a3bbedfe6650e20522c545a\": container with ID starting with c8bdb8ca08431286ee72e245809e203a98e6fb829a3bbedfe6650e20522c545a not found: ID does not exist" containerID="c8bdb8ca08431286ee72e245809e203a98e6fb829a3bbedfe6650e20522c545a" Oct 02 02:05:53 crc kubenswrapper[4885]: I1002 02:05:53.308552 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c8bdb8ca08431286ee72e245809e203a98e6fb829a3bbedfe6650e20522c545a"} err="failed to get container status \"c8bdb8ca08431286ee72e245809e203a98e6fb829a3bbedfe6650e20522c545a\": rpc error: code = NotFound desc = could not find container \"c8bdb8ca08431286ee72e245809e203a98e6fb829a3bbedfe6650e20522c545a\": container with ID starting with c8bdb8ca08431286ee72e245809e203a98e6fb829a3bbedfe6650e20522c545a not found: ID does not exist" Oct 02 02:05:53 crc kubenswrapper[4885]: I1002 02:05:53.311455 4885 scope.go:117] "RemoveContainer" containerID="4581427155c36a20f533490e126666c54ed86a65b174b7b13d1e54fc9afff5c5" Oct 02 02:05:53 crc kubenswrapper[4885]: E1002 02:05:53.312149 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4581427155c36a20f533490e126666c54ed86a65b174b7b13d1e54fc9afff5c5\": container with ID starting with 4581427155c36a20f533490e126666c54ed86a65b174b7b13d1e54fc9afff5c5 not found: ID does not exist" containerID="4581427155c36a20f533490e126666c54ed86a65b174b7b13d1e54fc9afff5c5" Oct 02 02:05:53 crc kubenswrapper[4885]: I1002 02:05:53.312182 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4581427155c36a20f533490e126666c54ed86a65b174b7b13d1e54fc9afff5c5"} err="failed to get container status \"4581427155c36a20f533490e126666c54ed86a65b174b7b13d1e54fc9afff5c5\": rpc error: code = NotFound desc = could not find container \"4581427155c36a20f533490e126666c54ed86a65b174b7b13d1e54fc9afff5c5\": container with ID starting with 4581427155c36a20f533490e126666c54ed86a65b174b7b13d1e54fc9afff5c5 not found: ID does not exist" Oct 02 02:05:53 crc kubenswrapper[4885]: I1002 02:05:53.312209 4885 scope.go:117] "RemoveContainer" containerID="ca16afc7b8004a174b9e14ba3f50325e85b8b8826a36d267971a64ae668ed1d5" Oct 02 02:05:53 crc kubenswrapper[4885]: E1002 02:05:53.312800 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ca16afc7b8004a174b9e14ba3f50325e85b8b8826a36d267971a64ae668ed1d5\": container with ID starting with ca16afc7b8004a174b9e14ba3f50325e85b8b8826a36d267971a64ae668ed1d5 not found: ID does not exist" containerID="ca16afc7b8004a174b9e14ba3f50325e85b8b8826a36d267971a64ae668ed1d5" Oct 02 02:05:53 crc kubenswrapper[4885]: I1002 02:05:53.312842 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ca16afc7b8004a174b9e14ba3f50325e85b8b8826a36d267971a64ae668ed1d5"} err="failed to get container status \"ca16afc7b8004a174b9e14ba3f50325e85b8b8826a36d267971a64ae668ed1d5\": rpc error: code = NotFound desc = could not find container \"ca16afc7b8004a174b9e14ba3f50325e85b8b8826a36d267971a64ae668ed1d5\": container with ID starting with ca16afc7b8004a174b9e14ba3f50325e85b8b8826a36d267971a64ae668ed1d5 not found: ID does not exist" Oct 02 02:05:53 crc kubenswrapper[4885]: I1002 02:05:53.312874 4885 scope.go:117] "RemoveContainer" containerID="3173daded4dc1af5a514fac97c9d595dbf68ceb3b7420cfb3d1f1a96772ac5fa" Oct 02 02:05:53 crc kubenswrapper[4885]: I1002 02:05:53.313250 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3173daded4dc1af5a514fac97c9d595dbf68ceb3b7420cfb3d1f1a96772ac5fa"} err="failed to get container status \"3173daded4dc1af5a514fac97c9d595dbf68ceb3b7420cfb3d1f1a96772ac5fa\": rpc error: code = NotFound desc = could not find container \"3173daded4dc1af5a514fac97c9d595dbf68ceb3b7420cfb3d1f1a96772ac5fa\": container with ID starting with 3173daded4dc1af5a514fac97c9d595dbf68ceb3b7420cfb3d1f1a96772ac5fa not found: ID does not exist" Oct 02 02:05:53 crc kubenswrapper[4885]: I1002 02:05:53.313280 4885 scope.go:117] "RemoveContainer" containerID="c8bdb8ca08431286ee72e245809e203a98e6fb829a3bbedfe6650e20522c545a" Oct 02 02:05:53 crc kubenswrapper[4885]: I1002 02:05:53.313598 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c8bdb8ca08431286ee72e245809e203a98e6fb829a3bbedfe6650e20522c545a"} err="failed to get container status \"c8bdb8ca08431286ee72e245809e203a98e6fb829a3bbedfe6650e20522c545a\": rpc error: code = NotFound desc = could not find container \"c8bdb8ca08431286ee72e245809e203a98e6fb829a3bbedfe6650e20522c545a\": container with ID starting with c8bdb8ca08431286ee72e245809e203a98e6fb829a3bbedfe6650e20522c545a not found: ID does not exist" Oct 02 02:05:53 crc kubenswrapper[4885]: I1002 02:05:53.313612 4885 scope.go:117] "RemoveContainer" containerID="4581427155c36a20f533490e126666c54ed86a65b174b7b13d1e54fc9afff5c5" Oct 02 02:05:53 crc kubenswrapper[4885]: I1002 02:05:53.313836 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4581427155c36a20f533490e126666c54ed86a65b174b7b13d1e54fc9afff5c5"} err="failed to get container status \"4581427155c36a20f533490e126666c54ed86a65b174b7b13d1e54fc9afff5c5\": rpc error: code = NotFound desc = could not find container \"4581427155c36a20f533490e126666c54ed86a65b174b7b13d1e54fc9afff5c5\": container with ID starting with 4581427155c36a20f533490e126666c54ed86a65b174b7b13d1e54fc9afff5c5 not found: ID does not exist" Oct 02 02:05:53 crc kubenswrapper[4885]: I1002 02:05:53.313888 4885 scope.go:117] "RemoveContainer" containerID="ca16afc7b8004a174b9e14ba3f50325e85b8b8826a36d267971a64ae668ed1d5" Oct 02 02:05:53 crc kubenswrapper[4885]: I1002 02:05:53.314093 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ca16afc7b8004a174b9e14ba3f50325e85b8b8826a36d267971a64ae668ed1d5"} err="failed to get container status \"ca16afc7b8004a174b9e14ba3f50325e85b8b8826a36d267971a64ae668ed1d5\": rpc error: code = NotFound desc = could not find container \"ca16afc7b8004a174b9e14ba3f50325e85b8b8826a36d267971a64ae668ed1d5\": container with ID starting with ca16afc7b8004a174b9e14ba3f50325e85b8b8826a36d267971a64ae668ed1d5 not found: ID does not exist" Oct 02 02:05:53 crc kubenswrapper[4885]: I1002 02:05:53.340233 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31\") " pod="openstack/ceilometer-0" Oct 02 02:05:53 crc kubenswrapper[4885]: I1002 02:05:53.340336 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sptk8\" (UniqueName: \"kubernetes.io/projected/3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31-kube-api-access-sptk8\") pod \"ceilometer-0\" (UID: \"3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31\") " pod="openstack/ceilometer-0" Oct 02 02:05:53 crc kubenswrapper[4885]: I1002 02:05:53.340355 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31-run-httpd\") pod \"ceilometer-0\" (UID: \"3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31\") " pod="openstack/ceilometer-0" Oct 02 02:05:53 crc kubenswrapper[4885]: I1002 02:05:53.340387 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31-config-data\") pod \"ceilometer-0\" (UID: \"3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31\") " pod="openstack/ceilometer-0" Oct 02 02:05:53 crc kubenswrapper[4885]: I1002 02:05:53.340411 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31-log-httpd\") pod \"ceilometer-0\" (UID: \"3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31\") " pod="openstack/ceilometer-0" Oct 02 02:05:53 crc kubenswrapper[4885]: I1002 02:05:53.340426 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31\") " pod="openstack/ceilometer-0" Oct 02 02:05:53 crc kubenswrapper[4885]: I1002 02:05:53.340449 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31-scripts\") pod \"ceilometer-0\" (UID: \"3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31\") " pod="openstack/ceilometer-0" Oct 02 02:05:53 crc kubenswrapper[4885]: I1002 02:05:53.341557 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31-run-httpd\") pod \"ceilometer-0\" (UID: \"3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31\") " pod="openstack/ceilometer-0" Oct 02 02:05:53 crc kubenswrapper[4885]: I1002 02:05:53.341753 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31-log-httpd\") pod \"ceilometer-0\" (UID: \"3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31\") " pod="openstack/ceilometer-0" Oct 02 02:05:53 crc kubenswrapper[4885]: I1002 02:05:53.344821 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31\") " pod="openstack/ceilometer-0" Oct 02 02:05:53 crc kubenswrapper[4885]: I1002 02:05:53.347392 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31-scripts\") pod \"ceilometer-0\" (UID: \"3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31\") " pod="openstack/ceilometer-0" Oct 02 02:05:53 crc kubenswrapper[4885]: I1002 02:05:53.350825 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31-config-data\") pod \"ceilometer-0\" (UID: \"3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31\") " pod="openstack/ceilometer-0" Oct 02 02:05:53 crc kubenswrapper[4885]: I1002 02:05:53.353676 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31\") " pod="openstack/ceilometer-0" Oct 02 02:05:53 crc kubenswrapper[4885]: I1002 02:05:53.356378 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sptk8\" (UniqueName: \"kubernetes.io/projected/3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31-kube-api-access-sptk8\") pod \"ceilometer-0\" (UID: \"3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31\") " pod="openstack/ceilometer-0" Oct 02 02:05:53 crc kubenswrapper[4885]: I1002 02:05:53.535802 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 02:05:53 crc kubenswrapper[4885]: I1002 02:05:53.612400 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-zcpc9"] Oct 02 02:05:54 crc kubenswrapper[4885]: I1002 02:05:54.027151 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 02:05:54 crc kubenswrapper[4885]: W1002 02:05:54.037385 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3a0eeb33_7f67_4d0c_9b18_3ccbf8715a31.slice/crio-6347f77e8d0ea5295e11111f03572d32eac3e9c5f0c3874548bd51be05fd980b WatchSource:0}: Error finding container 6347f77e8d0ea5295e11111f03572d32eac3e9c5f0c3874548bd51be05fd980b: Status 404 returned error can't find the container with id 6347f77e8d0ea5295e11111f03572d32eac3e9c5f0c3874548bd51be05fd980b Oct 02 02:05:54 crc kubenswrapper[4885]: I1002 02:05:54.061037 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b616d391-356d-4e48-bf47-be58358aa15f" path="/var/lib/kubelet/pods/b616d391-356d-4e48-bf47-be58358aa15f/volumes" Oct 02 02:05:54 crc kubenswrapper[4885]: I1002 02:05:54.121062 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31","Type":"ContainerStarted","Data":"6347f77e8d0ea5295e11111f03572d32eac3e9c5f0c3874548bd51be05fd980b"} Oct 02 02:05:54 crc kubenswrapper[4885]: I1002 02:05:54.122446 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-zcpc9" event={"ID":"cb797f9e-26a8-42fb-bac6-f5c145103aca","Type":"ContainerStarted","Data":"0bc418da516fd37b6a50f644e58bfc862a4f2cc8c20bba69c776533cf93eb76c"} Oct 02 02:05:55 crc kubenswrapper[4885]: I1002 02:05:55.133213 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31","Type":"ContainerStarted","Data":"37415b45895365cfac8d3cda583e36f42405ad139ab5a23a98f34186e2451267"} Oct 02 02:05:56 crc kubenswrapper[4885]: I1002 02:05:56.145287 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31","Type":"ContainerStarted","Data":"fa0d2db4bf10abe0c8e33d54ab6448de1f276f34da0265afacb000713267d8c8"} Oct 02 02:05:57 crc kubenswrapper[4885]: I1002 02:05:57.155074 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31","Type":"ContainerStarted","Data":"196cd6495bbe8e9aef57fcdd9cde8457cfc188166141d4cc6cd72a97e3e3d3a0"} Oct 02 02:05:57 crc kubenswrapper[4885]: I1002 02:05:57.730894 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 02:06:02 crc kubenswrapper[4885]: I1002 02:06:02.220214 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-zcpc9" event={"ID":"cb797f9e-26a8-42fb-bac6-f5c145103aca","Type":"ContainerStarted","Data":"afb02741a030d7ca36e1fee4413fde2d4f695ccf7f80c7624b2fd6e3270d7a1d"} Oct 02 02:06:02 crc kubenswrapper[4885]: I1002 02:06:02.223152 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31","Type":"ContainerStarted","Data":"3189eb217f435027c5682afe170fc3d3df2f758e8e7f6e08ef786038a88e19f3"} Oct 02 02:06:02 crc kubenswrapper[4885]: I1002 02:06:02.223327 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31" containerName="ceilometer-central-agent" containerID="cri-o://37415b45895365cfac8d3cda583e36f42405ad139ab5a23a98f34186e2451267" gracePeriod=30 Oct 02 02:06:02 crc kubenswrapper[4885]: I1002 02:06:02.223399 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31" containerName="ceilometer-notification-agent" containerID="cri-o://fa0d2db4bf10abe0c8e33d54ab6448de1f276f34da0265afacb000713267d8c8" gracePeriod=30 Oct 02 02:06:02 crc kubenswrapper[4885]: I1002 02:06:02.223414 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 02 02:06:02 crc kubenswrapper[4885]: I1002 02:06:02.223419 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31" containerName="sg-core" containerID="cri-o://196cd6495bbe8e9aef57fcdd9cde8457cfc188166141d4cc6cd72a97e3e3d3a0" gracePeriod=30 Oct 02 02:06:02 crc kubenswrapper[4885]: I1002 02:06:02.223492 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31" containerName="proxy-httpd" containerID="cri-o://3189eb217f435027c5682afe170fc3d3df2f758e8e7f6e08ef786038a88e19f3" gracePeriod=30 Oct 02 02:06:02 crc kubenswrapper[4885]: I1002 02:06:02.257781 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-zcpc9" podStartSLOduration=2.638838813 podStartE2EDuration="10.257756786s" podCreationTimestamp="2025-10-02 02:05:52 +0000 UTC" firstStartedPulling="2025-10-02 02:05:53.660086119 +0000 UTC m=+1142.471833548" lastFinishedPulling="2025-10-02 02:06:01.279004102 +0000 UTC m=+1150.090751521" observedRunningTime="2025-10-02 02:06:02.239874523 +0000 UTC m=+1151.051621952" watchObservedRunningTime="2025-10-02 02:06:02.257756786 +0000 UTC m=+1151.069504205" Oct 02 02:06:03 crc kubenswrapper[4885]: I1002 02:06:03.238656 4885 generic.go:334] "Generic (PLEG): container finished" podID="3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31" containerID="3189eb217f435027c5682afe170fc3d3df2f758e8e7f6e08ef786038a88e19f3" exitCode=0 Oct 02 02:06:03 crc kubenswrapper[4885]: I1002 02:06:03.238982 4885 generic.go:334] "Generic (PLEG): container finished" podID="3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31" containerID="196cd6495bbe8e9aef57fcdd9cde8457cfc188166141d4cc6cd72a97e3e3d3a0" exitCode=2 Oct 02 02:06:03 crc kubenswrapper[4885]: I1002 02:06:03.238993 4885 generic.go:334] "Generic (PLEG): container finished" podID="3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31" containerID="fa0d2db4bf10abe0c8e33d54ab6448de1f276f34da0265afacb000713267d8c8" exitCode=0 Oct 02 02:06:03 crc kubenswrapper[4885]: I1002 02:06:03.238750 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31","Type":"ContainerDied","Data":"3189eb217f435027c5682afe170fc3d3df2f758e8e7f6e08ef786038a88e19f3"} Oct 02 02:06:03 crc kubenswrapper[4885]: I1002 02:06:03.239054 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31","Type":"ContainerDied","Data":"196cd6495bbe8e9aef57fcdd9cde8457cfc188166141d4cc6cd72a97e3e3d3a0"} Oct 02 02:06:03 crc kubenswrapper[4885]: I1002 02:06:03.239078 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31","Type":"ContainerDied","Data":"fa0d2db4bf10abe0c8e33d54ab6448de1f276f34da0265afacb000713267d8c8"} Oct 02 02:06:04 crc kubenswrapper[4885]: I1002 02:06:04.253855 4885 generic.go:334] "Generic (PLEG): container finished" podID="3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31" containerID="37415b45895365cfac8d3cda583e36f42405ad139ab5a23a98f34186e2451267" exitCode=0 Oct 02 02:06:04 crc kubenswrapper[4885]: I1002 02:06:04.253931 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31","Type":"ContainerDied","Data":"37415b45895365cfac8d3cda583e36f42405ad139ab5a23a98f34186e2451267"} Oct 02 02:06:04 crc kubenswrapper[4885]: I1002 02:06:04.254137 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31","Type":"ContainerDied","Data":"6347f77e8d0ea5295e11111f03572d32eac3e9c5f0c3874548bd51be05fd980b"} Oct 02 02:06:04 crc kubenswrapper[4885]: I1002 02:06:04.254154 4885 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6347f77e8d0ea5295e11111f03572d32eac3e9c5f0c3874548bd51be05fd980b" Oct 02 02:06:04 crc kubenswrapper[4885]: I1002 02:06:04.306816 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 02:06:04 crc kubenswrapper[4885]: I1002 02:06:04.374963 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31-run-httpd\") pod \"3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31\" (UID: \"3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31\") " Oct 02 02:06:04 crc kubenswrapper[4885]: I1002 02:06:04.375018 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31-combined-ca-bundle\") pod \"3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31\" (UID: \"3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31\") " Oct 02 02:06:04 crc kubenswrapper[4885]: I1002 02:06:04.375072 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31-scripts\") pod \"3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31\" (UID: \"3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31\") " Oct 02 02:06:04 crc kubenswrapper[4885]: I1002 02:06:04.375106 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31-sg-core-conf-yaml\") pod \"3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31\" (UID: \"3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31\") " Oct 02 02:06:04 crc kubenswrapper[4885]: I1002 02:06:04.375235 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sptk8\" (UniqueName: \"kubernetes.io/projected/3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31-kube-api-access-sptk8\") pod \"3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31\" (UID: \"3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31\") " Oct 02 02:06:04 crc kubenswrapper[4885]: I1002 02:06:04.375286 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31-config-data\") pod \"3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31\" (UID: \"3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31\") " Oct 02 02:06:04 crc kubenswrapper[4885]: I1002 02:06:04.375317 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31-log-httpd\") pod \"3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31\" (UID: \"3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31\") " Oct 02 02:06:04 crc kubenswrapper[4885]: I1002 02:06:04.375568 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31" (UID: "3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:06:04 crc kubenswrapper[4885]: I1002 02:06:04.375823 4885 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 02:06:04 crc kubenswrapper[4885]: I1002 02:06:04.376127 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31" (UID: "3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:06:04 crc kubenswrapper[4885]: I1002 02:06:04.380514 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31-kube-api-access-sptk8" (OuterVolumeSpecName: "kube-api-access-sptk8") pod "3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31" (UID: "3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31"). InnerVolumeSpecName "kube-api-access-sptk8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:06:04 crc kubenswrapper[4885]: I1002 02:06:04.404435 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31-scripts" (OuterVolumeSpecName: "scripts") pod "3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31" (UID: "3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:06:04 crc kubenswrapper[4885]: I1002 02:06:04.419659 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31" (UID: "3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:06:04 crc kubenswrapper[4885]: I1002 02:06:04.468891 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31" (UID: "3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:06:04 crc kubenswrapper[4885]: I1002 02:06:04.476898 4885 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:06:04 crc kubenswrapper[4885]: I1002 02:06:04.476921 4885 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 02:06:04 crc kubenswrapper[4885]: I1002 02:06:04.476932 4885 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 02 02:06:04 crc kubenswrapper[4885]: I1002 02:06:04.476941 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sptk8\" (UniqueName: \"kubernetes.io/projected/3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31-kube-api-access-sptk8\") on node \"crc\" DevicePath \"\"" Oct 02 02:06:04 crc kubenswrapper[4885]: I1002 02:06:04.476949 4885 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 02:06:04 crc kubenswrapper[4885]: I1002 02:06:04.490167 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31-config-data" (OuterVolumeSpecName: "config-data") pod "3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31" (UID: "3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:06:04 crc kubenswrapper[4885]: I1002 02:06:04.578623 4885 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 02:06:05 crc kubenswrapper[4885]: I1002 02:06:05.265664 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 02:06:05 crc kubenswrapper[4885]: I1002 02:06:05.310249 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 02:06:05 crc kubenswrapper[4885]: I1002 02:06:05.329536 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 02 02:06:05 crc kubenswrapper[4885]: I1002 02:06:05.340746 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 02 02:06:05 crc kubenswrapper[4885]: E1002 02:06:05.341068 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31" containerName="proxy-httpd" Oct 02 02:06:05 crc kubenswrapper[4885]: I1002 02:06:05.341083 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31" containerName="proxy-httpd" Oct 02 02:06:05 crc kubenswrapper[4885]: E1002 02:06:05.341097 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31" containerName="ceilometer-central-agent" Oct 02 02:06:05 crc kubenswrapper[4885]: I1002 02:06:05.341104 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31" containerName="ceilometer-central-agent" Oct 02 02:06:05 crc kubenswrapper[4885]: E1002 02:06:05.341118 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31" containerName="sg-core" Oct 02 02:06:05 crc kubenswrapper[4885]: I1002 02:06:05.341124 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31" containerName="sg-core" Oct 02 02:06:05 crc kubenswrapper[4885]: E1002 02:06:05.341138 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31" containerName="ceilometer-notification-agent" Oct 02 02:06:05 crc kubenswrapper[4885]: I1002 02:06:05.341144 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31" containerName="ceilometer-notification-agent" Oct 02 02:06:05 crc kubenswrapper[4885]: I1002 02:06:05.341323 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31" containerName="ceilometer-central-agent" Oct 02 02:06:05 crc kubenswrapper[4885]: I1002 02:06:05.341342 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31" containerName="ceilometer-notification-agent" Oct 02 02:06:05 crc kubenswrapper[4885]: I1002 02:06:05.341354 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31" containerName="sg-core" Oct 02 02:06:05 crc kubenswrapper[4885]: I1002 02:06:05.341364 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31" containerName="proxy-httpd" Oct 02 02:06:05 crc kubenswrapper[4885]: I1002 02:06:05.342850 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 02:06:05 crc kubenswrapper[4885]: I1002 02:06:05.344737 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 02 02:06:05 crc kubenswrapper[4885]: I1002 02:06:05.346894 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 02 02:06:05 crc kubenswrapper[4885]: I1002 02:06:05.361241 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 02:06:05 crc kubenswrapper[4885]: I1002 02:06:05.399591 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7617cdd5-023a-427d-8365-52dbfc6cf20a-scripts\") pod \"ceilometer-0\" (UID: \"7617cdd5-023a-427d-8365-52dbfc6cf20a\") " pod="openstack/ceilometer-0" Oct 02 02:06:05 crc kubenswrapper[4885]: I1002 02:06:05.399772 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7617cdd5-023a-427d-8365-52dbfc6cf20a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7617cdd5-023a-427d-8365-52dbfc6cf20a\") " pod="openstack/ceilometer-0" Oct 02 02:06:05 crc kubenswrapper[4885]: I1002 02:06:05.399794 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7617cdd5-023a-427d-8365-52dbfc6cf20a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7617cdd5-023a-427d-8365-52dbfc6cf20a\") " pod="openstack/ceilometer-0" Oct 02 02:06:05 crc kubenswrapper[4885]: I1002 02:06:05.399846 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7617cdd5-023a-427d-8365-52dbfc6cf20a-config-data\") pod \"ceilometer-0\" (UID: \"7617cdd5-023a-427d-8365-52dbfc6cf20a\") " pod="openstack/ceilometer-0" Oct 02 02:06:05 crc kubenswrapper[4885]: I1002 02:06:05.399886 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pm989\" (UniqueName: \"kubernetes.io/projected/7617cdd5-023a-427d-8365-52dbfc6cf20a-kube-api-access-pm989\") pod \"ceilometer-0\" (UID: \"7617cdd5-023a-427d-8365-52dbfc6cf20a\") " pod="openstack/ceilometer-0" Oct 02 02:06:05 crc kubenswrapper[4885]: I1002 02:06:05.399916 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7617cdd5-023a-427d-8365-52dbfc6cf20a-log-httpd\") pod \"ceilometer-0\" (UID: \"7617cdd5-023a-427d-8365-52dbfc6cf20a\") " pod="openstack/ceilometer-0" Oct 02 02:06:05 crc kubenswrapper[4885]: I1002 02:06:05.399937 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7617cdd5-023a-427d-8365-52dbfc6cf20a-run-httpd\") pod \"ceilometer-0\" (UID: \"7617cdd5-023a-427d-8365-52dbfc6cf20a\") " pod="openstack/ceilometer-0" Oct 02 02:06:05 crc kubenswrapper[4885]: I1002 02:06:05.501554 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7617cdd5-023a-427d-8365-52dbfc6cf20a-config-data\") pod \"ceilometer-0\" (UID: \"7617cdd5-023a-427d-8365-52dbfc6cf20a\") " pod="openstack/ceilometer-0" Oct 02 02:06:05 crc kubenswrapper[4885]: I1002 02:06:05.501621 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pm989\" (UniqueName: \"kubernetes.io/projected/7617cdd5-023a-427d-8365-52dbfc6cf20a-kube-api-access-pm989\") pod \"ceilometer-0\" (UID: \"7617cdd5-023a-427d-8365-52dbfc6cf20a\") " pod="openstack/ceilometer-0" Oct 02 02:06:05 crc kubenswrapper[4885]: I1002 02:06:05.501657 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7617cdd5-023a-427d-8365-52dbfc6cf20a-log-httpd\") pod \"ceilometer-0\" (UID: \"7617cdd5-023a-427d-8365-52dbfc6cf20a\") " pod="openstack/ceilometer-0" Oct 02 02:06:05 crc kubenswrapper[4885]: I1002 02:06:05.501682 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7617cdd5-023a-427d-8365-52dbfc6cf20a-run-httpd\") pod \"ceilometer-0\" (UID: \"7617cdd5-023a-427d-8365-52dbfc6cf20a\") " pod="openstack/ceilometer-0" Oct 02 02:06:05 crc kubenswrapper[4885]: I1002 02:06:05.501741 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7617cdd5-023a-427d-8365-52dbfc6cf20a-scripts\") pod \"ceilometer-0\" (UID: \"7617cdd5-023a-427d-8365-52dbfc6cf20a\") " pod="openstack/ceilometer-0" Oct 02 02:06:05 crc kubenswrapper[4885]: I1002 02:06:05.501787 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7617cdd5-023a-427d-8365-52dbfc6cf20a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7617cdd5-023a-427d-8365-52dbfc6cf20a\") " pod="openstack/ceilometer-0" Oct 02 02:06:05 crc kubenswrapper[4885]: I1002 02:06:05.501802 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7617cdd5-023a-427d-8365-52dbfc6cf20a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7617cdd5-023a-427d-8365-52dbfc6cf20a\") " pod="openstack/ceilometer-0" Oct 02 02:06:05 crc kubenswrapper[4885]: I1002 02:06:05.503006 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7617cdd5-023a-427d-8365-52dbfc6cf20a-run-httpd\") pod \"ceilometer-0\" (UID: \"7617cdd5-023a-427d-8365-52dbfc6cf20a\") " pod="openstack/ceilometer-0" Oct 02 02:06:05 crc kubenswrapper[4885]: I1002 02:06:05.503040 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7617cdd5-023a-427d-8365-52dbfc6cf20a-log-httpd\") pod \"ceilometer-0\" (UID: \"7617cdd5-023a-427d-8365-52dbfc6cf20a\") " pod="openstack/ceilometer-0" Oct 02 02:06:05 crc kubenswrapper[4885]: I1002 02:06:05.506346 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7617cdd5-023a-427d-8365-52dbfc6cf20a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7617cdd5-023a-427d-8365-52dbfc6cf20a\") " pod="openstack/ceilometer-0" Oct 02 02:06:05 crc kubenswrapper[4885]: I1002 02:06:05.506655 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7617cdd5-023a-427d-8365-52dbfc6cf20a-config-data\") pod \"ceilometer-0\" (UID: \"7617cdd5-023a-427d-8365-52dbfc6cf20a\") " pod="openstack/ceilometer-0" Oct 02 02:06:05 crc kubenswrapper[4885]: I1002 02:06:05.507784 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7617cdd5-023a-427d-8365-52dbfc6cf20a-scripts\") pod \"ceilometer-0\" (UID: \"7617cdd5-023a-427d-8365-52dbfc6cf20a\") " pod="openstack/ceilometer-0" Oct 02 02:06:05 crc kubenswrapper[4885]: I1002 02:06:05.524357 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7617cdd5-023a-427d-8365-52dbfc6cf20a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7617cdd5-023a-427d-8365-52dbfc6cf20a\") " pod="openstack/ceilometer-0" Oct 02 02:06:05 crc kubenswrapper[4885]: I1002 02:06:05.541184 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pm989\" (UniqueName: \"kubernetes.io/projected/7617cdd5-023a-427d-8365-52dbfc6cf20a-kube-api-access-pm989\") pod \"ceilometer-0\" (UID: \"7617cdd5-023a-427d-8365-52dbfc6cf20a\") " pod="openstack/ceilometer-0" Oct 02 02:06:05 crc kubenswrapper[4885]: I1002 02:06:05.669212 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 02:06:06 crc kubenswrapper[4885]: I1002 02:06:06.030368 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 02:06:06 crc kubenswrapper[4885]: W1002 02:06:06.032634 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7617cdd5_023a_427d_8365_52dbfc6cf20a.slice/crio-6d9831c52506bd7eb51f0fa792de43fc2b6db5f95cba7927541cea46a4e8ee2b WatchSource:0}: Error finding container 6d9831c52506bd7eb51f0fa792de43fc2b6db5f95cba7927541cea46a4e8ee2b: Status 404 returned error can't find the container with id 6d9831c52506bd7eb51f0fa792de43fc2b6db5f95cba7927541cea46a4e8ee2b Oct 02 02:06:06 crc kubenswrapper[4885]: I1002 02:06:06.065571 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31" path="/var/lib/kubelet/pods/3a0eeb33-7f67-4d0c-9b18-3ccbf8715a31/volumes" Oct 02 02:06:06 crc kubenswrapper[4885]: I1002 02:06:06.281420 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7617cdd5-023a-427d-8365-52dbfc6cf20a","Type":"ContainerStarted","Data":"6d9831c52506bd7eb51f0fa792de43fc2b6db5f95cba7927541cea46a4e8ee2b"} Oct 02 02:06:07 crc kubenswrapper[4885]: I1002 02:06:07.304489 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7617cdd5-023a-427d-8365-52dbfc6cf20a","Type":"ContainerStarted","Data":"1a275dbd70e21d0358dfdb3b1ee5eb025a0342c7445566bddcc00930eaa277f3"} Oct 02 02:06:08 crc kubenswrapper[4885]: I1002 02:06:08.316330 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7617cdd5-023a-427d-8365-52dbfc6cf20a","Type":"ContainerStarted","Data":"4d15decee345cf96133f34e791bdd5f7eaec17200d66f3a76a057701de1be4c6"} Oct 02 02:06:08 crc kubenswrapper[4885]: I1002 02:06:08.316858 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7617cdd5-023a-427d-8365-52dbfc6cf20a","Type":"ContainerStarted","Data":"5467c6f206f6d2e7c0a79b0aa4ab9d08a4129ba7f16275ccee62fec86ad3bf2e"} Oct 02 02:06:10 crc kubenswrapper[4885]: I1002 02:06:10.353112 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7617cdd5-023a-427d-8365-52dbfc6cf20a","Type":"ContainerStarted","Data":"737b1e2d7170ba0c915e063b22c387f8c1b2dcc4d7cb99b2b87f8998b9369cbb"} Oct 02 02:06:10 crc kubenswrapper[4885]: I1002 02:06:10.354890 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 02 02:06:10 crc kubenswrapper[4885]: I1002 02:06:10.387032 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.814377441 podStartE2EDuration="5.387013873s" podCreationTimestamp="2025-10-02 02:06:05 +0000 UTC" firstStartedPulling="2025-10-02 02:06:06.035973477 +0000 UTC m=+1154.847720886" lastFinishedPulling="2025-10-02 02:06:09.608609919 +0000 UTC m=+1158.420357318" observedRunningTime="2025-10-02 02:06:10.37706564 +0000 UTC m=+1159.188813039" watchObservedRunningTime="2025-10-02 02:06:10.387013873 +0000 UTC m=+1159.198761282" Oct 02 02:06:12 crc kubenswrapper[4885]: I1002 02:06:12.376051 4885 generic.go:334] "Generic (PLEG): container finished" podID="cb797f9e-26a8-42fb-bac6-f5c145103aca" containerID="afb02741a030d7ca36e1fee4413fde2d4f695ccf7f80c7624b2fd6e3270d7a1d" exitCode=0 Oct 02 02:06:12 crc kubenswrapper[4885]: I1002 02:06:12.376117 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-zcpc9" event={"ID":"cb797f9e-26a8-42fb-bac6-f5c145103aca","Type":"ContainerDied","Data":"afb02741a030d7ca36e1fee4413fde2d4f695ccf7f80c7624b2fd6e3270d7a1d"} Oct 02 02:06:13 crc kubenswrapper[4885]: I1002 02:06:13.265385 4885 patch_prober.go:28] interesting pod/machine-config-daemon-rttx8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 02:06:13 crc kubenswrapper[4885]: I1002 02:06:13.265708 4885 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 02:06:13 crc kubenswrapper[4885]: I1002 02:06:13.762050 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-zcpc9" Oct 02 02:06:13 crc kubenswrapper[4885]: I1002 02:06:13.871150 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jj59f\" (UniqueName: \"kubernetes.io/projected/cb797f9e-26a8-42fb-bac6-f5c145103aca-kube-api-access-jj59f\") pod \"cb797f9e-26a8-42fb-bac6-f5c145103aca\" (UID: \"cb797f9e-26a8-42fb-bac6-f5c145103aca\") " Oct 02 02:06:13 crc kubenswrapper[4885]: I1002 02:06:13.871494 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb797f9e-26a8-42fb-bac6-f5c145103aca-config-data\") pod \"cb797f9e-26a8-42fb-bac6-f5c145103aca\" (UID: \"cb797f9e-26a8-42fb-bac6-f5c145103aca\") " Oct 02 02:06:13 crc kubenswrapper[4885]: I1002 02:06:13.871542 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cb797f9e-26a8-42fb-bac6-f5c145103aca-scripts\") pod \"cb797f9e-26a8-42fb-bac6-f5c145103aca\" (UID: \"cb797f9e-26a8-42fb-bac6-f5c145103aca\") " Oct 02 02:06:13 crc kubenswrapper[4885]: I1002 02:06:13.871831 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb797f9e-26a8-42fb-bac6-f5c145103aca-combined-ca-bundle\") pod \"cb797f9e-26a8-42fb-bac6-f5c145103aca\" (UID: \"cb797f9e-26a8-42fb-bac6-f5c145103aca\") " Oct 02 02:06:13 crc kubenswrapper[4885]: I1002 02:06:13.879015 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb797f9e-26a8-42fb-bac6-f5c145103aca-scripts" (OuterVolumeSpecName: "scripts") pod "cb797f9e-26a8-42fb-bac6-f5c145103aca" (UID: "cb797f9e-26a8-42fb-bac6-f5c145103aca"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:06:13 crc kubenswrapper[4885]: I1002 02:06:13.883746 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cb797f9e-26a8-42fb-bac6-f5c145103aca-kube-api-access-jj59f" (OuterVolumeSpecName: "kube-api-access-jj59f") pod "cb797f9e-26a8-42fb-bac6-f5c145103aca" (UID: "cb797f9e-26a8-42fb-bac6-f5c145103aca"). InnerVolumeSpecName "kube-api-access-jj59f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:06:13 crc kubenswrapper[4885]: I1002 02:06:13.901084 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb797f9e-26a8-42fb-bac6-f5c145103aca-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cb797f9e-26a8-42fb-bac6-f5c145103aca" (UID: "cb797f9e-26a8-42fb-bac6-f5c145103aca"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:06:13 crc kubenswrapper[4885]: I1002 02:06:13.913389 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb797f9e-26a8-42fb-bac6-f5c145103aca-config-data" (OuterVolumeSpecName: "config-data") pod "cb797f9e-26a8-42fb-bac6-f5c145103aca" (UID: "cb797f9e-26a8-42fb-bac6-f5c145103aca"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:06:13 crc kubenswrapper[4885]: I1002 02:06:13.974495 4885 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb797f9e-26a8-42fb-bac6-f5c145103aca-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:06:13 crc kubenswrapper[4885]: I1002 02:06:13.974638 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jj59f\" (UniqueName: \"kubernetes.io/projected/cb797f9e-26a8-42fb-bac6-f5c145103aca-kube-api-access-jj59f\") on node \"crc\" DevicePath \"\"" Oct 02 02:06:13 crc kubenswrapper[4885]: I1002 02:06:13.974700 4885 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb797f9e-26a8-42fb-bac6-f5c145103aca-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 02:06:13 crc kubenswrapper[4885]: I1002 02:06:13.974754 4885 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cb797f9e-26a8-42fb-bac6-f5c145103aca-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 02:06:14 crc kubenswrapper[4885]: I1002 02:06:14.399676 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-zcpc9" event={"ID":"cb797f9e-26a8-42fb-bac6-f5c145103aca","Type":"ContainerDied","Data":"0bc418da516fd37b6a50f644e58bfc862a4f2cc8c20bba69c776533cf93eb76c"} Oct 02 02:06:14 crc kubenswrapper[4885]: I1002 02:06:14.399746 4885 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0bc418da516fd37b6a50f644e58bfc862a4f2cc8c20bba69c776533cf93eb76c" Oct 02 02:06:14 crc kubenswrapper[4885]: I1002 02:06:14.399842 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-zcpc9" Oct 02 02:06:14 crc kubenswrapper[4885]: I1002 02:06:14.532987 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 02 02:06:14 crc kubenswrapper[4885]: E1002 02:06:14.533400 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb797f9e-26a8-42fb-bac6-f5c145103aca" containerName="nova-cell0-conductor-db-sync" Oct 02 02:06:14 crc kubenswrapper[4885]: I1002 02:06:14.533420 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb797f9e-26a8-42fb-bac6-f5c145103aca" containerName="nova-cell0-conductor-db-sync" Oct 02 02:06:14 crc kubenswrapper[4885]: I1002 02:06:14.533584 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb797f9e-26a8-42fb-bac6-f5c145103aca" containerName="nova-cell0-conductor-db-sync" Oct 02 02:06:14 crc kubenswrapper[4885]: I1002 02:06:14.534120 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 02 02:06:14 crc kubenswrapper[4885]: I1002 02:06:14.537121 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-fxj22" Oct 02 02:06:14 crc kubenswrapper[4885]: I1002 02:06:14.537511 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 02 02:06:14 crc kubenswrapper[4885]: I1002 02:06:14.550377 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 02 02:06:14 crc kubenswrapper[4885]: I1002 02:06:14.584389 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p694l\" (UniqueName: \"kubernetes.io/projected/504d46bb-b842-4362-9a00-4917a01cf509-kube-api-access-p694l\") pod \"nova-cell0-conductor-0\" (UID: \"504d46bb-b842-4362-9a00-4917a01cf509\") " pod="openstack/nova-cell0-conductor-0" Oct 02 02:06:14 crc kubenswrapper[4885]: I1002 02:06:14.584420 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/504d46bb-b842-4362-9a00-4917a01cf509-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"504d46bb-b842-4362-9a00-4917a01cf509\") " pod="openstack/nova-cell0-conductor-0" Oct 02 02:06:14 crc kubenswrapper[4885]: I1002 02:06:14.584449 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/504d46bb-b842-4362-9a00-4917a01cf509-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"504d46bb-b842-4362-9a00-4917a01cf509\") " pod="openstack/nova-cell0-conductor-0" Oct 02 02:06:14 crc kubenswrapper[4885]: I1002 02:06:14.686171 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p694l\" (UniqueName: \"kubernetes.io/projected/504d46bb-b842-4362-9a00-4917a01cf509-kube-api-access-p694l\") pod \"nova-cell0-conductor-0\" (UID: \"504d46bb-b842-4362-9a00-4917a01cf509\") " pod="openstack/nova-cell0-conductor-0" Oct 02 02:06:14 crc kubenswrapper[4885]: I1002 02:06:14.686473 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/504d46bb-b842-4362-9a00-4917a01cf509-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"504d46bb-b842-4362-9a00-4917a01cf509\") " pod="openstack/nova-cell0-conductor-0" Oct 02 02:06:14 crc kubenswrapper[4885]: I1002 02:06:14.686500 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/504d46bb-b842-4362-9a00-4917a01cf509-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"504d46bb-b842-4362-9a00-4917a01cf509\") " pod="openstack/nova-cell0-conductor-0" Oct 02 02:06:14 crc kubenswrapper[4885]: I1002 02:06:14.692982 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/504d46bb-b842-4362-9a00-4917a01cf509-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"504d46bb-b842-4362-9a00-4917a01cf509\") " pod="openstack/nova-cell0-conductor-0" Oct 02 02:06:14 crc kubenswrapper[4885]: I1002 02:06:14.693855 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/504d46bb-b842-4362-9a00-4917a01cf509-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"504d46bb-b842-4362-9a00-4917a01cf509\") " pod="openstack/nova-cell0-conductor-0" Oct 02 02:06:14 crc kubenswrapper[4885]: I1002 02:06:14.706850 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p694l\" (UniqueName: \"kubernetes.io/projected/504d46bb-b842-4362-9a00-4917a01cf509-kube-api-access-p694l\") pod \"nova-cell0-conductor-0\" (UID: \"504d46bb-b842-4362-9a00-4917a01cf509\") " pod="openstack/nova-cell0-conductor-0" Oct 02 02:06:14 crc kubenswrapper[4885]: I1002 02:06:14.888176 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 02 02:06:15 crc kubenswrapper[4885]: I1002 02:06:15.183043 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 02 02:06:15 crc kubenswrapper[4885]: I1002 02:06:15.410588 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"504d46bb-b842-4362-9a00-4917a01cf509","Type":"ContainerStarted","Data":"9dae1075678c017d1b9cdab3c6255e7f76a310fe3fb19c8f7a1f44e50c103a85"} Oct 02 02:06:15 crc kubenswrapper[4885]: I1002 02:06:15.410920 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"504d46bb-b842-4362-9a00-4917a01cf509","Type":"ContainerStarted","Data":"8e69d994d5fc6bd2cfc5ac760cedd806f8b4effb3c99e2e931dd1aa550ba9c9b"} Oct 02 02:06:15 crc kubenswrapper[4885]: I1002 02:06:15.410950 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Oct 02 02:06:15 crc kubenswrapper[4885]: I1002 02:06:15.442500 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=1.4424772940000001 podStartE2EDuration="1.442477294s" podCreationTimestamp="2025-10-02 02:06:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:06:15.432373126 +0000 UTC m=+1164.244120555" watchObservedRunningTime="2025-10-02 02:06:15.442477294 +0000 UTC m=+1164.254224723" Oct 02 02:06:24 crc kubenswrapper[4885]: I1002 02:06:24.932017 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Oct 02 02:06:25 crc kubenswrapper[4885]: I1002 02:06:25.648048 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-t94gt"] Oct 02 02:06:25 crc kubenswrapper[4885]: I1002 02:06:25.649223 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-t94gt" Oct 02 02:06:25 crc kubenswrapper[4885]: I1002 02:06:25.651280 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Oct 02 02:06:25 crc kubenswrapper[4885]: I1002 02:06:25.651282 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Oct 02 02:06:25 crc kubenswrapper[4885]: I1002 02:06:25.694473 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-t94gt"] Oct 02 02:06:25 crc kubenswrapper[4885]: I1002 02:06:25.742197 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-prjhc\" (UniqueName: \"kubernetes.io/projected/50bc98ab-1629-4591-8e85-bfa8f17962dc-kube-api-access-prjhc\") pod \"nova-cell0-cell-mapping-t94gt\" (UID: \"50bc98ab-1629-4591-8e85-bfa8f17962dc\") " pod="openstack/nova-cell0-cell-mapping-t94gt" Oct 02 02:06:25 crc kubenswrapper[4885]: I1002 02:06:25.742337 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50bc98ab-1629-4591-8e85-bfa8f17962dc-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-t94gt\" (UID: \"50bc98ab-1629-4591-8e85-bfa8f17962dc\") " pod="openstack/nova-cell0-cell-mapping-t94gt" Oct 02 02:06:25 crc kubenswrapper[4885]: I1002 02:06:25.742358 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50bc98ab-1629-4591-8e85-bfa8f17962dc-config-data\") pod \"nova-cell0-cell-mapping-t94gt\" (UID: \"50bc98ab-1629-4591-8e85-bfa8f17962dc\") " pod="openstack/nova-cell0-cell-mapping-t94gt" Oct 02 02:06:25 crc kubenswrapper[4885]: I1002 02:06:25.742436 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/50bc98ab-1629-4591-8e85-bfa8f17962dc-scripts\") pod \"nova-cell0-cell-mapping-t94gt\" (UID: \"50bc98ab-1629-4591-8e85-bfa8f17962dc\") " pod="openstack/nova-cell0-cell-mapping-t94gt" Oct 02 02:06:25 crc kubenswrapper[4885]: I1002 02:06:25.817014 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 02 02:06:25 crc kubenswrapper[4885]: I1002 02:06:25.818481 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 02:06:25 crc kubenswrapper[4885]: I1002 02:06:25.821628 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 02 02:06:25 crc kubenswrapper[4885]: I1002 02:06:25.842100 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 02 02:06:25 crc kubenswrapper[4885]: I1002 02:06:25.844299 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/50bc98ab-1629-4591-8e85-bfa8f17962dc-scripts\") pod \"nova-cell0-cell-mapping-t94gt\" (UID: \"50bc98ab-1629-4591-8e85-bfa8f17962dc\") " pod="openstack/nova-cell0-cell-mapping-t94gt" Oct 02 02:06:25 crc kubenswrapper[4885]: I1002 02:06:25.844471 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-prjhc\" (UniqueName: \"kubernetes.io/projected/50bc98ab-1629-4591-8e85-bfa8f17962dc-kube-api-access-prjhc\") pod \"nova-cell0-cell-mapping-t94gt\" (UID: \"50bc98ab-1629-4591-8e85-bfa8f17962dc\") " pod="openstack/nova-cell0-cell-mapping-t94gt" Oct 02 02:06:25 crc kubenswrapper[4885]: I1002 02:06:25.844530 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50bc98ab-1629-4591-8e85-bfa8f17962dc-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-t94gt\" (UID: \"50bc98ab-1629-4591-8e85-bfa8f17962dc\") " pod="openstack/nova-cell0-cell-mapping-t94gt" Oct 02 02:06:25 crc kubenswrapper[4885]: I1002 02:06:25.844551 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50bc98ab-1629-4591-8e85-bfa8f17962dc-config-data\") pod \"nova-cell0-cell-mapping-t94gt\" (UID: \"50bc98ab-1629-4591-8e85-bfa8f17962dc\") " pod="openstack/nova-cell0-cell-mapping-t94gt" Oct 02 02:06:25 crc kubenswrapper[4885]: I1002 02:06:25.852484 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50bc98ab-1629-4591-8e85-bfa8f17962dc-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-t94gt\" (UID: \"50bc98ab-1629-4591-8e85-bfa8f17962dc\") " pod="openstack/nova-cell0-cell-mapping-t94gt" Oct 02 02:06:25 crc kubenswrapper[4885]: I1002 02:06:25.859339 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50bc98ab-1629-4591-8e85-bfa8f17962dc-config-data\") pod \"nova-cell0-cell-mapping-t94gt\" (UID: \"50bc98ab-1629-4591-8e85-bfa8f17962dc\") " pod="openstack/nova-cell0-cell-mapping-t94gt" Oct 02 02:06:25 crc kubenswrapper[4885]: I1002 02:06:25.867000 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/50bc98ab-1629-4591-8e85-bfa8f17962dc-scripts\") pod \"nova-cell0-cell-mapping-t94gt\" (UID: \"50bc98ab-1629-4591-8e85-bfa8f17962dc\") " pod="openstack/nova-cell0-cell-mapping-t94gt" Oct 02 02:06:25 crc kubenswrapper[4885]: I1002 02:06:25.887005 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-prjhc\" (UniqueName: \"kubernetes.io/projected/50bc98ab-1629-4591-8e85-bfa8f17962dc-kube-api-access-prjhc\") pod \"nova-cell0-cell-mapping-t94gt\" (UID: \"50bc98ab-1629-4591-8e85-bfa8f17962dc\") " pod="openstack/nova-cell0-cell-mapping-t94gt" Oct 02 02:06:25 crc kubenswrapper[4885]: I1002 02:06:25.947953 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8dba3ac0-e4cf-4d37-baf4-952ff07f8f3f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"8dba3ac0-e4cf-4d37-baf4-952ff07f8f3f\") " pod="openstack/nova-api-0" Oct 02 02:06:25 crc kubenswrapper[4885]: I1002 02:06:25.948084 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8dba3ac0-e4cf-4d37-baf4-952ff07f8f3f-logs\") pod \"nova-api-0\" (UID: \"8dba3ac0-e4cf-4d37-baf4-952ff07f8f3f\") " pod="openstack/nova-api-0" Oct 02 02:06:25 crc kubenswrapper[4885]: I1002 02:06:25.948122 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8dba3ac0-e4cf-4d37-baf4-952ff07f8f3f-config-data\") pod \"nova-api-0\" (UID: \"8dba3ac0-e4cf-4d37-baf4-952ff07f8f3f\") " pod="openstack/nova-api-0" Oct 02 02:06:25 crc kubenswrapper[4885]: I1002 02:06:25.948146 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dgc5x\" (UniqueName: \"kubernetes.io/projected/8dba3ac0-e4cf-4d37-baf4-952ff07f8f3f-kube-api-access-dgc5x\") pod \"nova-api-0\" (UID: \"8dba3ac0-e4cf-4d37-baf4-952ff07f8f3f\") " pod="openstack/nova-api-0" Oct 02 02:06:25 crc kubenswrapper[4885]: I1002 02:06:25.967220 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 02 02:06:25 crc kubenswrapper[4885]: I1002 02:06:25.968948 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 02:06:25 crc kubenswrapper[4885]: I1002 02:06:25.982489 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 02:06:25 crc kubenswrapper[4885]: I1002 02:06:25.983694 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 02:06:25 crc kubenswrapper[4885]: I1002 02:06:25.990966 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 02 02:06:25 crc kubenswrapper[4885]: I1002 02:06:25.991757 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-t94gt" Oct 02 02:06:26 crc kubenswrapper[4885]: I1002 02:06:26.003279 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 02 02:06:26 crc kubenswrapper[4885]: I1002 02:06:26.020794 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 02:06:26 crc kubenswrapper[4885]: I1002 02:06:26.056010 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c5cf6dd-9dce-4990-a808-c25ee4041ce1-config-data\") pod \"nova-scheduler-0\" (UID: \"9c5cf6dd-9dce-4990-a808-c25ee4041ce1\") " pod="openstack/nova-scheduler-0" Oct 02 02:06:26 crc kubenswrapper[4885]: I1002 02:06:26.056064 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8dba3ac0-e4cf-4d37-baf4-952ff07f8f3f-config-data\") pod \"nova-api-0\" (UID: \"8dba3ac0-e4cf-4d37-baf4-952ff07f8f3f\") " pod="openstack/nova-api-0" Oct 02 02:06:26 crc kubenswrapper[4885]: I1002 02:06:26.056100 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dgc5x\" (UniqueName: \"kubernetes.io/projected/8dba3ac0-e4cf-4d37-baf4-952ff07f8f3f-kube-api-access-dgc5x\") pod \"nova-api-0\" (UID: \"8dba3ac0-e4cf-4d37-baf4-952ff07f8f3f\") " pod="openstack/nova-api-0" Oct 02 02:06:26 crc kubenswrapper[4885]: I1002 02:06:26.056132 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/988a6a93-aa98-4ab0-bad2-ec73966f2055-logs\") pod \"nova-metadata-0\" (UID: \"988a6a93-aa98-4ab0-bad2-ec73966f2055\") " pod="openstack/nova-metadata-0" Oct 02 02:06:26 crc kubenswrapper[4885]: I1002 02:06:26.056155 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/988a6a93-aa98-4ab0-bad2-ec73966f2055-config-data\") pod \"nova-metadata-0\" (UID: \"988a6a93-aa98-4ab0-bad2-ec73966f2055\") " pod="openstack/nova-metadata-0" Oct 02 02:06:26 crc kubenswrapper[4885]: I1002 02:06:26.056191 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8dba3ac0-e4cf-4d37-baf4-952ff07f8f3f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"8dba3ac0-e4cf-4d37-baf4-952ff07f8f3f\") " pod="openstack/nova-api-0" Oct 02 02:06:26 crc kubenswrapper[4885]: I1002 02:06:26.056241 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/988a6a93-aa98-4ab0-bad2-ec73966f2055-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"988a6a93-aa98-4ab0-bad2-ec73966f2055\") " pod="openstack/nova-metadata-0" Oct 02 02:06:26 crc kubenswrapper[4885]: I1002 02:06:26.056283 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rvjfk\" (UniqueName: \"kubernetes.io/projected/988a6a93-aa98-4ab0-bad2-ec73966f2055-kube-api-access-rvjfk\") pod \"nova-metadata-0\" (UID: \"988a6a93-aa98-4ab0-bad2-ec73966f2055\") " pod="openstack/nova-metadata-0" Oct 02 02:06:26 crc kubenswrapper[4885]: I1002 02:06:26.056322 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w9pz7\" (UniqueName: \"kubernetes.io/projected/9c5cf6dd-9dce-4990-a808-c25ee4041ce1-kube-api-access-w9pz7\") pod \"nova-scheduler-0\" (UID: \"9c5cf6dd-9dce-4990-a808-c25ee4041ce1\") " pod="openstack/nova-scheduler-0" Oct 02 02:06:26 crc kubenswrapper[4885]: I1002 02:06:26.056369 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c5cf6dd-9dce-4990-a808-c25ee4041ce1-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"9c5cf6dd-9dce-4990-a808-c25ee4041ce1\") " pod="openstack/nova-scheduler-0" Oct 02 02:06:26 crc kubenswrapper[4885]: I1002 02:06:26.056421 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8dba3ac0-e4cf-4d37-baf4-952ff07f8f3f-logs\") pod \"nova-api-0\" (UID: \"8dba3ac0-e4cf-4d37-baf4-952ff07f8f3f\") " pod="openstack/nova-api-0" Oct 02 02:06:26 crc kubenswrapper[4885]: I1002 02:06:26.056868 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8dba3ac0-e4cf-4d37-baf4-952ff07f8f3f-logs\") pod \"nova-api-0\" (UID: \"8dba3ac0-e4cf-4d37-baf4-952ff07f8f3f\") " pod="openstack/nova-api-0" Oct 02 02:06:26 crc kubenswrapper[4885]: I1002 02:06:26.078408 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 02:06:26 crc kubenswrapper[4885]: I1002 02:06:26.106996 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dgc5x\" (UniqueName: \"kubernetes.io/projected/8dba3ac0-e4cf-4d37-baf4-952ff07f8f3f-kube-api-access-dgc5x\") pod \"nova-api-0\" (UID: \"8dba3ac0-e4cf-4d37-baf4-952ff07f8f3f\") " pod="openstack/nova-api-0" Oct 02 02:06:26 crc kubenswrapper[4885]: I1002 02:06:26.113867 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8dba3ac0-e4cf-4d37-baf4-952ff07f8f3f-config-data\") pod \"nova-api-0\" (UID: \"8dba3ac0-e4cf-4d37-baf4-952ff07f8f3f\") " pod="openstack/nova-api-0" Oct 02 02:06:26 crc kubenswrapper[4885]: I1002 02:06:26.122328 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 02:06:26 crc kubenswrapper[4885]: I1002 02:06:26.123224 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8dba3ac0-e4cf-4d37-baf4-952ff07f8f3f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"8dba3ac0-e4cf-4d37-baf4-952ff07f8f3f\") " pod="openstack/nova-api-0" Oct 02 02:06:26 crc kubenswrapper[4885]: I1002 02:06:26.123598 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 02 02:06:26 crc kubenswrapper[4885]: I1002 02:06:26.131781 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 02 02:06:26 crc kubenswrapper[4885]: I1002 02:06:26.141657 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 02:06:26 crc kubenswrapper[4885]: I1002 02:06:26.146891 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 02:06:26 crc kubenswrapper[4885]: I1002 02:06:26.157646 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c5cf6dd-9dce-4990-a808-c25ee4041ce1-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"9c5cf6dd-9dce-4990-a808-c25ee4041ce1\") " pod="openstack/nova-scheduler-0" Oct 02 02:06:26 crc kubenswrapper[4885]: I1002 02:06:26.157726 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f450788-51fa-430e-a52e-770d02dd9ff0-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"3f450788-51fa-430e-a52e-770d02dd9ff0\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 02:06:26 crc kubenswrapper[4885]: I1002 02:06:26.157747 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mxkg6\" (UniqueName: \"kubernetes.io/projected/3f450788-51fa-430e-a52e-770d02dd9ff0-kube-api-access-mxkg6\") pod \"nova-cell1-novncproxy-0\" (UID: \"3f450788-51fa-430e-a52e-770d02dd9ff0\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 02:06:26 crc kubenswrapper[4885]: I1002 02:06:26.157784 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c5cf6dd-9dce-4990-a808-c25ee4041ce1-config-data\") pod \"nova-scheduler-0\" (UID: \"9c5cf6dd-9dce-4990-a808-c25ee4041ce1\") " pod="openstack/nova-scheduler-0" Oct 02 02:06:26 crc kubenswrapper[4885]: I1002 02:06:26.157825 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/988a6a93-aa98-4ab0-bad2-ec73966f2055-logs\") pod \"nova-metadata-0\" (UID: \"988a6a93-aa98-4ab0-bad2-ec73966f2055\") " pod="openstack/nova-metadata-0" Oct 02 02:06:26 crc kubenswrapper[4885]: I1002 02:06:26.157844 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/988a6a93-aa98-4ab0-bad2-ec73966f2055-config-data\") pod \"nova-metadata-0\" (UID: \"988a6a93-aa98-4ab0-bad2-ec73966f2055\") " pod="openstack/nova-metadata-0" Oct 02 02:06:26 crc kubenswrapper[4885]: I1002 02:06:26.157892 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/988a6a93-aa98-4ab0-bad2-ec73966f2055-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"988a6a93-aa98-4ab0-bad2-ec73966f2055\") " pod="openstack/nova-metadata-0" Oct 02 02:06:26 crc kubenswrapper[4885]: I1002 02:06:26.157917 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rvjfk\" (UniqueName: \"kubernetes.io/projected/988a6a93-aa98-4ab0-bad2-ec73966f2055-kube-api-access-rvjfk\") pod \"nova-metadata-0\" (UID: \"988a6a93-aa98-4ab0-bad2-ec73966f2055\") " pod="openstack/nova-metadata-0" Oct 02 02:06:26 crc kubenswrapper[4885]: I1002 02:06:26.157948 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w9pz7\" (UniqueName: \"kubernetes.io/projected/9c5cf6dd-9dce-4990-a808-c25ee4041ce1-kube-api-access-w9pz7\") pod \"nova-scheduler-0\" (UID: \"9c5cf6dd-9dce-4990-a808-c25ee4041ce1\") " pod="openstack/nova-scheduler-0" Oct 02 02:06:26 crc kubenswrapper[4885]: I1002 02:06:26.157965 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f450788-51fa-430e-a52e-770d02dd9ff0-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"3f450788-51fa-430e-a52e-770d02dd9ff0\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 02:06:26 crc kubenswrapper[4885]: I1002 02:06:26.160881 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/988a6a93-aa98-4ab0-bad2-ec73966f2055-logs\") pod \"nova-metadata-0\" (UID: \"988a6a93-aa98-4ab0-bad2-ec73966f2055\") " pod="openstack/nova-metadata-0" Oct 02 02:06:26 crc kubenswrapper[4885]: I1002 02:06:26.167960 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/988a6a93-aa98-4ab0-bad2-ec73966f2055-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"988a6a93-aa98-4ab0-bad2-ec73966f2055\") " pod="openstack/nova-metadata-0" Oct 02 02:06:26 crc kubenswrapper[4885]: I1002 02:06:26.177889 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/988a6a93-aa98-4ab0-bad2-ec73966f2055-config-data\") pod \"nova-metadata-0\" (UID: \"988a6a93-aa98-4ab0-bad2-ec73966f2055\") " pod="openstack/nova-metadata-0" Oct 02 02:06:26 crc kubenswrapper[4885]: I1002 02:06:26.187650 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-xm97j"] Oct 02 02:06:26 crc kubenswrapper[4885]: I1002 02:06:26.189190 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-xm97j" Oct 02 02:06:26 crc kubenswrapper[4885]: I1002 02:06:26.190420 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c5cf6dd-9dce-4990-a808-c25ee4041ce1-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"9c5cf6dd-9dce-4990-a808-c25ee4041ce1\") " pod="openstack/nova-scheduler-0" Oct 02 02:06:26 crc kubenswrapper[4885]: I1002 02:06:26.193865 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c5cf6dd-9dce-4990-a808-c25ee4041ce1-config-data\") pod \"nova-scheduler-0\" (UID: \"9c5cf6dd-9dce-4990-a808-c25ee4041ce1\") " pod="openstack/nova-scheduler-0" Oct 02 02:06:26 crc kubenswrapper[4885]: I1002 02:06:26.202759 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rvjfk\" (UniqueName: \"kubernetes.io/projected/988a6a93-aa98-4ab0-bad2-ec73966f2055-kube-api-access-rvjfk\") pod \"nova-metadata-0\" (UID: \"988a6a93-aa98-4ab0-bad2-ec73966f2055\") " pod="openstack/nova-metadata-0" Oct 02 02:06:26 crc kubenswrapper[4885]: I1002 02:06:26.203963 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w9pz7\" (UniqueName: \"kubernetes.io/projected/9c5cf6dd-9dce-4990-a808-c25ee4041ce1-kube-api-access-w9pz7\") pod \"nova-scheduler-0\" (UID: \"9c5cf6dd-9dce-4990-a808-c25ee4041ce1\") " pod="openstack/nova-scheduler-0" Oct 02 02:06:26 crc kubenswrapper[4885]: I1002 02:06:26.211762 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-xm97j"] Oct 02 02:06:26 crc kubenswrapper[4885]: I1002 02:06:26.260640 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e8d78e70-43db-4890-9351-f23f53116625-ovsdbserver-nb\") pod \"dnsmasq-dns-845d6d6f59-xm97j\" (UID: \"e8d78e70-43db-4890-9351-f23f53116625\") " pod="openstack/dnsmasq-dns-845d6d6f59-xm97j" Oct 02 02:06:26 crc kubenswrapper[4885]: I1002 02:06:26.260693 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f450788-51fa-430e-a52e-770d02dd9ff0-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"3f450788-51fa-430e-a52e-770d02dd9ff0\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 02:06:26 crc kubenswrapper[4885]: I1002 02:06:26.260716 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e8d78e70-43db-4890-9351-f23f53116625-dns-swift-storage-0\") pod \"dnsmasq-dns-845d6d6f59-xm97j\" (UID: \"e8d78e70-43db-4890-9351-f23f53116625\") " pod="openstack/dnsmasq-dns-845d6d6f59-xm97j" Oct 02 02:06:26 crc kubenswrapper[4885]: I1002 02:06:26.260761 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dwd7r\" (UniqueName: \"kubernetes.io/projected/e8d78e70-43db-4890-9351-f23f53116625-kube-api-access-dwd7r\") pod \"dnsmasq-dns-845d6d6f59-xm97j\" (UID: \"e8d78e70-43db-4890-9351-f23f53116625\") " pod="openstack/dnsmasq-dns-845d6d6f59-xm97j" Oct 02 02:06:26 crc kubenswrapper[4885]: I1002 02:06:26.260810 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f450788-51fa-430e-a52e-770d02dd9ff0-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"3f450788-51fa-430e-a52e-770d02dd9ff0\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 02:06:26 crc kubenswrapper[4885]: I1002 02:06:26.260833 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mxkg6\" (UniqueName: \"kubernetes.io/projected/3f450788-51fa-430e-a52e-770d02dd9ff0-kube-api-access-mxkg6\") pod \"nova-cell1-novncproxy-0\" (UID: \"3f450788-51fa-430e-a52e-770d02dd9ff0\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 02:06:26 crc kubenswrapper[4885]: I1002 02:06:26.260857 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8d78e70-43db-4890-9351-f23f53116625-config\") pod \"dnsmasq-dns-845d6d6f59-xm97j\" (UID: \"e8d78e70-43db-4890-9351-f23f53116625\") " pod="openstack/dnsmasq-dns-845d6d6f59-xm97j" Oct 02 02:06:26 crc kubenswrapper[4885]: I1002 02:06:26.260875 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e8d78e70-43db-4890-9351-f23f53116625-dns-svc\") pod \"dnsmasq-dns-845d6d6f59-xm97j\" (UID: \"e8d78e70-43db-4890-9351-f23f53116625\") " pod="openstack/dnsmasq-dns-845d6d6f59-xm97j" Oct 02 02:06:26 crc kubenswrapper[4885]: I1002 02:06:26.260902 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e8d78e70-43db-4890-9351-f23f53116625-ovsdbserver-sb\") pod \"dnsmasq-dns-845d6d6f59-xm97j\" (UID: \"e8d78e70-43db-4890-9351-f23f53116625\") " pod="openstack/dnsmasq-dns-845d6d6f59-xm97j" Oct 02 02:06:26 crc kubenswrapper[4885]: I1002 02:06:26.264243 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f450788-51fa-430e-a52e-770d02dd9ff0-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"3f450788-51fa-430e-a52e-770d02dd9ff0\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 02:06:26 crc kubenswrapper[4885]: I1002 02:06:26.265859 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f450788-51fa-430e-a52e-770d02dd9ff0-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"3f450788-51fa-430e-a52e-770d02dd9ff0\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 02:06:26 crc kubenswrapper[4885]: I1002 02:06:26.284037 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mxkg6\" (UniqueName: \"kubernetes.io/projected/3f450788-51fa-430e-a52e-770d02dd9ff0-kube-api-access-mxkg6\") pod \"nova-cell1-novncproxy-0\" (UID: \"3f450788-51fa-430e-a52e-770d02dd9ff0\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 02:06:26 crc kubenswrapper[4885]: I1002 02:06:26.286698 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 02:06:26 crc kubenswrapper[4885]: I1002 02:06:26.314436 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 02:06:26 crc kubenswrapper[4885]: I1002 02:06:26.362469 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8d78e70-43db-4890-9351-f23f53116625-config\") pod \"dnsmasq-dns-845d6d6f59-xm97j\" (UID: \"e8d78e70-43db-4890-9351-f23f53116625\") " pod="openstack/dnsmasq-dns-845d6d6f59-xm97j" Oct 02 02:06:26 crc kubenswrapper[4885]: I1002 02:06:26.362939 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e8d78e70-43db-4890-9351-f23f53116625-dns-svc\") pod \"dnsmasq-dns-845d6d6f59-xm97j\" (UID: \"e8d78e70-43db-4890-9351-f23f53116625\") " pod="openstack/dnsmasq-dns-845d6d6f59-xm97j" Oct 02 02:06:26 crc kubenswrapper[4885]: I1002 02:06:26.362969 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e8d78e70-43db-4890-9351-f23f53116625-ovsdbserver-sb\") pod \"dnsmasq-dns-845d6d6f59-xm97j\" (UID: \"e8d78e70-43db-4890-9351-f23f53116625\") " pod="openstack/dnsmasq-dns-845d6d6f59-xm97j" Oct 02 02:06:26 crc kubenswrapper[4885]: I1002 02:06:26.363095 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e8d78e70-43db-4890-9351-f23f53116625-ovsdbserver-nb\") pod \"dnsmasq-dns-845d6d6f59-xm97j\" (UID: \"e8d78e70-43db-4890-9351-f23f53116625\") " pod="openstack/dnsmasq-dns-845d6d6f59-xm97j" Oct 02 02:06:26 crc kubenswrapper[4885]: I1002 02:06:26.363129 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e8d78e70-43db-4890-9351-f23f53116625-dns-swift-storage-0\") pod \"dnsmasq-dns-845d6d6f59-xm97j\" (UID: \"e8d78e70-43db-4890-9351-f23f53116625\") " pod="openstack/dnsmasq-dns-845d6d6f59-xm97j" Oct 02 02:06:26 crc kubenswrapper[4885]: I1002 02:06:26.363207 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dwd7r\" (UniqueName: \"kubernetes.io/projected/e8d78e70-43db-4890-9351-f23f53116625-kube-api-access-dwd7r\") pod \"dnsmasq-dns-845d6d6f59-xm97j\" (UID: \"e8d78e70-43db-4890-9351-f23f53116625\") " pod="openstack/dnsmasq-dns-845d6d6f59-xm97j" Oct 02 02:06:26 crc kubenswrapper[4885]: I1002 02:06:26.363376 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8d78e70-43db-4890-9351-f23f53116625-config\") pod \"dnsmasq-dns-845d6d6f59-xm97j\" (UID: \"e8d78e70-43db-4890-9351-f23f53116625\") " pod="openstack/dnsmasq-dns-845d6d6f59-xm97j" Oct 02 02:06:26 crc kubenswrapper[4885]: I1002 02:06:26.364238 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e8d78e70-43db-4890-9351-f23f53116625-ovsdbserver-sb\") pod \"dnsmasq-dns-845d6d6f59-xm97j\" (UID: \"e8d78e70-43db-4890-9351-f23f53116625\") " pod="openstack/dnsmasq-dns-845d6d6f59-xm97j" Oct 02 02:06:26 crc kubenswrapper[4885]: I1002 02:06:26.364654 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e8d78e70-43db-4890-9351-f23f53116625-ovsdbserver-nb\") pod \"dnsmasq-dns-845d6d6f59-xm97j\" (UID: \"e8d78e70-43db-4890-9351-f23f53116625\") " pod="openstack/dnsmasq-dns-845d6d6f59-xm97j" Oct 02 02:06:26 crc kubenswrapper[4885]: I1002 02:06:26.365442 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e8d78e70-43db-4890-9351-f23f53116625-dns-swift-storage-0\") pod \"dnsmasq-dns-845d6d6f59-xm97j\" (UID: \"e8d78e70-43db-4890-9351-f23f53116625\") " pod="openstack/dnsmasq-dns-845d6d6f59-xm97j" Oct 02 02:06:26 crc kubenswrapper[4885]: I1002 02:06:26.365811 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e8d78e70-43db-4890-9351-f23f53116625-dns-svc\") pod \"dnsmasq-dns-845d6d6f59-xm97j\" (UID: \"e8d78e70-43db-4890-9351-f23f53116625\") " pod="openstack/dnsmasq-dns-845d6d6f59-xm97j" Oct 02 02:06:26 crc kubenswrapper[4885]: I1002 02:06:26.410056 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dwd7r\" (UniqueName: \"kubernetes.io/projected/e8d78e70-43db-4890-9351-f23f53116625-kube-api-access-dwd7r\") pod \"dnsmasq-dns-845d6d6f59-xm97j\" (UID: \"e8d78e70-43db-4890-9351-f23f53116625\") " pod="openstack/dnsmasq-dns-845d6d6f59-xm97j" Oct 02 02:06:26 crc kubenswrapper[4885]: I1002 02:06:26.509763 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 02 02:06:26 crc kubenswrapper[4885]: I1002 02:06:26.514942 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-xm97j" Oct 02 02:06:26 crc kubenswrapper[4885]: I1002 02:06:26.593208 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-t94gt"] Oct 02 02:06:26 crc kubenswrapper[4885]: I1002 02:06:26.835693 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 02 02:06:26 crc kubenswrapper[4885]: W1002 02:06:26.845801 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8dba3ac0_e4cf_4d37_baf4_952ff07f8f3f.slice/crio-63f31a77904c9eefacf8c288ec0c63864af11d7253b085e7d90233d2cdd904ca WatchSource:0}: Error finding container 63f31a77904c9eefacf8c288ec0c63864af11d7253b085e7d90233d2cdd904ca: Status 404 returned error can't find the container with id 63f31a77904c9eefacf8c288ec0c63864af11d7253b085e7d90233d2cdd904ca Oct 02 02:06:26 crc kubenswrapper[4885]: I1002 02:06:26.925605 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 02:06:26 crc kubenswrapper[4885]: I1002 02:06:26.976389 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-cm2bt"] Oct 02 02:06:26 crc kubenswrapper[4885]: I1002 02:06:26.981190 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-cm2bt" Oct 02 02:06:26 crc kubenswrapper[4885]: I1002 02:06:26.984749 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Oct 02 02:06:26 crc kubenswrapper[4885]: I1002 02:06:26.985039 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 02 02:06:27 crc kubenswrapper[4885]: I1002 02:06:27.012268 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-cm2bt"] Oct 02 02:06:27 crc kubenswrapper[4885]: I1002 02:06:27.019796 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 02:06:27 crc kubenswrapper[4885]: I1002 02:06:27.082488 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ffc2d43b-ac70-44d3-9fa5-3dc0f6159f46-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-cm2bt\" (UID: \"ffc2d43b-ac70-44d3-9fa5-3dc0f6159f46\") " pod="openstack/nova-cell1-conductor-db-sync-cm2bt" Oct 02 02:06:27 crc kubenswrapper[4885]: I1002 02:06:27.082830 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ffc2d43b-ac70-44d3-9fa5-3dc0f6159f46-config-data\") pod \"nova-cell1-conductor-db-sync-cm2bt\" (UID: \"ffc2d43b-ac70-44d3-9fa5-3dc0f6159f46\") " pod="openstack/nova-cell1-conductor-db-sync-cm2bt" Oct 02 02:06:27 crc kubenswrapper[4885]: I1002 02:06:27.083041 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ffc2d43b-ac70-44d3-9fa5-3dc0f6159f46-scripts\") pod \"nova-cell1-conductor-db-sync-cm2bt\" (UID: \"ffc2d43b-ac70-44d3-9fa5-3dc0f6159f46\") " pod="openstack/nova-cell1-conductor-db-sync-cm2bt" Oct 02 02:06:27 crc kubenswrapper[4885]: I1002 02:06:27.083188 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ff27s\" (UniqueName: \"kubernetes.io/projected/ffc2d43b-ac70-44d3-9fa5-3dc0f6159f46-kube-api-access-ff27s\") pod \"nova-cell1-conductor-db-sync-cm2bt\" (UID: \"ffc2d43b-ac70-44d3-9fa5-3dc0f6159f46\") " pod="openstack/nova-cell1-conductor-db-sync-cm2bt" Oct 02 02:06:27 crc kubenswrapper[4885]: I1002 02:06:27.135307 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 02:06:27 crc kubenswrapper[4885]: W1002 02:06:27.138998 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3f450788_51fa_430e_a52e_770d02dd9ff0.slice/crio-1d5790ea9a1dec217c85fbab4852cd3c354d8c98d42df670f4fb55766c96eb41 WatchSource:0}: Error finding container 1d5790ea9a1dec217c85fbab4852cd3c354d8c98d42df670f4fb55766c96eb41: Status 404 returned error can't find the container with id 1d5790ea9a1dec217c85fbab4852cd3c354d8c98d42df670f4fb55766c96eb41 Oct 02 02:06:27 crc kubenswrapper[4885]: I1002 02:06:27.162707 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-xm97j"] Oct 02 02:06:27 crc kubenswrapper[4885]: I1002 02:06:27.183852 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ffc2d43b-ac70-44d3-9fa5-3dc0f6159f46-scripts\") pod \"nova-cell1-conductor-db-sync-cm2bt\" (UID: \"ffc2d43b-ac70-44d3-9fa5-3dc0f6159f46\") " pod="openstack/nova-cell1-conductor-db-sync-cm2bt" Oct 02 02:06:27 crc kubenswrapper[4885]: I1002 02:06:27.183910 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ff27s\" (UniqueName: \"kubernetes.io/projected/ffc2d43b-ac70-44d3-9fa5-3dc0f6159f46-kube-api-access-ff27s\") pod \"nova-cell1-conductor-db-sync-cm2bt\" (UID: \"ffc2d43b-ac70-44d3-9fa5-3dc0f6159f46\") " pod="openstack/nova-cell1-conductor-db-sync-cm2bt" Oct 02 02:06:27 crc kubenswrapper[4885]: I1002 02:06:27.184016 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ffc2d43b-ac70-44d3-9fa5-3dc0f6159f46-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-cm2bt\" (UID: \"ffc2d43b-ac70-44d3-9fa5-3dc0f6159f46\") " pod="openstack/nova-cell1-conductor-db-sync-cm2bt" Oct 02 02:06:27 crc kubenswrapper[4885]: I1002 02:06:27.184056 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ffc2d43b-ac70-44d3-9fa5-3dc0f6159f46-config-data\") pod \"nova-cell1-conductor-db-sync-cm2bt\" (UID: \"ffc2d43b-ac70-44d3-9fa5-3dc0f6159f46\") " pod="openstack/nova-cell1-conductor-db-sync-cm2bt" Oct 02 02:06:27 crc kubenswrapper[4885]: I1002 02:06:27.194849 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ffc2d43b-ac70-44d3-9fa5-3dc0f6159f46-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-cm2bt\" (UID: \"ffc2d43b-ac70-44d3-9fa5-3dc0f6159f46\") " pod="openstack/nova-cell1-conductor-db-sync-cm2bt" Oct 02 02:06:27 crc kubenswrapper[4885]: I1002 02:06:27.203087 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ff27s\" (UniqueName: \"kubernetes.io/projected/ffc2d43b-ac70-44d3-9fa5-3dc0f6159f46-kube-api-access-ff27s\") pod \"nova-cell1-conductor-db-sync-cm2bt\" (UID: \"ffc2d43b-ac70-44d3-9fa5-3dc0f6159f46\") " pod="openstack/nova-cell1-conductor-db-sync-cm2bt" Oct 02 02:06:27 crc kubenswrapper[4885]: I1002 02:06:27.209410 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ffc2d43b-ac70-44d3-9fa5-3dc0f6159f46-config-data\") pod \"nova-cell1-conductor-db-sync-cm2bt\" (UID: \"ffc2d43b-ac70-44d3-9fa5-3dc0f6159f46\") " pod="openstack/nova-cell1-conductor-db-sync-cm2bt" Oct 02 02:06:27 crc kubenswrapper[4885]: I1002 02:06:27.213972 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ffc2d43b-ac70-44d3-9fa5-3dc0f6159f46-scripts\") pod \"nova-cell1-conductor-db-sync-cm2bt\" (UID: \"ffc2d43b-ac70-44d3-9fa5-3dc0f6159f46\") " pod="openstack/nova-cell1-conductor-db-sync-cm2bt" Oct 02 02:06:27 crc kubenswrapper[4885]: I1002 02:06:27.341676 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-cm2bt" Oct 02 02:06:27 crc kubenswrapper[4885]: I1002 02:06:27.571044 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-t94gt" event={"ID":"50bc98ab-1629-4591-8e85-bfa8f17962dc","Type":"ContainerStarted","Data":"4c8b11178b78970976d78c639d18fb811b383b7789114f921f64a5148460950c"} Oct 02 02:06:27 crc kubenswrapper[4885]: I1002 02:06:27.571814 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-t94gt" event={"ID":"50bc98ab-1629-4591-8e85-bfa8f17962dc","Type":"ContainerStarted","Data":"d3de1d6e534de28c0b2f70751585c3fbfae6c958776cd3ab60ac3b89a5f7b299"} Oct 02 02:06:27 crc kubenswrapper[4885]: I1002 02:06:27.574985 4885 generic.go:334] "Generic (PLEG): container finished" podID="e8d78e70-43db-4890-9351-f23f53116625" containerID="d68a7424c831e1a782715dc1bd7924fc00e70ad464803683bd68ab2294f92457" exitCode=0 Oct 02 02:06:27 crc kubenswrapper[4885]: I1002 02:06:27.575031 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-xm97j" event={"ID":"e8d78e70-43db-4890-9351-f23f53116625","Type":"ContainerDied","Data":"d68a7424c831e1a782715dc1bd7924fc00e70ad464803683bd68ab2294f92457"} Oct 02 02:06:27 crc kubenswrapper[4885]: I1002 02:06:27.575048 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-xm97j" event={"ID":"e8d78e70-43db-4890-9351-f23f53116625","Type":"ContainerStarted","Data":"6f9a1c017ff51b7b46bfb07a9aa3e4fab503a93b6488053c8a64f6f0e0d9468c"} Oct 02 02:06:27 crc kubenswrapper[4885]: I1002 02:06:27.577168 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"988a6a93-aa98-4ab0-bad2-ec73966f2055","Type":"ContainerStarted","Data":"4a705f48dc4ea19776a8067dbafe443dec228ff2ab3c9908b3c654cccb92a6dd"} Oct 02 02:06:27 crc kubenswrapper[4885]: I1002 02:06:27.578607 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"9c5cf6dd-9dce-4990-a808-c25ee4041ce1","Type":"ContainerStarted","Data":"fe00358be026f53aeb637f82e6883af09ac4cec2709d62c9d13111e080a0fb3a"} Oct 02 02:06:27 crc kubenswrapper[4885]: I1002 02:06:27.580968 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8dba3ac0-e4cf-4d37-baf4-952ff07f8f3f","Type":"ContainerStarted","Data":"63f31a77904c9eefacf8c288ec0c63864af11d7253b085e7d90233d2cdd904ca"} Oct 02 02:06:27 crc kubenswrapper[4885]: I1002 02:06:27.582007 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"3f450788-51fa-430e-a52e-770d02dd9ff0","Type":"ContainerStarted","Data":"1d5790ea9a1dec217c85fbab4852cd3c354d8c98d42df670f4fb55766c96eb41"} Oct 02 02:06:27 crc kubenswrapper[4885]: I1002 02:06:27.595098 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-t94gt" podStartSLOduration=2.595080023 podStartE2EDuration="2.595080023s" podCreationTimestamp="2025-10-02 02:06:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:06:27.590417611 +0000 UTC m=+1176.402165020" watchObservedRunningTime="2025-10-02 02:06:27.595080023 +0000 UTC m=+1176.406827422" Oct 02 02:06:28 crc kubenswrapper[4885]: I1002 02:06:27.781050 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-cm2bt"] Oct 02 02:06:28 crc kubenswrapper[4885]: W1002 02:06:27.797400 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podffc2d43b_ac70_44d3_9fa5_3dc0f6159f46.slice/crio-fedb6345713aae2d9a1df14e6c41474bb78fb2d82f4750a5d2a9e79aef7f8d12 WatchSource:0}: Error finding container fedb6345713aae2d9a1df14e6c41474bb78fb2d82f4750a5d2a9e79aef7f8d12: Status 404 returned error can't find the container with id fedb6345713aae2d9a1df14e6c41474bb78fb2d82f4750a5d2a9e79aef7f8d12 Oct 02 02:06:28 crc kubenswrapper[4885]: I1002 02:06:28.592006 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-cm2bt" event={"ID":"ffc2d43b-ac70-44d3-9fa5-3dc0f6159f46","Type":"ContainerStarted","Data":"af23b39faeaa1d212bba186d27a3acdf681f31ba75d794163d93887c5df66788"} Oct 02 02:06:28 crc kubenswrapper[4885]: I1002 02:06:28.592288 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-cm2bt" event={"ID":"ffc2d43b-ac70-44d3-9fa5-3dc0f6159f46","Type":"ContainerStarted","Data":"fedb6345713aae2d9a1df14e6c41474bb78fb2d82f4750a5d2a9e79aef7f8d12"} Oct 02 02:06:28 crc kubenswrapper[4885]: I1002 02:06:28.596076 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-xm97j" event={"ID":"e8d78e70-43db-4890-9351-f23f53116625","Type":"ContainerStarted","Data":"cd44bce807ff26ccb11e5f121e3962a161f00ef43d5350fb2324ef439f2d8371"} Oct 02 02:06:28 crc kubenswrapper[4885]: I1002 02:06:28.596117 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-845d6d6f59-xm97j" Oct 02 02:06:28 crc kubenswrapper[4885]: I1002 02:06:28.633780 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-cm2bt" podStartSLOduration=2.633761511 podStartE2EDuration="2.633761511s" podCreationTimestamp="2025-10-02 02:06:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:06:28.604974445 +0000 UTC m=+1177.416721844" watchObservedRunningTime="2025-10-02 02:06:28.633761511 +0000 UTC m=+1177.445508900" Oct 02 02:06:28 crc kubenswrapper[4885]: I1002 02:06:28.637374 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-845d6d6f59-xm97j" podStartSLOduration=2.63736438 podStartE2EDuration="2.63736438s" podCreationTimestamp="2025-10-02 02:06:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:06:28.622431496 +0000 UTC m=+1177.434178895" watchObservedRunningTime="2025-10-02 02:06:28.63736438 +0000 UTC m=+1177.449111769" Oct 02 02:06:29 crc kubenswrapper[4885]: I1002 02:06:29.510958 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 02:06:29 crc kubenswrapper[4885]: I1002 02:06:29.517893 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 02:06:33 crc kubenswrapper[4885]: I1002 02:06:33.637818 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-t94gt" event={"ID":"50bc98ab-1629-4591-8e85-bfa8f17962dc","Type":"ContainerDied","Data":"4c8b11178b78970976d78c639d18fb811b383b7789114f921f64a5148460950c"} Oct 02 02:06:33 crc kubenswrapper[4885]: I1002 02:06:33.638575 4885 generic.go:334] "Generic (PLEG): container finished" podID="50bc98ab-1629-4591-8e85-bfa8f17962dc" containerID="4c8b11178b78970976d78c639d18fb811b383b7789114f921f64a5148460950c" exitCode=0 Oct 02 02:06:33 crc kubenswrapper[4885]: I1002 02:06:33.642150 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"988a6a93-aa98-4ab0-bad2-ec73966f2055","Type":"ContainerStarted","Data":"95555fa01dde2ea8c41af5c08a03a3d6d7324baf053eab062395b4b2dfd30b4d"} Oct 02 02:06:33 crc kubenswrapper[4885]: I1002 02:06:33.642202 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"988a6a93-aa98-4ab0-bad2-ec73966f2055","Type":"ContainerStarted","Data":"18a2dc8bd7395baf9497c1d6457f96e0f7feb00ccb9c7c7d22e8cc72fb9cd60c"} Oct 02 02:06:33 crc kubenswrapper[4885]: I1002 02:06:33.642253 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="988a6a93-aa98-4ab0-bad2-ec73966f2055" containerName="nova-metadata-log" containerID="cri-o://18a2dc8bd7395baf9497c1d6457f96e0f7feb00ccb9c7c7d22e8cc72fb9cd60c" gracePeriod=30 Oct 02 02:06:33 crc kubenswrapper[4885]: I1002 02:06:33.642328 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="988a6a93-aa98-4ab0-bad2-ec73966f2055" containerName="nova-metadata-metadata" containerID="cri-o://95555fa01dde2ea8c41af5c08a03a3d6d7324baf053eab062395b4b2dfd30b4d" gracePeriod=30 Oct 02 02:06:33 crc kubenswrapper[4885]: I1002 02:06:33.651499 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"9c5cf6dd-9dce-4990-a808-c25ee4041ce1","Type":"ContainerStarted","Data":"328986948d9fecf68235cca5eabdac788859b2f522129d45fe4db81cb822d4ff"} Oct 02 02:06:33 crc kubenswrapper[4885]: I1002 02:06:33.663484 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8dba3ac0-e4cf-4d37-baf4-952ff07f8f3f","Type":"ContainerStarted","Data":"9744816df676453098b6107d1501e370943b8f22064e77a11b9403b6f27dd433"} Oct 02 02:06:33 crc kubenswrapper[4885]: I1002 02:06:33.663848 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8dba3ac0-e4cf-4d37-baf4-952ff07f8f3f","Type":"ContainerStarted","Data":"70e93fabba24e4509ac499a1783817b7563b50e7ac82c2d2471f009f68a48c69"} Oct 02 02:06:33 crc kubenswrapper[4885]: I1002 02:06:33.666171 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"3f450788-51fa-430e-a52e-770d02dd9ff0","Type":"ContainerStarted","Data":"62a53bb82dd9556b15940442cc66e0553ed172da19d9e4dce1714428edf18e64"} Oct 02 02:06:33 crc kubenswrapper[4885]: I1002 02:06:33.666252 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="3f450788-51fa-430e-a52e-770d02dd9ff0" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://62a53bb82dd9556b15940442cc66e0553ed172da19d9e4dce1714428edf18e64" gracePeriod=30 Oct 02 02:06:33 crc kubenswrapper[4885]: I1002 02:06:33.693510 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=3.279073108 podStartE2EDuration="8.693489731s" podCreationTimestamp="2025-10-02 02:06:25 +0000 UTC" firstStartedPulling="2025-10-02 02:06:26.956793959 +0000 UTC m=+1175.768541358" lastFinishedPulling="2025-10-02 02:06:32.371210562 +0000 UTC m=+1181.182957981" observedRunningTime="2025-10-02 02:06:33.685930801 +0000 UTC m=+1182.497678230" watchObservedRunningTime="2025-10-02 02:06:33.693489731 +0000 UTC m=+1182.505237140" Oct 02 02:06:33 crc kubenswrapper[4885]: I1002 02:06:33.719280 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.330833541 podStartE2EDuration="8.719234104s" podCreationTimestamp="2025-10-02 02:06:25 +0000 UTC" firstStartedPulling="2025-10-02 02:06:26.992502324 +0000 UTC m=+1175.804249713" lastFinishedPulling="2025-10-02 02:06:32.380902887 +0000 UTC m=+1181.192650276" observedRunningTime="2025-10-02 02:06:33.712356454 +0000 UTC m=+1182.524103863" watchObservedRunningTime="2025-10-02 02:06:33.719234104 +0000 UTC m=+1182.530981543" Oct 02 02:06:33 crc kubenswrapper[4885]: I1002 02:06:33.739940 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.226031635 podStartE2EDuration="8.739920742s" podCreationTimestamp="2025-10-02 02:06:25 +0000 UTC" firstStartedPulling="2025-10-02 02:06:26.859506622 +0000 UTC m=+1175.671254021" lastFinishedPulling="2025-10-02 02:06:32.373395699 +0000 UTC m=+1181.185143128" observedRunningTime="2025-10-02 02:06:33.735316153 +0000 UTC m=+1182.547063582" watchObservedRunningTime="2025-10-02 02:06:33.739920742 +0000 UTC m=+1182.551668151" Oct 02 02:06:33 crc kubenswrapper[4885]: I1002 02:06:33.760744 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=3.530664287 podStartE2EDuration="8.760723845s" podCreationTimestamp="2025-10-02 02:06:25 +0000 UTC" firstStartedPulling="2025-10-02 02:06:27.14103113 +0000 UTC m=+1175.952778529" lastFinishedPulling="2025-10-02 02:06:32.371090648 +0000 UTC m=+1181.182838087" observedRunningTime="2025-10-02 02:06:33.751134723 +0000 UTC m=+1182.562882132" watchObservedRunningTime="2025-10-02 02:06:33.760723845 +0000 UTC m=+1182.572471254" Oct 02 02:06:34 crc kubenswrapper[4885]: I1002 02:06:34.678046 4885 generic.go:334] "Generic (PLEG): container finished" podID="988a6a93-aa98-4ab0-bad2-ec73966f2055" containerID="95555fa01dde2ea8c41af5c08a03a3d6d7324baf053eab062395b4b2dfd30b4d" exitCode=0 Oct 02 02:06:34 crc kubenswrapper[4885]: I1002 02:06:34.678193 4885 generic.go:334] "Generic (PLEG): container finished" podID="988a6a93-aa98-4ab0-bad2-ec73966f2055" containerID="18a2dc8bd7395baf9497c1d6457f96e0f7feb00ccb9c7c7d22e8cc72fb9cd60c" exitCode=143 Oct 02 02:06:34 crc kubenswrapper[4885]: I1002 02:06:34.679152 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"988a6a93-aa98-4ab0-bad2-ec73966f2055","Type":"ContainerDied","Data":"95555fa01dde2ea8c41af5c08a03a3d6d7324baf053eab062395b4b2dfd30b4d"} Oct 02 02:06:34 crc kubenswrapper[4885]: I1002 02:06:34.679213 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"988a6a93-aa98-4ab0-bad2-ec73966f2055","Type":"ContainerDied","Data":"18a2dc8bd7395baf9497c1d6457f96e0f7feb00ccb9c7c7d22e8cc72fb9cd60c"} Oct 02 02:06:34 crc kubenswrapper[4885]: I1002 02:06:34.807033 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 02:06:34 crc kubenswrapper[4885]: I1002 02:06:34.940201 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rvjfk\" (UniqueName: \"kubernetes.io/projected/988a6a93-aa98-4ab0-bad2-ec73966f2055-kube-api-access-rvjfk\") pod \"988a6a93-aa98-4ab0-bad2-ec73966f2055\" (UID: \"988a6a93-aa98-4ab0-bad2-ec73966f2055\") " Oct 02 02:06:34 crc kubenswrapper[4885]: I1002 02:06:34.940331 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/988a6a93-aa98-4ab0-bad2-ec73966f2055-combined-ca-bundle\") pod \"988a6a93-aa98-4ab0-bad2-ec73966f2055\" (UID: \"988a6a93-aa98-4ab0-bad2-ec73966f2055\") " Oct 02 02:06:34 crc kubenswrapper[4885]: I1002 02:06:34.940491 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/988a6a93-aa98-4ab0-bad2-ec73966f2055-logs\") pod \"988a6a93-aa98-4ab0-bad2-ec73966f2055\" (UID: \"988a6a93-aa98-4ab0-bad2-ec73966f2055\") " Oct 02 02:06:34 crc kubenswrapper[4885]: I1002 02:06:34.940531 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/988a6a93-aa98-4ab0-bad2-ec73966f2055-config-data\") pod \"988a6a93-aa98-4ab0-bad2-ec73966f2055\" (UID: \"988a6a93-aa98-4ab0-bad2-ec73966f2055\") " Oct 02 02:06:34 crc kubenswrapper[4885]: I1002 02:06:34.941464 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/988a6a93-aa98-4ab0-bad2-ec73966f2055-logs" (OuterVolumeSpecName: "logs") pod "988a6a93-aa98-4ab0-bad2-ec73966f2055" (UID: "988a6a93-aa98-4ab0-bad2-ec73966f2055"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:06:34 crc kubenswrapper[4885]: I1002 02:06:34.952058 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/988a6a93-aa98-4ab0-bad2-ec73966f2055-kube-api-access-rvjfk" (OuterVolumeSpecName: "kube-api-access-rvjfk") pod "988a6a93-aa98-4ab0-bad2-ec73966f2055" (UID: "988a6a93-aa98-4ab0-bad2-ec73966f2055"). InnerVolumeSpecName "kube-api-access-rvjfk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:06:34 crc kubenswrapper[4885]: I1002 02:06:34.973459 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/988a6a93-aa98-4ab0-bad2-ec73966f2055-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "988a6a93-aa98-4ab0-bad2-ec73966f2055" (UID: "988a6a93-aa98-4ab0-bad2-ec73966f2055"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:06:34 crc kubenswrapper[4885]: I1002 02:06:34.979124 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/988a6a93-aa98-4ab0-bad2-ec73966f2055-config-data" (OuterVolumeSpecName: "config-data") pod "988a6a93-aa98-4ab0-bad2-ec73966f2055" (UID: "988a6a93-aa98-4ab0-bad2-ec73966f2055"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:06:35 crc kubenswrapper[4885]: I1002 02:06:35.042470 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rvjfk\" (UniqueName: \"kubernetes.io/projected/988a6a93-aa98-4ab0-bad2-ec73966f2055-kube-api-access-rvjfk\") on node \"crc\" DevicePath \"\"" Oct 02 02:06:35 crc kubenswrapper[4885]: I1002 02:06:35.042495 4885 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/988a6a93-aa98-4ab0-bad2-ec73966f2055-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:06:35 crc kubenswrapper[4885]: I1002 02:06:35.042503 4885 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/988a6a93-aa98-4ab0-bad2-ec73966f2055-logs\") on node \"crc\" DevicePath \"\"" Oct 02 02:06:35 crc kubenswrapper[4885]: I1002 02:06:35.042512 4885 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/988a6a93-aa98-4ab0-bad2-ec73966f2055-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 02:06:35 crc kubenswrapper[4885]: I1002 02:06:35.043895 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-t94gt" Oct 02 02:06:35 crc kubenswrapper[4885]: I1002 02:06:35.143695 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50bc98ab-1629-4591-8e85-bfa8f17962dc-config-data\") pod \"50bc98ab-1629-4591-8e85-bfa8f17962dc\" (UID: \"50bc98ab-1629-4591-8e85-bfa8f17962dc\") " Oct 02 02:06:35 crc kubenswrapper[4885]: I1002 02:06:35.143815 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-prjhc\" (UniqueName: \"kubernetes.io/projected/50bc98ab-1629-4591-8e85-bfa8f17962dc-kube-api-access-prjhc\") pod \"50bc98ab-1629-4591-8e85-bfa8f17962dc\" (UID: \"50bc98ab-1629-4591-8e85-bfa8f17962dc\") " Oct 02 02:06:35 crc kubenswrapper[4885]: I1002 02:06:35.143931 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50bc98ab-1629-4591-8e85-bfa8f17962dc-combined-ca-bundle\") pod \"50bc98ab-1629-4591-8e85-bfa8f17962dc\" (UID: \"50bc98ab-1629-4591-8e85-bfa8f17962dc\") " Oct 02 02:06:35 crc kubenswrapper[4885]: I1002 02:06:35.144045 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/50bc98ab-1629-4591-8e85-bfa8f17962dc-scripts\") pod \"50bc98ab-1629-4591-8e85-bfa8f17962dc\" (UID: \"50bc98ab-1629-4591-8e85-bfa8f17962dc\") " Oct 02 02:06:35 crc kubenswrapper[4885]: I1002 02:06:35.150341 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50bc98ab-1629-4591-8e85-bfa8f17962dc-scripts" (OuterVolumeSpecName: "scripts") pod "50bc98ab-1629-4591-8e85-bfa8f17962dc" (UID: "50bc98ab-1629-4591-8e85-bfa8f17962dc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:06:35 crc kubenswrapper[4885]: I1002 02:06:35.151168 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/50bc98ab-1629-4591-8e85-bfa8f17962dc-kube-api-access-prjhc" (OuterVolumeSpecName: "kube-api-access-prjhc") pod "50bc98ab-1629-4591-8e85-bfa8f17962dc" (UID: "50bc98ab-1629-4591-8e85-bfa8f17962dc"). InnerVolumeSpecName "kube-api-access-prjhc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:06:35 crc kubenswrapper[4885]: I1002 02:06:35.166630 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50bc98ab-1629-4591-8e85-bfa8f17962dc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "50bc98ab-1629-4591-8e85-bfa8f17962dc" (UID: "50bc98ab-1629-4591-8e85-bfa8f17962dc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:06:35 crc kubenswrapper[4885]: I1002 02:06:35.202893 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50bc98ab-1629-4591-8e85-bfa8f17962dc-config-data" (OuterVolumeSpecName: "config-data") pod "50bc98ab-1629-4591-8e85-bfa8f17962dc" (UID: "50bc98ab-1629-4591-8e85-bfa8f17962dc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:06:35 crc kubenswrapper[4885]: I1002 02:06:35.246637 4885 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50bc98ab-1629-4591-8e85-bfa8f17962dc-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 02:06:35 crc kubenswrapper[4885]: I1002 02:06:35.246667 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-prjhc\" (UniqueName: \"kubernetes.io/projected/50bc98ab-1629-4591-8e85-bfa8f17962dc-kube-api-access-prjhc\") on node \"crc\" DevicePath \"\"" Oct 02 02:06:35 crc kubenswrapper[4885]: I1002 02:06:35.246679 4885 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50bc98ab-1629-4591-8e85-bfa8f17962dc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:06:35 crc kubenswrapper[4885]: I1002 02:06:35.246689 4885 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/50bc98ab-1629-4591-8e85-bfa8f17962dc-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 02:06:35 crc kubenswrapper[4885]: I1002 02:06:35.682360 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 02 02:06:35 crc kubenswrapper[4885]: I1002 02:06:35.710413 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-t94gt" Oct 02 02:06:35 crc kubenswrapper[4885]: I1002 02:06:35.710407 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-t94gt" event={"ID":"50bc98ab-1629-4591-8e85-bfa8f17962dc","Type":"ContainerDied","Data":"d3de1d6e534de28c0b2f70751585c3fbfae6c958776cd3ab60ac3b89a5f7b299"} Oct 02 02:06:35 crc kubenswrapper[4885]: I1002 02:06:35.710624 4885 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d3de1d6e534de28c0b2f70751585c3fbfae6c958776cd3ab60ac3b89a5f7b299" Oct 02 02:06:35 crc kubenswrapper[4885]: I1002 02:06:35.713715 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"988a6a93-aa98-4ab0-bad2-ec73966f2055","Type":"ContainerDied","Data":"4a705f48dc4ea19776a8067dbafe443dec228ff2ab3c9908b3c654cccb92a6dd"} Oct 02 02:06:35 crc kubenswrapper[4885]: I1002 02:06:35.713747 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 02:06:35 crc kubenswrapper[4885]: I1002 02:06:35.713797 4885 scope.go:117] "RemoveContainer" containerID="95555fa01dde2ea8c41af5c08a03a3d6d7324baf053eab062395b4b2dfd30b4d" Oct 02 02:06:35 crc kubenswrapper[4885]: I1002 02:06:35.773922 4885 scope.go:117] "RemoveContainer" containerID="18a2dc8bd7395baf9497c1d6457f96e0f7feb00ccb9c7c7d22e8cc72fb9cd60c" Oct 02 02:06:35 crc kubenswrapper[4885]: I1002 02:06:35.839093 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 02:06:35 crc kubenswrapper[4885]: I1002 02:06:35.851645 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 02:06:35 crc kubenswrapper[4885]: I1002 02:06:35.863834 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 02 02:06:35 crc kubenswrapper[4885]: E1002 02:06:35.864210 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="988a6a93-aa98-4ab0-bad2-ec73966f2055" containerName="nova-metadata-metadata" Oct 02 02:06:35 crc kubenswrapper[4885]: I1002 02:06:35.864227 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="988a6a93-aa98-4ab0-bad2-ec73966f2055" containerName="nova-metadata-metadata" Oct 02 02:06:35 crc kubenswrapper[4885]: E1002 02:06:35.864239 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="988a6a93-aa98-4ab0-bad2-ec73966f2055" containerName="nova-metadata-log" Oct 02 02:06:35 crc kubenswrapper[4885]: I1002 02:06:35.864246 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="988a6a93-aa98-4ab0-bad2-ec73966f2055" containerName="nova-metadata-log" Oct 02 02:06:35 crc kubenswrapper[4885]: E1002 02:06:35.864330 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50bc98ab-1629-4591-8e85-bfa8f17962dc" containerName="nova-manage" Oct 02 02:06:35 crc kubenswrapper[4885]: I1002 02:06:35.864338 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="50bc98ab-1629-4591-8e85-bfa8f17962dc" containerName="nova-manage" Oct 02 02:06:35 crc kubenswrapper[4885]: I1002 02:06:35.864523 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="50bc98ab-1629-4591-8e85-bfa8f17962dc" containerName="nova-manage" Oct 02 02:06:35 crc kubenswrapper[4885]: I1002 02:06:35.864535 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="988a6a93-aa98-4ab0-bad2-ec73966f2055" containerName="nova-metadata-log" Oct 02 02:06:35 crc kubenswrapper[4885]: I1002 02:06:35.864544 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="988a6a93-aa98-4ab0-bad2-ec73966f2055" containerName="nova-metadata-metadata" Oct 02 02:06:35 crc kubenswrapper[4885]: I1002 02:06:35.865575 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 02:06:35 crc kubenswrapper[4885]: I1002 02:06:35.868322 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ded68855-3099-4c7a-9239-c1b5e2f12e63-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"ded68855-3099-4c7a-9239-c1b5e2f12e63\") " pod="openstack/nova-metadata-0" Oct 02 02:06:35 crc kubenswrapper[4885]: I1002 02:06:35.868437 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7bmlq\" (UniqueName: \"kubernetes.io/projected/ded68855-3099-4c7a-9239-c1b5e2f12e63-kube-api-access-7bmlq\") pod \"nova-metadata-0\" (UID: \"ded68855-3099-4c7a-9239-c1b5e2f12e63\") " pod="openstack/nova-metadata-0" Oct 02 02:06:35 crc kubenswrapper[4885]: I1002 02:06:35.868482 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ded68855-3099-4c7a-9239-c1b5e2f12e63-config-data\") pod \"nova-metadata-0\" (UID: \"ded68855-3099-4c7a-9239-c1b5e2f12e63\") " pod="openstack/nova-metadata-0" Oct 02 02:06:35 crc kubenswrapper[4885]: I1002 02:06:35.868530 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ded68855-3099-4c7a-9239-c1b5e2f12e63-logs\") pod \"nova-metadata-0\" (UID: \"ded68855-3099-4c7a-9239-c1b5e2f12e63\") " pod="openstack/nova-metadata-0" Oct 02 02:06:35 crc kubenswrapper[4885]: I1002 02:06:35.868564 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/ded68855-3099-4c7a-9239-c1b5e2f12e63-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"ded68855-3099-4c7a-9239-c1b5e2f12e63\") " pod="openstack/nova-metadata-0" Oct 02 02:06:35 crc kubenswrapper[4885]: I1002 02:06:35.869367 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 02 02:06:35 crc kubenswrapper[4885]: I1002 02:06:35.869573 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 02 02:06:35 crc kubenswrapper[4885]: I1002 02:06:35.888799 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 02:06:35 crc kubenswrapper[4885]: I1002 02:06:35.966719 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 02 02:06:35 crc kubenswrapper[4885]: I1002 02:06:35.966905 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="8dba3ac0-e4cf-4d37-baf4-952ff07f8f3f" containerName="nova-api-log" containerID="cri-o://70e93fabba24e4509ac499a1783817b7563b50e7ac82c2d2471f009f68a48c69" gracePeriod=30 Oct 02 02:06:35 crc kubenswrapper[4885]: I1002 02:06:35.967286 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="8dba3ac0-e4cf-4d37-baf4-952ff07f8f3f" containerName="nova-api-api" containerID="cri-o://9744816df676453098b6107d1501e370943b8f22064e77a11b9403b6f27dd433" gracePeriod=30 Oct 02 02:06:35 crc kubenswrapper[4885]: I1002 02:06:35.971708 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/ded68855-3099-4c7a-9239-c1b5e2f12e63-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"ded68855-3099-4c7a-9239-c1b5e2f12e63\") " pod="openstack/nova-metadata-0" Oct 02 02:06:35 crc kubenswrapper[4885]: I1002 02:06:35.971814 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ded68855-3099-4c7a-9239-c1b5e2f12e63-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"ded68855-3099-4c7a-9239-c1b5e2f12e63\") " pod="openstack/nova-metadata-0" Oct 02 02:06:35 crc kubenswrapper[4885]: I1002 02:06:35.971876 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7bmlq\" (UniqueName: \"kubernetes.io/projected/ded68855-3099-4c7a-9239-c1b5e2f12e63-kube-api-access-7bmlq\") pod \"nova-metadata-0\" (UID: \"ded68855-3099-4c7a-9239-c1b5e2f12e63\") " pod="openstack/nova-metadata-0" Oct 02 02:06:35 crc kubenswrapper[4885]: I1002 02:06:35.971907 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ded68855-3099-4c7a-9239-c1b5e2f12e63-config-data\") pod \"nova-metadata-0\" (UID: \"ded68855-3099-4c7a-9239-c1b5e2f12e63\") " pod="openstack/nova-metadata-0" Oct 02 02:06:35 crc kubenswrapper[4885]: I1002 02:06:35.971942 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ded68855-3099-4c7a-9239-c1b5e2f12e63-logs\") pod \"nova-metadata-0\" (UID: \"ded68855-3099-4c7a-9239-c1b5e2f12e63\") " pod="openstack/nova-metadata-0" Oct 02 02:06:35 crc kubenswrapper[4885]: I1002 02:06:35.972323 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ded68855-3099-4c7a-9239-c1b5e2f12e63-logs\") pod \"nova-metadata-0\" (UID: \"ded68855-3099-4c7a-9239-c1b5e2f12e63\") " pod="openstack/nova-metadata-0" Oct 02 02:06:35 crc kubenswrapper[4885]: I1002 02:06:35.983337 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ded68855-3099-4c7a-9239-c1b5e2f12e63-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"ded68855-3099-4c7a-9239-c1b5e2f12e63\") " pod="openstack/nova-metadata-0" Oct 02 02:06:35 crc kubenswrapper[4885]: I1002 02:06:35.984811 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/ded68855-3099-4c7a-9239-c1b5e2f12e63-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"ded68855-3099-4c7a-9239-c1b5e2f12e63\") " pod="openstack/nova-metadata-0" Oct 02 02:06:35 crc kubenswrapper[4885]: I1002 02:06:35.987845 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ded68855-3099-4c7a-9239-c1b5e2f12e63-config-data\") pod \"nova-metadata-0\" (UID: \"ded68855-3099-4c7a-9239-c1b5e2f12e63\") " pod="openstack/nova-metadata-0" Oct 02 02:06:36 crc kubenswrapper[4885]: I1002 02:06:36.018635 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7bmlq\" (UniqueName: \"kubernetes.io/projected/ded68855-3099-4c7a-9239-c1b5e2f12e63-kube-api-access-7bmlq\") pod \"nova-metadata-0\" (UID: \"ded68855-3099-4c7a-9239-c1b5e2f12e63\") " pod="openstack/nova-metadata-0" Oct 02 02:06:36 crc kubenswrapper[4885]: I1002 02:06:36.029318 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 02:06:36 crc kubenswrapper[4885]: I1002 02:06:36.030035 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 02:06:36 crc kubenswrapper[4885]: I1002 02:06:36.047246 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 02:06:36 crc kubenswrapper[4885]: I1002 02:06:36.048057 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="9c5cf6dd-9dce-4990-a808-c25ee4041ce1" containerName="nova-scheduler-scheduler" containerID="cri-o://328986948d9fecf68235cca5eabdac788859b2f522129d45fe4db81cb822d4ff" gracePeriod=30 Oct 02 02:06:36 crc kubenswrapper[4885]: I1002 02:06:36.072126 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="988a6a93-aa98-4ab0-bad2-ec73966f2055" path="/var/lib/kubelet/pods/988a6a93-aa98-4ab0-bad2-ec73966f2055/volumes" Oct 02 02:06:36 crc kubenswrapper[4885]: I1002 02:06:36.315565 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 02 02:06:36 crc kubenswrapper[4885]: I1002 02:06:36.511076 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 02 02:06:36 crc kubenswrapper[4885]: I1002 02:06:36.517972 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-845d6d6f59-xm97j" Oct 02 02:06:36 crc kubenswrapper[4885]: I1002 02:06:36.597658 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-85ntb"] Oct 02 02:06:36 crc kubenswrapper[4885]: I1002 02:06:36.597857 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5784cf869f-85ntb" podUID="7a9422e4-cbc9-4eb7-885d-eb42b1017cab" containerName="dnsmasq-dns" containerID="cri-o://347578f3963a42175c1618b8d901e7db107730b159cc81edce216f280b6faa66" gracePeriod=10 Oct 02 02:06:36 crc kubenswrapper[4885]: I1002 02:06:36.633943 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 02:06:36 crc kubenswrapper[4885]: W1002 02:06:36.662107 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podded68855_3099_4c7a_9239_c1b5e2f12e63.slice/crio-32ffa290cee09ab7fa4e23ff41ae1532e9093de46d254d986ce9d28b106b9578 WatchSource:0}: Error finding container 32ffa290cee09ab7fa4e23ff41ae1532e9093de46d254d986ce9d28b106b9578: Status 404 returned error can't find the container with id 32ffa290cee09ab7fa4e23ff41ae1532e9093de46d254d986ce9d28b106b9578 Oct 02 02:06:36 crc kubenswrapper[4885]: I1002 02:06:36.730905 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ded68855-3099-4c7a-9239-c1b5e2f12e63","Type":"ContainerStarted","Data":"32ffa290cee09ab7fa4e23ff41ae1532e9093de46d254d986ce9d28b106b9578"} Oct 02 02:06:36 crc kubenswrapper[4885]: I1002 02:06:36.736172 4885 generic.go:334] "Generic (PLEG): container finished" podID="7a9422e4-cbc9-4eb7-885d-eb42b1017cab" containerID="347578f3963a42175c1618b8d901e7db107730b159cc81edce216f280b6faa66" exitCode=0 Oct 02 02:06:36 crc kubenswrapper[4885]: I1002 02:06:36.736218 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-85ntb" event={"ID":"7a9422e4-cbc9-4eb7-885d-eb42b1017cab","Type":"ContainerDied","Data":"347578f3963a42175c1618b8d901e7db107730b159cc81edce216f280b6faa66"} Oct 02 02:06:36 crc kubenswrapper[4885]: I1002 02:06:36.738284 4885 generic.go:334] "Generic (PLEG): container finished" podID="8dba3ac0-e4cf-4d37-baf4-952ff07f8f3f" containerID="9744816df676453098b6107d1501e370943b8f22064e77a11b9403b6f27dd433" exitCode=0 Oct 02 02:06:36 crc kubenswrapper[4885]: I1002 02:06:36.738300 4885 generic.go:334] "Generic (PLEG): container finished" podID="8dba3ac0-e4cf-4d37-baf4-952ff07f8f3f" containerID="70e93fabba24e4509ac499a1783817b7563b50e7ac82c2d2471f009f68a48c69" exitCode=143 Oct 02 02:06:36 crc kubenswrapper[4885]: I1002 02:06:36.738314 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8dba3ac0-e4cf-4d37-baf4-952ff07f8f3f","Type":"ContainerDied","Data":"9744816df676453098b6107d1501e370943b8f22064e77a11b9403b6f27dd433"} Oct 02 02:06:36 crc kubenswrapper[4885]: I1002 02:06:36.738328 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8dba3ac0-e4cf-4d37-baf4-952ff07f8f3f","Type":"ContainerDied","Data":"70e93fabba24e4509ac499a1783817b7563b50e7ac82c2d2471f009f68a48c69"} Oct 02 02:06:36 crc kubenswrapper[4885]: I1002 02:06:36.962406 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 02:06:37 crc kubenswrapper[4885]: I1002 02:06:37.005168 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8dba3ac0-e4cf-4d37-baf4-952ff07f8f3f-logs\") pod \"8dba3ac0-e4cf-4d37-baf4-952ff07f8f3f\" (UID: \"8dba3ac0-e4cf-4d37-baf4-952ff07f8f3f\") " Oct 02 02:06:37 crc kubenswrapper[4885]: I1002 02:06:37.005233 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8dba3ac0-e4cf-4d37-baf4-952ff07f8f3f-config-data\") pod \"8dba3ac0-e4cf-4d37-baf4-952ff07f8f3f\" (UID: \"8dba3ac0-e4cf-4d37-baf4-952ff07f8f3f\") " Oct 02 02:06:37 crc kubenswrapper[4885]: I1002 02:06:37.005484 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dgc5x\" (UniqueName: \"kubernetes.io/projected/8dba3ac0-e4cf-4d37-baf4-952ff07f8f3f-kube-api-access-dgc5x\") pod \"8dba3ac0-e4cf-4d37-baf4-952ff07f8f3f\" (UID: \"8dba3ac0-e4cf-4d37-baf4-952ff07f8f3f\") " Oct 02 02:06:37 crc kubenswrapper[4885]: I1002 02:06:37.005519 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8dba3ac0-e4cf-4d37-baf4-952ff07f8f3f-combined-ca-bundle\") pod \"8dba3ac0-e4cf-4d37-baf4-952ff07f8f3f\" (UID: \"8dba3ac0-e4cf-4d37-baf4-952ff07f8f3f\") " Oct 02 02:06:37 crc kubenswrapper[4885]: I1002 02:06:37.008441 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8dba3ac0-e4cf-4d37-baf4-952ff07f8f3f-logs" (OuterVolumeSpecName: "logs") pod "8dba3ac0-e4cf-4d37-baf4-952ff07f8f3f" (UID: "8dba3ac0-e4cf-4d37-baf4-952ff07f8f3f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:06:37 crc kubenswrapper[4885]: I1002 02:06:37.013486 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8dba3ac0-e4cf-4d37-baf4-952ff07f8f3f-kube-api-access-dgc5x" (OuterVolumeSpecName: "kube-api-access-dgc5x") pod "8dba3ac0-e4cf-4d37-baf4-952ff07f8f3f" (UID: "8dba3ac0-e4cf-4d37-baf4-952ff07f8f3f"). InnerVolumeSpecName "kube-api-access-dgc5x". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:06:37 crc kubenswrapper[4885]: I1002 02:06:37.040272 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8dba3ac0-e4cf-4d37-baf4-952ff07f8f3f-config-data" (OuterVolumeSpecName: "config-data") pod "8dba3ac0-e4cf-4d37-baf4-952ff07f8f3f" (UID: "8dba3ac0-e4cf-4d37-baf4-952ff07f8f3f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:06:37 crc kubenswrapper[4885]: I1002 02:06:37.047157 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-85ntb" Oct 02 02:06:37 crc kubenswrapper[4885]: I1002 02:06:37.048227 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8dba3ac0-e4cf-4d37-baf4-952ff07f8f3f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8dba3ac0-e4cf-4d37-baf4-952ff07f8f3f" (UID: "8dba3ac0-e4cf-4d37-baf4-952ff07f8f3f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:06:37 crc kubenswrapper[4885]: I1002 02:06:37.109971 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7a9422e4-cbc9-4eb7-885d-eb42b1017cab-dns-swift-storage-0\") pod \"7a9422e4-cbc9-4eb7-885d-eb42b1017cab\" (UID: \"7a9422e4-cbc9-4eb7-885d-eb42b1017cab\") " Oct 02 02:06:37 crc kubenswrapper[4885]: I1002 02:06:37.110087 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7a9422e4-cbc9-4eb7-885d-eb42b1017cab-ovsdbserver-nb\") pod \"7a9422e4-cbc9-4eb7-885d-eb42b1017cab\" (UID: \"7a9422e4-cbc9-4eb7-885d-eb42b1017cab\") " Oct 02 02:06:37 crc kubenswrapper[4885]: I1002 02:06:37.110189 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7a9422e4-cbc9-4eb7-885d-eb42b1017cab-dns-svc\") pod \"7a9422e4-cbc9-4eb7-885d-eb42b1017cab\" (UID: \"7a9422e4-cbc9-4eb7-885d-eb42b1017cab\") " Oct 02 02:06:37 crc kubenswrapper[4885]: I1002 02:06:37.110238 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a9422e4-cbc9-4eb7-885d-eb42b1017cab-config\") pod \"7a9422e4-cbc9-4eb7-885d-eb42b1017cab\" (UID: \"7a9422e4-cbc9-4eb7-885d-eb42b1017cab\") " Oct 02 02:06:37 crc kubenswrapper[4885]: I1002 02:06:37.110268 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8h5f4\" (UniqueName: \"kubernetes.io/projected/7a9422e4-cbc9-4eb7-885d-eb42b1017cab-kube-api-access-8h5f4\") pod \"7a9422e4-cbc9-4eb7-885d-eb42b1017cab\" (UID: \"7a9422e4-cbc9-4eb7-885d-eb42b1017cab\") " Oct 02 02:06:37 crc kubenswrapper[4885]: I1002 02:06:37.110316 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7a9422e4-cbc9-4eb7-885d-eb42b1017cab-ovsdbserver-sb\") pod \"7a9422e4-cbc9-4eb7-885d-eb42b1017cab\" (UID: \"7a9422e4-cbc9-4eb7-885d-eb42b1017cab\") " Oct 02 02:06:37 crc kubenswrapper[4885]: I1002 02:06:37.110665 4885 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8dba3ac0-e4cf-4d37-baf4-952ff07f8f3f-logs\") on node \"crc\" DevicePath \"\"" Oct 02 02:06:37 crc kubenswrapper[4885]: I1002 02:06:37.110676 4885 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8dba3ac0-e4cf-4d37-baf4-952ff07f8f3f-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 02:06:37 crc kubenswrapper[4885]: I1002 02:06:37.110685 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dgc5x\" (UniqueName: \"kubernetes.io/projected/8dba3ac0-e4cf-4d37-baf4-952ff07f8f3f-kube-api-access-dgc5x\") on node \"crc\" DevicePath \"\"" Oct 02 02:06:37 crc kubenswrapper[4885]: I1002 02:06:37.110694 4885 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8dba3ac0-e4cf-4d37-baf4-952ff07f8f3f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:06:37 crc kubenswrapper[4885]: I1002 02:06:37.116331 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a9422e4-cbc9-4eb7-885d-eb42b1017cab-kube-api-access-8h5f4" (OuterVolumeSpecName: "kube-api-access-8h5f4") pod "7a9422e4-cbc9-4eb7-885d-eb42b1017cab" (UID: "7a9422e4-cbc9-4eb7-885d-eb42b1017cab"). InnerVolumeSpecName "kube-api-access-8h5f4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:06:37 crc kubenswrapper[4885]: I1002 02:06:37.167868 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7a9422e4-cbc9-4eb7-885d-eb42b1017cab-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7a9422e4-cbc9-4eb7-885d-eb42b1017cab" (UID: "7a9422e4-cbc9-4eb7-885d-eb42b1017cab"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:06:37 crc kubenswrapper[4885]: I1002 02:06:37.178451 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7a9422e4-cbc9-4eb7-885d-eb42b1017cab-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "7a9422e4-cbc9-4eb7-885d-eb42b1017cab" (UID: "7a9422e4-cbc9-4eb7-885d-eb42b1017cab"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:06:37 crc kubenswrapper[4885]: I1002 02:06:37.182971 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7a9422e4-cbc9-4eb7-885d-eb42b1017cab-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7a9422e4-cbc9-4eb7-885d-eb42b1017cab" (UID: "7a9422e4-cbc9-4eb7-885d-eb42b1017cab"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:06:37 crc kubenswrapper[4885]: I1002 02:06:37.190063 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7a9422e4-cbc9-4eb7-885d-eb42b1017cab-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7a9422e4-cbc9-4eb7-885d-eb42b1017cab" (UID: "7a9422e4-cbc9-4eb7-885d-eb42b1017cab"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:06:37 crc kubenswrapper[4885]: I1002 02:06:37.195391 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7a9422e4-cbc9-4eb7-885d-eb42b1017cab-config" (OuterVolumeSpecName: "config") pod "7a9422e4-cbc9-4eb7-885d-eb42b1017cab" (UID: "7a9422e4-cbc9-4eb7-885d-eb42b1017cab"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:06:37 crc kubenswrapper[4885]: I1002 02:06:37.212095 4885 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7a9422e4-cbc9-4eb7-885d-eb42b1017cab-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 02:06:37 crc kubenswrapper[4885]: I1002 02:06:37.212126 4885 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7a9422e4-cbc9-4eb7-885d-eb42b1017cab-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 02:06:37 crc kubenswrapper[4885]: I1002 02:06:37.212141 4885 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a9422e4-cbc9-4eb7-885d-eb42b1017cab-config\") on node \"crc\" DevicePath \"\"" Oct 02 02:06:37 crc kubenswrapper[4885]: I1002 02:06:37.212151 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8h5f4\" (UniqueName: \"kubernetes.io/projected/7a9422e4-cbc9-4eb7-885d-eb42b1017cab-kube-api-access-8h5f4\") on node \"crc\" DevicePath \"\"" Oct 02 02:06:37 crc kubenswrapper[4885]: I1002 02:06:37.212163 4885 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7a9422e4-cbc9-4eb7-885d-eb42b1017cab-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 02:06:37 crc kubenswrapper[4885]: I1002 02:06:37.212174 4885 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7a9422e4-cbc9-4eb7-885d-eb42b1017cab-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 02 02:06:37 crc kubenswrapper[4885]: I1002 02:06:37.226864 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 02:06:37 crc kubenswrapper[4885]: I1002 02:06:37.313076 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c5cf6dd-9dce-4990-a808-c25ee4041ce1-config-data\") pod \"9c5cf6dd-9dce-4990-a808-c25ee4041ce1\" (UID: \"9c5cf6dd-9dce-4990-a808-c25ee4041ce1\") " Oct 02 02:06:37 crc kubenswrapper[4885]: I1002 02:06:37.313135 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9pz7\" (UniqueName: \"kubernetes.io/projected/9c5cf6dd-9dce-4990-a808-c25ee4041ce1-kube-api-access-w9pz7\") pod \"9c5cf6dd-9dce-4990-a808-c25ee4041ce1\" (UID: \"9c5cf6dd-9dce-4990-a808-c25ee4041ce1\") " Oct 02 02:06:37 crc kubenswrapper[4885]: I1002 02:06:37.313164 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c5cf6dd-9dce-4990-a808-c25ee4041ce1-combined-ca-bundle\") pod \"9c5cf6dd-9dce-4990-a808-c25ee4041ce1\" (UID: \"9c5cf6dd-9dce-4990-a808-c25ee4041ce1\") " Oct 02 02:06:37 crc kubenswrapper[4885]: I1002 02:06:37.317521 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c5cf6dd-9dce-4990-a808-c25ee4041ce1-kube-api-access-w9pz7" (OuterVolumeSpecName: "kube-api-access-w9pz7") pod "9c5cf6dd-9dce-4990-a808-c25ee4041ce1" (UID: "9c5cf6dd-9dce-4990-a808-c25ee4041ce1"). InnerVolumeSpecName "kube-api-access-w9pz7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:06:37 crc kubenswrapper[4885]: I1002 02:06:37.339388 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c5cf6dd-9dce-4990-a808-c25ee4041ce1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9c5cf6dd-9dce-4990-a808-c25ee4041ce1" (UID: "9c5cf6dd-9dce-4990-a808-c25ee4041ce1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:06:37 crc kubenswrapper[4885]: I1002 02:06:37.339784 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c5cf6dd-9dce-4990-a808-c25ee4041ce1-config-data" (OuterVolumeSpecName: "config-data") pod "9c5cf6dd-9dce-4990-a808-c25ee4041ce1" (UID: "9c5cf6dd-9dce-4990-a808-c25ee4041ce1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:06:37 crc kubenswrapper[4885]: I1002 02:06:37.415723 4885 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c5cf6dd-9dce-4990-a808-c25ee4041ce1-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 02:06:37 crc kubenswrapper[4885]: I1002 02:06:37.415760 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9pz7\" (UniqueName: \"kubernetes.io/projected/9c5cf6dd-9dce-4990-a808-c25ee4041ce1-kube-api-access-w9pz7\") on node \"crc\" DevicePath \"\"" Oct 02 02:06:37 crc kubenswrapper[4885]: I1002 02:06:37.415774 4885 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c5cf6dd-9dce-4990-a808-c25ee4041ce1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:06:37 crc kubenswrapper[4885]: I1002 02:06:37.748765 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-85ntb" event={"ID":"7a9422e4-cbc9-4eb7-885d-eb42b1017cab","Type":"ContainerDied","Data":"37eece54cd8a4c574dcc228201218ea61821ce6ce6068bfa666592c9a6828440"} Oct 02 02:06:37 crc kubenswrapper[4885]: I1002 02:06:37.748844 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-85ntb" Oct 02 02:06:37 crc kubenswrapper[4885]: I1002 02:06:37.749069 4885 scope.go:117] "RemoveContainer" containerID="347578f3963a42175c1618b8d901e7db107730b159cc81edce216f280b6faa66" Oct 02 02:06:37 crc kubenswrapper[4885]: I1002 02:06:37.752612 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 02:06:37 crc kubenswrapper[4885]: I1002 02:06:37.752644 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8dba3ac0-e4cf-4d37-baf4-952ff07f8f3f","Type":"ContainerDied","Data":"63f31a77904c9eefacf8c288ec0c63864af11d7253b085e7d90233d2cdd904ca"} Oct 02 02:06:37 crc kubenswrapper[4885]: I1002 02:06:37.758150 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ded68855-3099-4c7a-9239-c1b5e2f12e63","Type":"ContainerStarted","Data":"80bb7bfd8498043c9aee2dd1ffdbe9cd51f22691f4e8ba3f00d0b91892dfa24c"} Oct 02 02:06:37 crc kubenswrapper[4885]: I1002 02:06:37.758194 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ded68855-3099-4c7a-9239-c1b5e2f12e63","Type":"ContainerStarted","Data":"c7b1579fd3177d152948ca37a1637447fb5e907f1e2d5ae7c4897e522fb78787"} Oct 02 02:06:37 crc kubenswrapper[4885]: I1002 02:06:37.758330 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="ded68855-3099-4c7a-9239-c1b5e2f12e63" containerName="nova-metadata-metadata" containerID="cri-o://80bb7bfd8498043c9aee2dd1ffdbe9cd51f22691f4e8ba3f00d0b91892dfa24c" gracePeriod=30 Oct 02 02:06:37 crc kubenswrapper[4885]: I1002 02:06:37.758331 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="ded68855-3099-4c7a-9239-c1b5e2f12e63" containerName="nova-metadata-log" containerID="cri-o://c7b1579fd3177d152948ca37a1637447fb5e907f1e2d5ae7c4897e522fb78787" gracePeriod=30 Oct 02 02:06:37 crc kubenswrapper[4885]: I1002 02:06:37.761690 4885 generic.go:334] "Generic (PLEG): container finished" podID="9c5cf6dd-9dce-4990-a808-c25ee4041ce1" containerID="328986948d9fecf68235cca5eabdac788859b2f522129d45fe4db81cb822d4ff" exitCode=0 Oct 02 02:06:37 crc kubenswrapper[4885]: I1002 02:06:37.761735 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"9c5cf6dd-9dce-4990-a808-c25ee4041ce1","Type":"ContainerDied","Data":"328986948d9fecf68235cca5eabdac788859b2f522129d45fe4db81cb822d4ff"} Oct 02 02:06:37 crc kubenswrapper[4885]: I1002 02:06:37.761773 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 02:06:37 crc kubenswrapper[4885]: I1002 02:06:37.761775 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"9c5cf6dd-9dce-4990-a808-c25ee4041ce1","Type":"ContainerDied","Data":"fe00358be026f53aeb637f82e6883af09ac4cec2709d62c9d13111e080a0fb3a"} Oct 02 02:06:37 crc kubenswrapper[4885]: I1002 02:06:37.781960 4885 scope.go:117] "RemoveContainer" containerID="fdc94990ef2b2029d14e73d89279ee34dec65d2cb37895e045e79f39df25994e" Oct 02 02:06:37 crc kubenswrapper[4885]: I1002 02:06:37.792350 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.792328769 podStartE2EDuration="2.792328769s" podCreationTimestamp="2025-10-02 02:06:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:06:37.779589941 +0000 UTC m=+1186.591337330" watchObservedRunningTime="2025-10-02 02:06:37.792328769 +0000 UTC m=+1186.604076178" Oct 02 02:06:37 crc kubenswrapper[4885]: I1002 02:06:37.810522 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-85ntb"] Oct 02 02:06:37 crc kubenswrapper[4885]: I1002 02:06:37.810948 4885 scope.go:117] "RemoveContainer" containerID="9744816df676453098b6107d1501e370943b8f22064e77a11b9403b6f27dd433" Oct 02 02:06:37 crc kubenswrapper[4885]: I1002 02:06:37.817288 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-85ntb"] Oct 02 02:06:37 crc kubenswrapper[4885]: I1002 02:06:37.825354 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 02 02:06:37 crc kubenswrapper[4885]: I1002 02:06:37.832685 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 02 02:06:37 crc kubenswrapper[4885]: I1002 02:06:37.846002 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 02:06:37 crc kubenswrapper[4885]: I1002 02:06:37.848633 4885 scope.go:117] "RemoveContainer" containerID="70e93fabba24e4509ac499a1783817b7563b50e7ac82c2d2471f009f68a48c69" Oct 02 02:06:37 crc kubenswrapper[4885]: I1002 02:06:37.853128 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 02 02:06:37 crc kubenswrapper[4885]: E1002 02:06:37.853552 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a9422e4-cbc9-4eb7-885d-eb42b1017cab" containerName="dnsmasq-dns" Oct 02 02:06:37 crc kubenswrapper[4885]: I1002 02:06:37.853566 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a9422e4-cbc9-4eb7-885d-eb42b1017cab" containerName="dnsmasq-dns" Oct 02 02:06:37 crc kubenswrapper[4885]: E1002 02:06:37.853581 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8dba3ac0-e4cf-4d37-baf4-952ff07f8f3f" containerName="nova-api-log" Oct 02 02:06:37 crc kubenswrapper[4885]: I1002 02:06:37.853587 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="8dba3ac0-e4cf-4d37-baf4-952ff07f8f3f" containerName="nova-api-log" Oct 02 02:06:37 crc kubenswrapper[4885]: E1002 02:06:37.853598 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8dba3ac0-e4cf-4d37-baf4-952ff07f8f3f" containerName="nova-api-api" Oct 02 02:06:37 crc kubenswrapper[4885]: I1002 02:06:37.853606 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="8dba3ac0-e4cf-4d37-baf4-952ff07f8f3f" containerName="nova-api-api" Oct 02 02:06:37 crc kubenswrapper[4885]: E1002 02:06:37.853620 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c5cf6dd-9dce-4990-a808-c25ee4041ce1" containerName="nova-scheduler-scheduler" Oct 02 02:06:37 crc kubenswrapper[4885]: I1002 02:06:37.853626 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c5cf6dd-9dce-4990-a808-c25ee4041ce1" containerName="nova-scheduler-scheduler" Oct 02 02:06:37 crc kubenswrapper[4885]: E1002 02:06:37.853647 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a9422e4-cbc9-4eb7-885d-eb42b1017cab" containerName="init" Oct 02 02:06:37 crc kubenswrapper[4885]: I1002 02:06:37.853653 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a9422e4-cbc9-4eb7-885d-eb42b1017cab" containerName="init" Oct 02 02:06:37 crc kubenswrapper[4885]: I1002 02:06:37.853810 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a9422e4-cbc9-4eb7-885d-eb42b1017cab" containerName="dnsmasq-dns" Oct 02 02:06:37 crc kubenswrapper[4885]: I1002 02:06:37.853829 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c5cf6dd-9dce-4990-a808-c25ee4041ce1" containerName="nova-scheduler-scheduler" Oct 02 02:06:37 crc kubenswrapper[4885]: I1002 02:06:37.853838 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="8dba3ac0-e4cf-4d37-baf4-952ff07f8f3f" containerName="nova-api-log" Oct 02 02:06:37 crc kubenswrapper[4885]: I1002 02:06:37.853850 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="8dba3ac0-e4cf-4d37-baf4-952ff07f8f3f" containerName="nova-api-api" Oct 02 02:06:37 crc kubenswrapper[4885]: I1002 02:06:37.854770 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 02:06:37 crc kubenswrapper[4885]: I1002 02:06:37.862649 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 02 02:06:37 crc kubenswrapper[4885]: I1002 02:06:37.876401 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 02:06:37 crc kubenswrapper[4885]: I1002 02:06:37.876487 4885 scope.go:117] "RemoveContainer" containerID="328986948d9fecf68235cca5eabdac788859b2f522129d45fe4db81cb822d4ff" Oct 02 02:06:37 crc kubenswrapper[4885]: I1002 02:06:37.897408 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 02 02:06:37 crc kubenswrapper[4885]: I1002 02:06:37.922627 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 02:06:37 crc kubenswrapper[4885]: I1002 02:06:37.925815 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h4jp2\" (UniqueName: \"kubernetes.io/projected/77273755-f99f-4c79-a730-6e021de17e85-kube-api-access-h4jp2\") pod \"nova-api-0\" (UID: \"77273755-f99f-4c79-a730-6e021de17e85\") " pod="openstack/nova-api-0" Oct 02 02:06:37 crc kubenswrapper[4885]: I1002 02:06:37.925920 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/77273755-f99f-4c79-a730-6e021de17e85-logs\") pod \"nova-api-0\" (UID: \"77273755-f99f-4c79-a730-6e021de17e85\") " pod="openstack/nova-api-0" Oct 02 02:06:37 crc kubenswrapper[4885]: I1002 02:06:37.925971 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77273755-f99f-4c79-a730-6e021de17e85-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"77273755-f99f-4c79-a730-6e021de17e85\") " pod="openstack/nova-api-0" Oct 02 02:06:37 crc kubenswrapper[4885]: I1002 02:06:37.926010 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77273755-f99f-4c79-a730-6e021de17e85-config-data\") pod \"nova-api-0\" (UID: \"77273755-f99f-4c79-a730-6e021de17e85\") " pod="openstack/nova-api-0" Oct 02 02:06:37 crc kubenswrapper[4885]: I1002 02:06:37.926115 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 02:06:37 crc kubenswrapper[4885]: I1002 02:06:37.929761 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 02:06:37 crc kubenswrapper[4885]: I1002 02:06:37.933843 4885 scope.go:117] "RemoveContainer" containerID="328986948d9fecf68235cca5eabdac788859b2f522129d45fe4db81cb822d4ff" Oct 02 02:06:37 crc kubenswrapper[4885]: I1002 02:06:37.933911 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 02 02:06:37 crc kubenswrapper[4885]: E1002 02:06:37.934232 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"328986948d9fecf68235cca5eabdac788859b2f522129d45fe4db81cb822d4ff\": container with ID starting with 328986948d9fecf68235cca5eabdac788859b2f522129d45fe4db81cb822d4ff not found: ID does not exist" containerID="328986948d9fecf68235cca5eabdac788859b2f522129d45fe4db81cb822d4ff" Oct 02 02:06:37 crc kubenswrapper[4885]: I1002 02:06:37.934282 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"328986948d9fecf68235cca5eabdac788859b2f522129d45fe4db81cb822d4ff"} err="failed to get container status \"328986948d9fecf68235cca5eabdac788859b2f522129d45fe4db81cb822d4ff\": rpc error: code = NotFound desc = could not find container \"328986948d9fecf68235cca5eabdac788859b2f522129d45fe4db81cb822d4ff\": container with ID starting with 328986948d9fecf68235cca5eabdac788859b2f522129d45fe4db81cb822d4ff not found: ID does not exist" Oct 02 02:06:38 crc kubenswrapper[4885]: I1002 02:06:38.028029 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gzs6m\" (UniqueName: \"kubernetes.io/projected/6895f52a-cc09-4eac-9f57-32f91106afb9-kube-api-access-gzs6m\") pod \"nova-scheduler-0\" (UID: \"6895f52a-cc09-4eac-9f57-32f91106afb9\") " pod="openstack/nova-scheduler-0" Oct 02 02:06:38 crc kubenswrapper[4885]: I1002 02:06:38.028179 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77273755-f99f-4c79-a730-6e021de17e85-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"77273755-f99f-4c79-a730-6e021de17e85\") " pod="openstack/nova-api-0" Oct 02 02:06:38 crc kubenswrapper[4885]: I1002 02:06:38.028336 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77273755-f99f-4c79-a730-6e021de17e85-config-data\") pod \"nova-api-0\" (UID: \"77273755-f99f-4c79-a730-6e021de17e85\") " pod="openstack/nova-api-0" Oct 02 02:06:38 crc kubenswrapper[4885]: I1002 02:06:38.028435 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h4jp2\" (UniqueName: \"kubernetes.io/projected/77273755-f99f-4c79-a730-6e021de17e85-kube-api-access-h4jp2\") pod \"nova-api-0\" (UID: \"77273755-f99f-4c79-a730-6e021de17e85\") " pod="openstack/nova-api-0" Oct 02 02:06:38 crc kubenswrapper[4885]: I1002 02:06:38.028527 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6895f52a-cc09-4eac-9f57-32f91106afb9-config-data\") pod \"nova-scheduler-0\" (UID: \"6895f52a-cc09-4eac-9f57-32f91106afb9\") " pod="openstack/nova-scheduler-0" Oct 02 02:06:38 crc kubenswrapper[4885]: I1002 02:06:38.028710 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6895f52a-cc09-4eac-9f57-32f91106afb9-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"6895f52a-cc09-4eac-9f57-32f91106afb9\") " pod="openstack/nova-scheduler-0" Oct 02 02:06:38 crc kubenswrapper[4885]: I1002 02:06:38.028867 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/77273755-f99f-4c79-a730-6e021de17e85-logs\") pod \"nova-api-0\" (UID: \"77273755-f99f-4c79-a730-6e021de17e85\") " pod="openstack/nova-api-0" Oct 02 02:06:38 crc kubenswrapper[4885]: I1002 02:06:38.029163 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/77273755-f99f-4c79-a730-6e021de17e85-logs\") pod \"nova-api-0\" (UID: \"77273755-f99f-4c79-a730-6e021de17e85\") " pod="openstack/nova-api-0" Oct 02 02:06:38 crc kubenswrapper[4885]: I1002 02:06:38.033087 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77273755-f99f-4c79-a730-6e021de17e85-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"77273755-f99f-4c79-a730-6e021de17e85\") " pod="openstack/nova-api-0" Oct 02 02:06:38 crc kubenswrapper[4885]: I1002 02:06:38.036893 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77273755-f99f-4c79-a730-6e021de17e85-config-data\") pod \"nova-api-0\" (UID: \"77273755-f99f-4c79-a730-6e021de17e85\") " pod="openstack/nova-api-0" Oct 02 02:06:38 crc kubenswrapper[4885]: I1002 02:06:38.046684 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h4jp2\" (UniqueName: \"kubernetes.io/projected/77273755-f99f-4c79-a730-6e021de17e85-kube-api-access-h4jp2\") pod \"nova-api-0\" (UID: \"77273755-f99f-4c79-a730-6e021de17e85\") " pod="openstack/nova-api-0" Oct 02 02:06:38 crc kubenswrapper[4885]: I1002 02:06:38.066485 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7a9422e4-cbc9-4eb7-885d-eb42b1017cab" path="/var/lib/kubelet/pods/7a9422e4-cbc9-4eb7-885d-eb42b1017cab/volumes" Oct 02 02:06:38 crc kubenswrapper[4885]: I1002 02:06:38.067110 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8dba3ac0-e4cf-4d37-baf4-952ff07f8f3f" path="/var/lib/kubelet/pods/8dba3ac0-e4cf-4d37-baf4-952ff07f8f3f/volumes" Oct 02 02:06:38 crc kubenswrapper[4885]: I1002 02:06:38.067797 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c5cf6dd-9dce-4990-a808-c25ee4041ce1" path="/var/lib/kubelet/pods/9c5cf6dd-9dce-4990-a808-c25ee4041ce1/volumes" Oct 02 02:06:38 crc kubenswrapper[4885]: I1002 02:06:38.130824 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6895f52a-cc09-4eac-9f57-32f91106afb9-config-data\") pod \"nova-scheduler-0\" (UID: \"6895f52a-cc09-4eac-9f57-32f91106afb9\") " pod="openstack/nova-scheduler-0" Oct 02 02:06:38 crc kubenswrapper[4885]: I1002 02:06:38.131236 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6895f52a-cc09-4eac-9f57-32f91106afb9-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"6895f52a-cc09-4eac-9f57-32f91106afb9\") " pod="openstack/nova-scheduler-0" Oct 02 02:06:38 crc kubenswrapper[4885]: I1002 02:06:38.131442 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gzs6m\" (UniqueName: \"kubernetes.io/projected/6895f52a-cc09-4eac-9f57-32f91106afb9-kube-api-access-gzs6m\") pod \"nova-scheduler-0\" (UID: \"6895f52a-cc09-4eac-9f57-32f91106afb9\") " pod="openstack/nova-scheduler-0" Oct 02 02:06:38 crc kubenswrapper[4885]: I1002 02:06:38.135858 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6895f52a-cc09-4eac-9f57-32f91106afb9-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"6895f52a-cc09-4eac-9f57-32f91106afb9\") " pod="openstack/nova-scheduler-0" Oct 02 02:06:38 crc kubenswrapper[4885]: I1002 02:06:38.138366 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6895f52a-cc09-4eac-9f57-32f91106afb9-config-data\") pod \"nova-scheduler-0\" (UID: \"6895f52a-cc09-4eac-9f57-32f91106afb9\") " pod="openstack/nova-scheduler-0" Oct 02 02:06:38 crc kubenswrapper[4885]: I1002 02:06:38.149833 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gzs6m\" (UniqueName: \"kubernetes.io/projected/6895f52a-cc09-4eac-9f57-32f91106afb9-kube-api-access-gzs6m\") pod \"nova-scheduler-0\" (UID: \"6895f52a-cc09-4eac-9f57-32f91106afb9\") " pod="openstack/nova-scheduler-0" Oct 02 02:06:38 crc kubenswrapper[4885]: I1002 02:06:38.222309 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 02:06:38 crc kubenswrapper[4885]: I1002 02:06:38.228181 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 02:06:38 crc kubenswrapper[4885]: I1002 02:06:38.327725 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 02:06:38 crc kubenswrapper[4885]: I1002 02:06:38.335606 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ded68855-3099-4c7a-9239-c1b5e2f12e63-combined-ca-bundle\") pod \"ded68855-3099-4c7a-9239-c1b5e2f12e63\" (UID: \"ded68855-3099-4c7a-9239-c1b5e2f12e63\") " Oct 02 02:06:38 crc kubenswrapper[4885]: I1002 02:06:38.335691 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ded68855-3099-4c7a-9239-c1b5e2f12e63-logs\") pod \"ded68855-3099-4c7a-9239-c1b5e2f12e63\" (UID: \"ded68855-3099-4c7a-9239-c1b5e2f12e63\") " Oct 02 02:06:38 crc kubenswrapper[4885]: I1002 02:06:38.335725 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ded68855-3099-4c7a-9239-c1b5e2f12e63-config-data\") pod \"ded68855-3099-4c7a-9239-c1b5e2f12e63\" (UID: \"ded68855-3099-4c7a-9239-c1b5e2f12e63\") " Oct 02 02:06:38 crc kubenswrapper[4885]: I1002 02:06:38.335939 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7bmlq\" (UniqueName: \"kubernetes.io/projected/ded68855-3099-4c7a-9239-c1b5e2f12e63-kube-api-access-7bmlq\") pod \"ded68855-3099-4c7a-9239-c1b5e2f12e63\" (UID: \"ded68855-3099-4c7a-9239-c1b5e2f12e63\") " Oct 02 02:06:38 crc kubenswrapper[4885]: I1002 02:06:38.336014 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/ded68855-3099-4c7a-9239-c1b5e2f12e63-nova-metadata-tls-certs\") pod \"ded68855-3099-4c7a-9239-c1b5e2f12e63\" (UID: \"ded68855-3099-4c7a-9239-c1b5e2f12e63\") " Oct 02 02:06:38 crc kubenswrapper[4885]: I1002 02:06:38.337144 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ded68855-3099-4c7a-9239-c1b5e2f12e63-logs" (OuterVolumeSpecName: "logs") pod "ded68855-3099-4c7a-9239-c1b5e2f12e63" (UID: "ded68855-3099-4c7a-9239-c1b5e2f12e63"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:06:38 crc kubenswrapper[4885]: I1002 02:06:38.344889 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ded68855-3099-4c7a-9239-c1b5e2f12e63-kube-api-access-7bmlq" (OuterVolumeSpecName: "kube-api-access-7bmlq") pod "ded68855-3099-4c7a-9239-c1b5e2f12e63" (UID: "ded68855-3099-4c7a-9239-c1b5e2f12e63"). InnerVolumeSpecName "kube-api-access-7bmlq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:06:38 crc kubenswrapper[4885]: I1002 02:06:38.364131 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ded68855-3099-4c7a-9239-c1b5e2f12e63-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ded68855-3099-4c7a-9239-c1b5e2f12e63" (UID: "ded68855-3099-4c7a-9239-c1b5e2f12e63"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:06:38 crc kubenswrapper[4885]: I1002 02:06:38.390982 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ded68855-3099-4c7a-9239-c1b5e2f12e63-config-data" (OuterVolumeSpecName: "config-data") pod "ded68855-3099-4c7a-9239-c1b5e2f12e63" (UID: "ded68855-3099-4c7a-9239-c1b5e2f12e63"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:06:38 crc kubenswrapper[4885]: I1002 02:06:38.398573 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ded68855-3099-4c7a-9239-c1b5e2f12e63-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "ded68855-3099-4c7a-9239-c1b5e2f12e63" (UID: "ded68855-3099-4c7a-9239-c1b5e2f12e63"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:06:38 crc kubenswrapper[4885]: I1002 02:06:38.474611 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7bmlq\" (UniqueName: \"kubernetes.io/projected/ded68855-3099-4c7a-9239-c1b5e2f12e63-kube-api-access-7bmlq\") on node \"crc\" DevicePath \"\"" Oct 02 02:06:38 crc kubenswrapper[4885]: I1002 02:06:38.474861 4885 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/ded68855-3099-4c7a-9239-c1b5e2f12e63-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 02:06:38 crc kubenswrapper[4885]: I1002 02:06:38.474871 4885 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ded68855-3099-4c7a-9239-c1b5e2f12e63-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:06:38 crc kubenswrapper[4885]: I1002 02:06:38.474879 4885 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ded68855-3099-4c7a-9239-c1b5e2f12e63-logs\") on node \"crc\" DevicePath \"\"" Oct 02 02:06:38 crc kubenswrapper[4885]: I1002 02:06:38.474888 4885 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ded68855-3099-4c7a-9239-c1b5e2f12e63-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 02:06:38 crc kubenswrapper[4885]: I1002 02:06:38.772691 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 02 02:06:38 crc kubenswrapper[4885]: I1002 02:06:38.776736 4885 generic.go:334] "Generic (PLEG): container finished" podID="ffc2d43b-ac70-44d3-9fa5-3dc0f6159f46" containerID="af23b39faeaa1d212bba186d27a3acdf681f31ba75d794163d93887c5df66788" exitCode=0 Oct 02 02:06:38 crc kubenswrapper[4885]: I1002 02:06:38.776813 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-cm2bt" event={"ID":"ffc2d43b-ac70-44d3-9fa5-3dc0f6159f46","Type":"ContainerDied","Data":"af23b39faeaa1d212bba186d27a3acdf681f31ba75d794163d93887c5df66788"} Oct 02 02:06:38 crc kubenswrapper[4885]: I1002 02:06:38.780350 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"77273755-f99f-4c79-a730-6e021de17e85","Type":"ContainerStarted","Data":"d8e6af61c85a1aadfbea122b344513b580c0ea16fd4bb38865bf1e1f3f0e0333"} Oct 02 02:06:38 crc kubenswrapper[4885]: I1002 02:06:38.797523 4885 generic.go:334] "Generic (PLEG): container finished" podID="ded68855-3099-4c7a-9239-c1b5e2f12e63" containerID="80bb7bfd8498043c9aee2dd1ffdbe9cd51f22691f4e8ba3f00d0b91892dfa24c" exitCode=0 Oct 02 02:06:38 crc kubenswrapper[4885]: I1002 02:06:38.797551 4885 generic.go:334] "Generic (PLEG): container finished" podID="ded68855-3099-4c7a-9239-c1b5e2f12e63" containerID="c7b1579fd3177d152948ca37a1637447fb5e907f1e2d5ae7c4897e522fb78787" exitCode=143 Oct 02 02:06:38 crc kubenswrapper[4885]: I1002 02:06:38.797593 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ded68855-3099-4c7a-9239-c1b5e2f12e63","Type":"ContainerDied","Data":"80bb7bfd8498043c9aee2dd1ffdbe9cd51f22691f4e8ba3f00d0b91892dfa24c"} Oct 02 02:06:38 crc kubenswrapper[4885]: I1002 02:06:38.797615 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ded68855-3099-4c7a-9239-c1b5e2f12e63","Type":"ContainerDied","Data":"c7b1579fd3177d152948ca37a1637447fb5e907f1e2d5ae7c4897e522fb78787"} Oct 02 02:06:38 crc kubenswrapper[4885]: I1002 02:06:38.797625 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ded68855-3099-4c7a-9239-c1b5e2f12e63","Type":"ContainerDied","Data":"32ffa290cee09ab7fa4e23ff41ae1532e9093de46d254d986ce9d28b106b9578"} Oct 02 02:06:38 crc kubenswrapper[4885]: I1002 02:06:38.797639 4885 scope.go:117] "RemoveContainer" containerID="80bb7bfd8498043c9aee2dd1ffdbe9cd51f22691f4e8ba3f00d0b91892dfa24c" Oct 02 02:06:38 crc kubenswrapper[4885]: I1002 02:06:38.797719 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 02:06:38 crc kubenswrapper[4885]: I1002 02:06:38.833209 4885 scope.go:117] "RemoveContainer" containerID="c7b1579fd3177d152948ca37a1637447fb5e907f1e2d5ae7c4897e522fb78787" Oct 02 02:06:38 crc kubenswrapper[4885]: I1002 02:06:38.883823 4885 scope.go:117] "RemoveContainer" containerID="80bb7bfd8498043c9aee2dd1ffdbe9cd51f22691f4e8ba3f00d0b91892dfa24c" Oct 02 02:06:38 crc kubenswrapper[4885]: E1002 02:06:38.885571 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"80bb7bfd8498043c9aee2dd1ffdbe9cd51f22691f4e8ba3f00d0b91892dfa24c\": container with ID starting with 80bb7bfd8498043c9aee2dd1ffdbe9cd51f22691f4e8ba3f00d0b91892dfa24c not found: ID does not exist" containerID="80bb7bfd8498043c9aee2dd1ffdbe9cd51f22691f4e8ba3f00d0b91892dfa24c" Oct 02 02:06:38 crc kubenswrapper[4885]: I1002 02:06:38.885601 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80bb7bfd8498043c9aee2dd1ffdbe9cd51f22691f4e8ba3f00d0b91892dfa24c"} err="failed to get container status \"80bb7bfd8498043c9aee2dd1ffdbe9cd51f22691f4e8ba3f00d0b91892dfa24c\": rpc error: code = NotFound desc = could not find container \"80bb7bfd8498043c9aee2dd1ffdbe9cd51f22691f4e8ba3f00d0b91892dfa24c\": container with ID starting with 80bb7bfd8498043c9aee2dd1ffdbe9cd51f22691f4e8ba3f00d0b91892dfa24c not found: ID does not exist" Oct 02 02:06:38 crc kubenswrapper[4885]: I1002 02:06:38.885622 4885 scope.go:117] "RemoveContainer" containerID="c7b1579fd3177d152948ca37a1637447fb5e907f1e2d5ae7c4897e522fb78787" Oct 02 02:06:38 crc kubenswrapper[4885]: E1002 02:06:38.887559 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c7b1579fd3177d152948ca37a1637447fb5e907f1e2d5ae7c4897e522fb78787\": container with ID starting with c7b1579fd3177d152948ca37a1637447fb5e907f1e2d5ae7c4897e522fb78787 not found: ID does not exist" containerID="c7b1579fd3177d152948ca37a1637447fb5e907f1e2d5ae7c4897e522fb78787" Oct 02 02:06:38 crc kubenswrapper[4885]: I1002 02:06:38.887582 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c7b1579fd3177d152948ca37a1637447fb5e907f1e2d5ae7c4897e522fb78787"} err="failed to get container status \"c7b1579fd3177d152948ca37a1637447fb5e907f1e2d5ae7c4897e522fb78787\": rpc error: code = NotFound desc = could not find container \"c7b1579fd3177d152948ca37a1637447fb5e907f1e2d5ae7c4897e522fb78787\": container with ID starting with c7b1579fd3177d152948ca37a1637447fb5e907f1e2d5ae7c4897e522fb78787 not found: ID does not exist" Oct 02 02:06:38 crc kubenswrapper[4885]: I1002 02:06:38.887596 4885 scope.go:117] "RemoveContainer" containerID="80bb7bfd8498043c9aee2dd1ffdbe9cd51f22691f4e8ba3f00d0b91892dfa24c" Oct 02 02:06:38 crc kubenswrapper[4885]: I1002 02:06:38.887818 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80bb7bfd8498043c9aee2dd1ffdbe9cd51f22691f4e8ba3f00d0b91892dfa24c"} err="failed to get container status \"80bb7bfd8498043c9aee2dd1ffdbe9cd51f22691f4e8ba3f00d0b91892dfa24c\": rpc error: code = NotFound desc = could not find container \"80bb7bfd8498043c9aee2dd1ffdbe9cd51f22691f4e8ba3f00d0b91892dfa24c\": container with ID starting with 80bb7bfd8498043c9aee2dd1ffdbe9cd51f22691f4e8ba3f00d0b91892dfa24c not found: ID does not exist" Oct 02 02:06:38 crc kubenswrapper[4885]: I1002 02:06:38.887833 4885 scope.go:117] "RemoveContainer" containerID="c7b1579fd3177d152948ca37a1637447fb5e907f1e2d5ae7c4897e522fb78787" Oct 02 02:06:38 crc kubenswrapper[4885]: I1002 02:06:38.888020 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c7b1579fd3177d152948ca37a1637447fb5e907f1e2d5ae7c4897e522fb78787"} err="failed to get container status \"c7b1579fd3177d152948ca37a1637447fb5e907f1e2d5ae7c4897e522fb78787\": rpc error: code = NotFound desc = could not find container \"c7b1579fd3177d152948ca37a1637447fb5e907f1e2d5ae7c4897e522fb78787\": container with ID starting with c7b1579fd3177d152948ca37a1637447fb5e907f1e2d5ae7c4897e522fb78787 not found: ID does not exist" Oct 02 02:06:38 crc kubenswrapper[4885]: I1002 02:06:38.899488 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 02:06:38 crc kubenswrapper[4885]: W1002 02:06:38.918135 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6895f52a_cc09_4eac_9f57_32f91106afb9.slice/crio-9adaa4894606bb5b8256c9368f407b32063c570f722a4768c232c7f5fd03469a WatchSource:0}: Error finding container 9adaa4894606bb5b8256c9368f407b32063c570f722a4768c232c7f5fd03469a: Status 404 returned error can't find the container with id 9adaa4894606bb5b8256c9368f407b32063c570f722a4768c232c7f5fd03469a Oct 02 02:06:38 crc kubenswrapper[4885]: I1002 02:06:38.920593 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 02:06:38 crc kubenswrapper[4885]: I1002 02:06:38.943420 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 02:06:38 crc kubenswrapper[4885]: I1002 02:06:38.962254 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 02 02:06:38 crc kubenswrapper[4885]: E1002 02:06:38.962850 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ded68855-3099-4c7a-9239-c1b5e2f12e63" containerName="nova-metadata-metadata" Oct 02 02:06:38 crc kubenswrapper[4885]: I1002 02:06:38.962866 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="ded68855-3099-4c7a-9239-c1b5e2f12e63" containerName="nova-metadata-metadata" Oct 02 02:06:38 crc kubenswrapper[4885]: E1002 02:06:38.962909 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ded68855-3099-4c7a-9239-c1b5e2f12e63" containerName="nova-metadata-log" Oct 02 02:06:38 crc kubenswrapper[4885]: I1002 02:06:38.962916 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="ded68855-3099-4c7a-9239-c1b5e2f12e63" containerName="nova-metadata-log" Oct 02 02:06:38 crc kubenswrapper[4885]: I1002 02:06:38.963075 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="ded68855-3099-4c7a-9239-c1b5e2f12e63" containerName="nova-metadata-metadata" Oct 02 02:06:38 crc kubenswrapper[4885]: I1002 02:06:38.963093 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="ded68855-3099-4c7a-9239-c1b5e2f12e63" containerName="nova-metadata-log" Oct 02 02:06:38 crc kubenswrapper[4885]: I1002 02:06:38.963964 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 02:06:38 crc kubenswrapper[4885]: I1002 02:06:38.968990 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 02 02:06:38 crc kubenswrapper[4885]: I1002 02:06:38.969597 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 02 02:06:39 crc kubenswrapper[4885]: I1002 02:06:39.000102 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2h5c6\" (UniqueName: \"kubernetes.io/projected/3555c77f-8b14-48c5-a9e1-42e8856dfe19-kube-api-access-2h5c6\") pod \"nova-metadata-0\" (UID: \"3555c77f-8b14-48c5-a9e1-42e8856dfe19\") " pod="openstack/nova-metadata-0" Oct 02 02:06:39 crc kubenswrapper[4885]: I1002 02:06:39.000147 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3555c77f-8b14-48c5-a9e1-42e8856dfe19-config-data\") pod \"nova-metadata-0\" (UID: \"3555c77f-8b14-48c5-a9e1-42e8856dfe19\") " pod="openstack/nova-metadata-0" Oct 02 02:06:39 crc kubenswrapper[4885]: I1002 02:06:39.000172 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3555c77f-8b14-48c5-a9e1-42e8856dfe19-logs\") pod \"nova-metadata-0\" (UID: \"3555c77f-8b14-48c5-a9e1-42e8856dfe19\") " pod="openstack/nova-metadata-0" Oct 02 02:06:39 crc kubenswrapper[4885]: I1002 02:06:39.000194 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3555c77f-8b14-48c5-a9e1-42e8856dfe19-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"3555c77f-8b14-48c5-a9e1-42e8856dfe19\") " pod="openstack/nova-metadata-0" Oct 02 02:06:39 crc kubenswrapper[4885]: I1002 02:06:39.000230 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/3555c77f-8b14-48c5-a9e1-42e8856dfe19-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"3555c77f-8b14-48c5-a9e1-42e8856dfe19\") " pod="openstack/nova-metadata-0" Oct 02 02:06:39 crc kubenswrapper[4885]: I1002 02:06:39.029856 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 02:06:39 crc kubenswrapper[4885]: I1002 02:06:39.108237 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2h5c6\" (UniqueName: \"kubernetes.io/projected/3555c77f-8b14-48c5-a9e1-42e8856dfe19-kube-api-access-2h5c6\") pod \"nova-metadata-0\" (UID: \"3555c77f-8b14-48c5-a9e1-42e8856dfe19\") " pod="openstack/nova-metadata-0" Oct 02 02:06:39 crc kubenswrapper[4885]: I1002 02:06:39.108307 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3555c77f-8b14-48c5-a9e1-42e8856dfe19-config-data\") pod \"nova-metadata-0\" (UID: \"3555c77f-8b14-48c5-a9e1-42e8856dfe19\") " pod="openstack/nova-metadata-0" Oct 02 02:06:39 crc kubenswrapper[4885]: I1002 02:06:39.108334 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3555c77f-8b14-48c5-a9e1-42e8856dfe19-logs\") pod \"nova-metadata-0\" (UID: \"3555c77f-8b14-48c5-a9e1-42e8856dfe19\") " pod="openstack/nova-metadata-0" Oct 02 02:06:39 crc kubenswrapper[4885]: I1002 02:06:39.108368 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3555c77f-8b14-48c5-a9e1-42e8856dfe19-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"3555c77f-8b14-48c5-a9e1-42e8856dfe19\") " pod="openstack/nova-metadata-0" Oct 02 02:06:39 crc kubenswrapper[4885]: I1002 02:06:39.108473 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/3555c77f-8b14-48c5-a9e1-42e8856dfe19-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"3555c77f-8b14-48c5-a9e1-42e8856dfe19\") " pod="openstack/nova-metadata-0" Oct 02 02:06:39 crc kubenswrapper[4885]: I1002 02:06:39.109013 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3555c77f-8b14-48c5-a9e1-42e8856dfe19-logs\") pod \"nova-metadata-0\" (UID: \"3555c77f-8b14-48c5-a9e1-42e8856dfe19\") " pod="openstack/nova-metadata-0" Oct 02 02:06:39 crc kubenswrapper[4885]: I1002 02:06:39.119037 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3555c77f-8b14-48c5-a9e1-42e8856dfe19-config-data\") pod \"nova-metadata-0\" (UID: \"3555c77f-8b14-48c5-a9e1-42e8856dfe19\") " pod="openstack/nova-metadata-0" Oct 02 02:06:39 crc kubenswrapper[4885]: I1002 02:06:39.119498 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3555c77f-8b14-48c5-a9e1-42e8856dfe19-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"3555c77f-8b14-48c5-a9e1-42e8856dfe19\") " pod="openstack/nova-metadata-0" Oct 02 02:06:39 crc kubenswrapper[4885]: I1002 02:06:39.125798 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/3555c77f-8b14-48c5-a9e1-42e8856dfe19-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"3555c77f-8b14-48c5-a9e1-42e8856dfe19\") " pod="openstack/nova-metadata-0" Oct 02 02:06:39 crc kubenswrapper[4885]: I1002 02:06:39.155771 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2h5c6\" (UniqueName: \"kubernetes.io/projected/3555c77f-8b14-48c5-a9e1-42e8856dfe19-kube-api-access-2h5c6\") pod \"nova-metadata-0\" (UID: \"3555c77f-8b14-48c5-a9e1-42e8856dfe19\") " pod="openstack/nova-metadata-0" Oct 02 02:06:39 crc kubenswrapper[4885]: I1002 02:06:39.330522 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 02:06:39 crc kubenswrapper[4885]: I1002 02:06:39.829049 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"77273755-f99f-4c79-a730-6e021de17e85","Type":"ContainerStarted","Data":"5a52eb88553b095bb2a574b5316b37218677f9a500d33dacaec88407cd140131"} Oct 02 02:06:39 crc kubenswrapper[4885]: I1002 02:06:39.829439 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"77273755-f99f-4c79-a730-6e021de17e85","Type":"ContainerStarted","Data":"21b6364614c91b5470857491a02ccefcda4e2881eccde16bcdcc041a8f8fab10"} Oct 02 02:06:39 crc kubenswrapper[4885]: I1002 02:06:39.829842 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 02:06:39 crc kubenswrapper[4885]: I1002 02:06:39.855551 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"6895f52a-cc09-4eac-9f57-32f91106afb9","Type":"ContainerStarted","Data":"77022366cfd5951f2c9ecb6edf329467419834685af3580f17b8fdf85b2921cb"} Oct 02 02:06:39 crc kubenswrapper[4885]: I1002 02:06:39.855613 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"6895f52a-cc09-4eac-9f57-32f91106afb9","Type":"ContainerStarted","Data":"9adaa4894606bb5b8256c9368f407b32063c570f722a4768c232c7f5fd03469a"} Oct 02 02:06:39 crc kubenswrapper[4885]: I1002 02:06:39.858306 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.858295036 podStartE2EDuration="2.858295036s" podCreationTimestamp="2025-10-02 02:06:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:06:39.856469621 +0000 UTC m=+1188.668217060" watchObservedRunningTime="2025-10-02 02:06:39.858295036 +0000 UTC m=+1188.670042435" Oct 02 02:06:39 crc kubenswrapper[4885]: I1002 02:06:39.885118 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.885099471 podStartE2EDuration="2.885099471s" podCreationTimestamp="2025-10-02 02:06:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:06:39.873607662 +0000 UTC m=+1188.685355081" watchObservedRunningTime="2025-10-02 02:06:39.885099471 +0000 UTC m=+1188.696846870" Oct 02 02:06:40 crc kubenswrapper[4885]: I1002 02:06:40.074192 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ded68855-3099-4c7a-9239-c1b5e2f12e63" path="/var/lib/kubelet/pods/ded68855-3099-4c7a-9239-c1b5e2f12e63/volumes" Oct 02 02:06:40 crc kubenswrapper[4885]: I1002 02:06:40.154781 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 02 02:06:40 crc kubenswrapper[4885]: I1002 02:06:40.155354 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="faff5405-00c6-4d75-bbca-77f9230c2811" containerName="kube-state-metrics" containerID="cri-o://80e36c78fd1078a8f7272a7a784beeca583573d3867f310d6a61e997fce2d26a" gracePeriod=30 Oct 02 02:06:40 crc kubenswrapper[4885]: I1002 02:06:40.264350 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-cm2bt" Oct 02 02:06:40 crc kubenswrapper[4885]: I1002 02:06:40.336362 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ff27s\" (UniqueName: \"kubernetes.io/projected/ffc2d43b-ac70-44d3-9fa5-3dc0f6159f46-kube-api-access-ff27s\") pod \"ffc2d43b-ac70-44d3-9fa5-3dc0f6159f46\" (UID: \"ffc2d43b-ac70-44d3-9fa5-3dc0f6159f46\") " Oct 02 02:06:40 crc kubenswrapper[4885]: I1002 02:06:40.336498 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ffc2d43b-ac70-44d3-9fa5-3dc0f6159f46-combined-ca-bundle\") pod \"ffc2d43b-ac70-44d3-9fa5-3dc0f6159f46\" (UID: \"ffc2d43b-ac70-44d3-9fa5-3dc0f6159f46\") " Oct 02 02:06:40 crc kubenswrapper[4885]: I1002 02:06:40.336529 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ffc2d43b-ac70-44d3-9fa5-3dc0f6159f46-scripts\") pod \"ffc2d43b-ac70-44d3-9fa5-3dc0f6159f46\" (UID: \"ffc2d43b-ac70-44d3-9fa5-3dc0f6159f46\") " Oct 02 02:06:40 crc kubenswrapper[4885]: I1002 02:06:40.336552 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ffc2d43b-ac70-44d3-9fa5-3dc0f6159f46-config-data\") pod \"ffc2d43b-ac70-44d3-9fa5-3dc0f6159f46\" (UID: \"ffc2d43b-ac70-44d3-9fa5-3dc0f6159f46\") " Oct 02 02:06:40 crc kubenswrapper[4885]: I1002 02:06:40.342763 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ffc2d43b-ac70-44d3-9fa5-3dc0f6159f46-kube-api-access-ff27s" (OuterVolumeSpecName: "kube-api-access-ff27s") pod "ffc2d43b-ac70-44d3-9fa5-3dc0f6159f46" (UID: "ffc2d43b-ac70-44d3-9fa5-3dc0f6159f46"). InnerVolumeSpecName "kube-api-access-ff27s". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:06:40 crc kubenswrapper[4885]: I1002 02:06:40.346415 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ffc2d43b-ac70-44d3-9fa5-3dc0f6159f46-scripts" (OuterVolumeSpecName: "scripts") pod "ffc2d43b-ac70-44d3-9fa5-3dc0f6159f46" (UID: "ffc2d43b-ac70-44d3-9fa5-3dc0f6159f46"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:06:40 crc kubenswrapper[4885]: I1002 02:06:40.373478 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ffc2d43b-ac70-44d3-9fa5-3dc0f6159f46-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ffc2d43b-ac70-44d3-9fa5-3dc0f6159f46" (UID: "ffc2d43b-ac70-44d3-9fa5-3dc0f6159f46"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:06:40 crc kubenswrapper[4885]: I1002 02:06:40.390193 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ffc2d43b-ac70-44d3-9fa5-3dc0f6159f46-config-data" (OuterVolumeSpecName: "config-data") pod "ffc2d43b-ac70-44d3-9fa5-3dc0f6159f46" (UID: "ffc2d43b-ac70-44d3-9fa5-3dc0f6159f46"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:06:40 crc kubenswrapper[4885]: I1002 02:06:40.438986 4885 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ffc2d43b-ac70-44d3-9fa5-3dc0f6159f46-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:06:40 crc kubenswrapper[4885]: I1002 02:06:40.439017 4885 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ffc2d43b-ac70-44d3-9fa5-3dc0f6159f46-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 02:06:40 crc kubenswrapper[4885]: I1002 02:06:40.439026 4885 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ffc2d43b-ac70-44d3-9fa5-3dc0f6159f46-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 02:06:40 crc kubenswrapper[4885]: I1002 02:06:40.439034 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ff27s\" (UniqueName: \"kubernetes.io/projected/ffc2d43b-ac70-44d3-9fa5-3dc0f6159f46-kube-api-access-ff27s\") on node \"crc\" DevicePath \"\"" Oct 02 02:06:40 crc kubenswrapper[4885]: I1002 02:06:40.572974 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 02 02:06:40 crc kubenswrapper[4885]: I1002 02:06:40.745947 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2ngf5\" (UniqueName: \"kubernetes.io/projected/faff5405-00c6-4d75-bbca-77f9230c2811-kube-api-access-2ngf5\") pod \"faff5405-00c6-4d75-bbca-77f9230c2811\" (UID: \"faff5405-00c6-4d75-bbca-77f9230c2811\") " Oct 02 02:06:40 crc kubenswrapper[4885]: I1002 02:06:40.749059 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/faff5405-00c6-4d75-bbca-77f9230c2811-kube-api-access-2ngf5" (OuterVolumeSpecName: "kube-api-access-2ngf5") pod "faff5405-00c6-4d75-bbca-77f9230c2811" (UID: "faff5405-00c6-4d75-bbca-77f9230c2811"). InnerVolumeSpecName "kube-api-access-2ngf5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:06:40 crc kubenswrapper[4885]: I1002 02:06:40.849081 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2ngf5\" (UniqueName: \"kubernetes.io/projected/faff5405-00c6-4d75-bbca-77f9230c2811-kube-api-access-2ngf5\") on node \"crc\" DevicePath \"\"" Oct 02 02:06:40 crc kubenswrapper[4885]: I1002 02:06:40.872539 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 02 02:06:40 crc kubenswrapper[4885]: E1002 02:06:40.872965 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ffc2d43b-ac70-44d3-9fa5-3dc0f6159f46" containerName="nova-cell1-conductor-db-sync" Oct 02 02:06:40 crc kubenswrapper[4885]: I1002 02:06:40.872982 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="ffc2d43b-ac70-44d3-9fa5-3dc0f6159f46" containerName="nova-cell1-conductor-db-sync" Oct 02 02:06:40 crc kubenswrapper[4885]: E1002 02:06:40.873003 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="faff5405-00c6-4d75-bbca-77f9230c2811" containerName="kube-state-metrics" Oct 02 02:06:40 crc kubenswrapper[4885]: I1002 02:06:40.873009 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="faff5405-00c6-4d75-bbca-77f9230c2811" containerName="kube-state-metrics" Oct 02 02:06:40 crc kubenswrapper[4885]: I1002 02:06:40.873196 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="ffc2d43b-ac70-44d3-9fa5-3dc0f6159f46" containerName="nova-cell1-conductor-db-sync" Oct 02 02:06:40 crc kubenswrapper[4885]: I1002 02:06:40.873205 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="faff5405-00c6-4d75-bbca-77f9230c2811" containerName="kube-state-metrics" Oct 02 02:06:40 crc kubenswrapper[4885]: I1002 02:06:40.873862 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 02 02:06:40 crc kubenswrapper[4885]: I1002 02:06:40.882765 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 02 02:06:40 crc kubenswrapper[4885]: I1002 02:06:40.884398 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-cm2bt" event={"ID":"ffc2d43b-ac70-44d3-9fa5-3dc0f6159f46","Type":"ContainerDied","Data":"fedb6345713aae2d9a1df14e6c41474bb78fb2d82f4750a5d2a9e79aef7f8d12"} Oct 02 02:06:40 crc kubenswrapper[4885]: I1002 02:06:40.884438 4885 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fedb6345713aae2d9a1df14e6c41474bb78fb2d82f4750a5d2a9e79aef7f8d12" Oct 02 02:06:40 crc kubenswrapper[4885]: I1002 02:06:40.884537 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-cm2bt" Oct 02 02:06:40 crc kubenswrapper[4885]: I1002 02:06:40.891608 4885 generic.go:334] "Generic (PLEG): container finished" podID="faff5405-00c6-4d75-bbca-77f9230c2811" containerID="80e36c78fd1078a8f7272a7a784beeca583573d3867f310d6a61e997fce2d26a" exitCode=2 Oct 02 02:06:40 crc kubenswrapper[4885]: I1002 02:06:40.891688 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"faff5405-00c6-4d75-bbca-77f9230c2811","Type":"ContainerDied","Data":"80e36c78fd1078a8f7272a7a784beeca583573d3867f310d6a61e997fce2d26a"} Oct 02 02:06:40 crc kubenswrapper[4885]: I1002 02:06:40.891714 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"faff5405-00c6-4d75-bbca-77f9230c2811","Type":"ContainerDied","Data":"49bd68dd50e5f1991581195877fe3325d00d374d02ade0c4582071dc275342dc"} Oct 02 02:06:40 crc kubenswrapper[4885]: I1002 02:06:40.891731 4885 scope.go:117] "RemoveContainer" containerID="80e36c78fd1078a8f7272a7a784beeca583573d3867f310d6a61e997fce2d26a" Oct 02 02:06:40 crc kubenswrapper[4885]: I1002 02:06:40.891911 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 02 02:06:40 crc kubenswrapper[4885]: I1002 02:06:40.908505 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3555c77f-8b14-48c5-a9e1-42e8856dfe19","Type":"ContainerStarted","Data":"ba1f877c8c7a3dc96a1a7cbad1d575537e1a8afaf77c80ebce637df8113780d6"} Oct 02 02:06:40 crc kubenswrapper[4885]: I1002 02:06:40.909445 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3555c77f-8b14-48c5-a9e1-42e8856dfe19","Type":"ContainerStarted","Data":"f52ccc636851ade8d152837715411fd00cf473e5b713d8ff129c166e1d595b01"} Oct 02 02:06:40 crc kubenswrapper[4885]: I1002 02:06:40.909458 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3555c77f-8b14-48c5-a9e1-42e8856dfe19","Type":"ContainerStarted","Data":"2e53ca06a2d7abcfee83265ec4e450bf329550e5d651058c1dfbd19240513dc5"} Oct 02 02:06:40 crc kubenswrapper[4885]: I1002 02:06:40.932886 4885 scope.go:117] "RemoveContainer" containerID="80e36c78fd1078a8f7272a7a784beeca583573d3867f310d6a61e997fce2d26a" Oct 02 02:06:40 crc kubenswrapper[4885]: E1002 02:06:40.933380 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"80e36c78fd1078a8f7272a7a784beeca583573d3867f310d6a61e997fce2d26a\": container with ID starting with 80e36c78fd1078a8f7272a7a784beeca583573d3867f310d6a61e997fce2d26a not found: ID does not exist" containerID="80e36c78fd1078a8f7272a7a784beeca583573d3867f310d6a61e997fce2d26a" Oct 02 02:06:40 crc kubenswrapper[4885]: I1002 02:06:40.933411 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80e36c78fd1078a8f7272a7a784beeca583573d3867f310d6a61e997fce2d26a"} err="failed to get container status \"80e36c78fd1078a8f7272a7a784beeca583573d3867f310d6a61e997fce2d26a\": rpc error: code = NotFound desc = could not find container \"80e36c78fd1078a8f7272a7a784beeca583573d3867f310d6a61e997fce2d26a\": container with ID starting with 80e36c78fd1078a8f7272a7a784beeca583573d3867f310d6a61e997fce2d26a not found: ID does not exist" Oct 02 02:06:40 crc kubenswrapper[4885]: I1002 02:06:40.944813 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.944796707 podStartE2EDuration="2.944796707s" podCreationTimestamp="2025-10-02 02:06:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:06:40.932682799 +0000 UTC m=+1189.744430218" watchObservedRunningTime="2025-10-02 02:06:40.944796707 +0000 UTC m=+1189.756544106" Oct 02 02:06:40 crc kubenswrapper[4885]: I1002 02:06:40.950843 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6f37337-ce77-444d-9efa-852d0db3aabb-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"b6f37337-ce77-444d-9efa-852d0db3aabb\") " pod="openstack/nova-cell1-conductor-0" Oct 02 02:06:40 crc kubenswrapper[4885]: I1002 02:06:40.951553 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6f37337-ce77-444d-9efa-852d0db3aabb-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"b6f37337-ce77-444d-9efa-852d0db3aabb\") " pod="openstack/nova-cell1-conductor-0" Oct 02 02:06:40 crc kubenswrapper[4885]: I1002 02:06:40.951633 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g8wj8\" (UniqueName: \"kubernetes.io/projected/b6f37337-ce77-444d-9efa-852d0db3aabb-kube-api-access-g8wj8\") pod \"nova-cell1-conductor-0\" (UID: \"b6f37337-ce77-444d-9efa-852d0db3aabb\") " pod="openstack/nova-cell1-conductor-0" Oct 02 02:06:40 crc kubenswrapper[4885]: I1002 02:06:40.954347 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 02 02:06:40 crc kubenswrapper[4885]: I1002 02:06:40.965773 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 02 02:06:40 crc kubenswrapper[4885]: I1002 02:06:40.977088 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 02 02:06:40 crc kubenswrapper[4885]: I1002 02:06:40.978330 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 02 02:06:40 crc kubenswrapper[4885]: I1002 02:06:40.980312 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Oct 02 02:06:40 crc kubenswrapper[4885]: I1002 02:06:40.984034 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Oct 02 02:06:40 crc kubenswrapper[4885]: I1002 02:06:40.987473 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 02 02:06:41 crc kubenswrapper[4885]: I1002 02:06:41.052845 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6f37337-ce77-444d-9efa-852d0db3aabb-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"b6f37337-ce77-444d-9efa-852d0db3aabb\") " pod="openstack/nova-cell1-conductor-0" Oct 02 02:06:41 crc kubenswrapper[4885]: I1002 02:06:41.052897 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/b5cb9a09-d2c8-4138-9eea-f404bd2e5034-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"b5cb9a09-d2c8-4138-9eea-f404bd2e5034\") " pod="openstack/kube-state-metrics-0" Oct 02 02:06:41 crc kubenswrapper[4885]: I1002 02:06:41.052913 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/b5cb9a09-d2c8-4138-9eea-f404bd2e5034-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"b5cb9a09-d2c8-4138-9eea-f404bd2e5034\") " pod="openstack/kube-state-metrics-0" Oct 02 02:06:41 crc kubenswrapper[4885]: I1002 02:06:41.052941 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6f37337-ce77-444d-9efa-852d0db3aabb-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"b6f37337-ce77-444d-9efa-852d0db3aabb\") " pod="openstack/nova-cell1-conductor-0" Oct 02 02:06:41 crc kubenswrapper[4885]: I1002 02:06:41.052960 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5cb9a09-d2c8-4138-9eea-f404bd2e5034-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"b5cb9a09-d2c8-4138-9eea-f404bd2e5034\") " pod="openstack/kube-state-metrics-0" Oct 02 02:06:41 crc kubenswrapper[4885]: I1002 02:06:41.053198 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mv2sp\" (UniqueName: \"kubernetes.io/projected/b5cb9a09-d2c8-4138-9eea-f404bd2e5034-kube-api-access-mv2sp\") pod \"kube-state-metrics-0\" (UID: \"b5cb9a09-d2c8-4138-9eea-f404bd2e5034\") " pod="openstack/kube-state-metrics-0" Oct 02 02:06:41 crc kubenswrapper[4885]: I1002 02:06:41.053303 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g8wj8\" (UniqueName: \"kubernetes.io/projected/b6f37337-ce77-444d-9efa-852d0db3aabb-kube-api-access-g8wj8\") pod \"nova-cell1-conductor-0\" (UID: \"b6f37337-ce77-444d-9efa-852d0db3aabb\") " pod="openstack/nova-cell1-conductor-0" Oct 02 02:06:41 crc kubenswrapper[4885]: I1002 02:06:41.056134 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6f37337-ce77-444d-9efa-852d0db3aabb-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"b6f37337-ce77-444d-9efa-852d0db3aabb\") " pod="openstack/nova-cell1-conductor-0" Oct 02 02:06:41 crc kubenswrapper[4885]: I1002 02:06:41.060916 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6f37337-ce77-444d-9efa-852d0db3aabb-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"b6f37337-ce77-444d-9efa-852d0db3aabb\") " pod="openstack/nova-cell1-conductor-0" Oct 02 02:06:41 crc kubenswrapper[4885]: I1002 02:06:41.081775 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g8wj8\" (UniqueName: \"kubernetes.io/projected/b6f37337-ce77-444d-9efa-852d0db3aabb-kube-api-access-g8wj8\") pod \"nova-cell1-conductor-0\" (UID: \"b6f37337-ce77-444d-9efa-852d0db3aabb\") " pod="openstack/nova-cell1-conductor-0" Oct 02 02:06:41 crc kubenswrapper[4885]: I1002 02:06:41.155439 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mv2sp\" (UniqueName: \"kubernetes.io/projected/b5cb9a09-d2c8-4138-9eea-f404bd2e5034-kube-api-access-mv2sp\") pod \"kube-state-metrics-0\" (UID: \"b5cb9a09-d2c8-4138-9eea-f404bd2e5034\") " pod="openstack/kube-state-metrics-0" Oct 02 02:06:41 crc kubenswrapper[4885]: I1002 02:06:41.155616 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/b5cb9a09-d2c8-4138-9eea-f404bd2e5034-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"b5cb9a09-d2c8-4138-9eea-f404bd2e5034\") " pod="openstack/kube-state-metrics-0" Oct 02 02:06:41 crc kubenswrapper[4885]: I1002 02:06:41.155678 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/b5cb9a09-d2c8-4138-9eea-f404bd2e5034-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"b5cb9a09-d2c8-4138-9eea-f404bd2e5034\") " pod="openstack/kube-state-metrics-0" Oct 02 02:06:41 crc kubenswrapper[4885]: I1002 02:06:41.156301 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5cb9a09-d2c8-4138-9eea-f404bd2e5034-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"b5cb9a09-d2c8-4138-9eea-f404bd2e5034\") " pod="openstack/kube-state-metrics-0" Oct 02 02:06:41 crc kubenswrapper[4885]: I1002 02:06:41.160187 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/b5cb9a09-d2c8-4138-9eea-f404bd2e5034-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"b5cb9a09-d2c8-4138-9eea-f404bd2e5034\") " pod="openstack/kube-state-metrics-0" Oct 02 02:06:41 crc kubenswrapper[4885]: I1002 02:06:41.160831 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/b5cb9a09-d2c8-4138-9eea-f404bd2e5034-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"b5cb9a09-d2c8-4138-9eea-f404bd2e5034\") " pod="openstack/kube-state-metrics-0" Oct 02 02:06:41 crc kubenswrapper[4885]: I1002 02:06:41.161285 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5cb9a09-d2c8-4138-9eea-f404bd2e5034-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"b5cb9a09-d2c8-4138-9eea-f404bd2e5034\") " pod="openstack/kube-state-metrics-0" Oct 02 02:06:41 crc kubenswrapper[4885]: I1002 02:06:41.179590 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mv2sp\" (UniqueName: \"kubernetes.io/projected/b5cb9a09-d2c8-4138-9eea-f404bd2e5034-kube-api-access-mv2sp\") pod \"kube-state-metrics-0\" (UID: \"b5cb9a09-d2c8-4138-9eea-f404bd2e5034\") " pod="openstack/kube-state-metrics-0" Oct 02 02:06:41 crc kubenswrapper[4885]: I1002 02:06:41.194020 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 02 02:06:41 crc kubenswrapper[4885]: I1002 02:06:41.301607 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 02 02:06:41 crc kubenswrapper[4885]: I1002 02:06:41.627235 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 02 02:06:41 crc kubenswrapper[4885]: I1002 02:06:41.816454 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 02 02:06:41 crc kubenswrapper[4885]: W1002 02:06:41.820817 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb5cb9a09_d2c8_4138_9eea_f404bd2e5034.slice/crio-f93e734af3f3b98d6324a3839faf5189bebb3e9f2a0816f3977a002c547ee4b0 WatchSource:0}: Error finding container f93e734af3f3b98d6324a3839faf5189bebb3e9f2a0816f3977a002c547ee4b0: Status 404 returned error can't find the container with id f93e734af3f3b98d6324a3839faf5189bebb3e9f2a0816f3977a002c547ee4b0 Oct 02 02:06:41 crc kubenswrapper[4885]: I1002 02:06:41.823299 4885 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 02:06:41 crc kubenswrapper[4885]: I1002 02:06:41.887436 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 02:06:41 crc kubenswrapper[4885]: I1002 02:06:41.887728 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7617cdd5-023a-427d-8365-52dbfc6cf20a" containerName="ceilometer-central-agent" containerID="cri-o://1a275dbd70e21d0358dfdb3b1ee5eb025a0342c7445566bddcc00930eaa277f3" gracePeriod=30 Oct 02 02:06:41 crc kubenswrapper[4885]: I1002 02:06:41.887770 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7617cdd5-023a-427d-8365-52dbfc6cf20a" containerName="proxy-httpd" containerID="cri-o://737b1e2d7170ba0c915e063b22c387f8c1b2dcc4d7cb99b2b87f8998b9369cbb" gracePeriod=30 Oct 02 02:06:41 crc kubenswrapper[4885]: I1002 02:06:41.887877 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7617cdd5-023a-427d-8365-52dbfc6cf20a" containerName="sg-core" containerID="cri-o://4d15decee345cf96133f34e791bdd5f7eaec17200d66f3a76a057701de1be4c6" gracePeriod=30 Oct 02 02:06:41 crc kubenswrapper[4885]: I1002 02:06:41.887895 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7617cdd5-023a-427d-8365-52dbfc6cf20a" containerName="ceilometer-notification-agent" containerID="cri-o://5467c6f206f6d2e7c0a79b0aa4ab9d08a4129ba7f16275ccee62fec86ad3bf2e" gracePeriod=30 Oct 02 02:06:41 crc kubenswrapper[4885]: I1002 02:06:41.943551 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"b6f37337-ce77-444d-9efa-852d0db3aabb","Type":"ContainerStarted","Data":"70c560f0ca12d2978d00c0fac295f637e561dd49af6c485f33430aaccad04ce9"} Oct 02 02:06:41 crc kubenswrapper[4885]: I1002 02:06:41.950378 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"b5cb9a09-d2c8-4138-9eea-f404bd2e5034","Type":"ContainerStarted","Data":"f93e734af3f3b98d6324a3839faf5189bebb3e9f2a0816f3977a002c547ee4b0"} Oct 02 02:06:42 crc kubenswrapper[4885]: I1002 02:06:42.061895 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="faff5405-00c6-4d75-bbca-77f9230c2811" path="/var/lib/kubelet/pods/faff5405-00c6-4d75-bbca-77f9230c2811/volumes" Oct 02 02:06:42 crc kubenswrapper[4885]: I1002 02:06:42.963080 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"b6f37337-ce77-444d-9efa-852d0db3aabb","Type":"ContainerStarted","Data":"7cf89bb4c0077aac6296f42ff6a2f7d5d0a5c6544d905f69f5218233ffaa9e11"} Oct 02 02:06:42 crc kubenswrapper[4885]: I1002 02:06:42.963438 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Oct 02 02:06:42 crc kubenswrapper[4885]: I1002 02:06:42.964838 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"b5cb9a09-d2c8-4138-9eea-f404bd2e5034","Type":"ContainerStarted","Data":"11d84422932866b2bcff16c0bebde8cad6a8647303a2b40414a0952f3a4bbfa7"} Oct 02 02:06:42 crc kubenswrapper[4885]: I1002 02:06:42.965233 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 02 02:06:42 crc kubenswrapper[4885]: I1002 02:06:42.967341 4885 generic.go:334] "Generic (PLEG): container finished" podID="7617cdd5-023a-427d-8365-52dbfc6cf20a" containerID="737b1e2d7170ba0c915e063b22c387f8c1b2dcc4d7cb99b2b87f8998b9369cbb" exitCode=0 Oct 02 02:06:42 crc kubenswrapper[4885]: I1002 02:06:42.967367 4885 generic.go:334] "Generic (PLEG): container finished" podID="7617cdd5-023a-427d-8365-52dbfc6cf20a" containerID="4d15decee345cf96133f34e791bdd5f7eaec17200d66f3a76a057701de1be4c6" exitCode=2 Oct 02 02:06:42 crc kubenswrapper[4885]: I1002 02:06:42.967376 4885 generic.go:334] "Generic (PLEG): container finished" podID="7617cdd5-023a-427d-8365-52dbfc6cf20a" containerID="1a275dbd70e21d0358dfdb3b1ee5eb025a0342c7445566bddcc00930eaa277f3" exitCode=0 Oct 02 02:06:42 crc kubenswrapper[4885]: I1002 02:06:42.967397 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7617cdd5-023a-427d-8365-52dbfc6cf20a","Type":"ContainerDied","Data":"737b1e2d7170ba0c915e063b22c387f8c1b2dcc4d7cb99b2b87f8998b9369cbb"} Oct 02 02:06:42 crc kubenswrapper[4885]: I1002 02:06:42.967421 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7617cdd5-023a-427d-8365-52dbfc6cf20a","Type":"ContainerDied","Data":"4d15decee345cf96133f34e791bdd5f7eaec17200d66f3a76a057701de1be4c6"} Oct 02 02:06:42 crc kubenswrapper[4885]: I1002 02:06:42.967435 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7617cdd5-023a-427d-8365-52dbfc6cf20a","Type":"ContainerDied","Data":"1a275dbd70e21d0358dfdb3b1ee5eb025a0342c7445566bddcc00930eaa277f3"} Oct 02 02:06:42 crc kubenswrapper[4885]: I1002 02:06:42.984592 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.984578188 podStartE2EDuration="2.984578188s" podCreationTimestamp="2025-10-02 02:06:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:06:42.977568576 +0000 UTC m=+1191.789315975" watchObservedRunningTime="2025-10-02 02:06:42.984578188 +0000 UTC m=+1191.796325577" Oct 02 02:06:43 crc kubenswrapper[4885]: I1002 02:06:43.009545 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.636346052 podStartE2EDuration="3.009515317s" podCreationTimestamp="2025-10-02 02:06:40 +0000 UTC" firstStartedPulling="2025-10-02 02:06:41.823042226 +0000 UTC m=+1190.634789625" lastFinishedPulling="2025-10-02 02:06:42.196211491 +0000 UTC m=+1191.007958890" observedRunningTime="2025-10-02 02:06:42.998892863 +0000 UTC m=+1191.810640272" watchObservedRunningTime="2025-10-02 02:06:43.009515317 +0000 UTC m=+1191.821262765" Oct 02 02:06:43 crc kubenswrapper[4885]: I1002 02:06:43.265838 4885 patch_prober.go:28] interesting pod/machine-config-daemon-rttx8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 02:06:43 crc kubenswrapper[4885]: I1002 02:06:43.266150 4885 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 02:06:43 crc kubenswrapper[4885]: I1002 02:06:43.266200 4885 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" Oct 02 02:06:43 crc kubenswrapper[4885]: I1002 02:06:43.266954 4885 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b3432e8413cbb9773cfd116ef5983ce1a5243df994d7ca0829906833f1dfa903"} pod="openshift-machine-config-operator/machine-config-daemon-rttx8" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 02:06:43 crc kubenswrapper[4885]: I1002 02:06:43.267018 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" containerName="machine-config-daemon" containerID="cri-o://b3432e8413cbb9773cfd116ef5983ce1a5243df994d7ca0829906833f1dfa903" gracePeriod=600 Oct 02 02:06:43 crc kubenswrapper[4885]: I1002 02:06:43.328344 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 02 02:06:43 crc kubenswrapper[4885]: I1002 02:06:43.586530 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 02:06:43 crc kubenswrapper[4885]: I1002 02:06:43.610695 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7617cdd5-023a-427d-8365-52dbfc6cf20a-config-data\") pod \"7617cdd5-023a-427d-8365-52dbfc6cf20a\" (UID: \"7617cdd5-023a-427d-8365-52dbfc6cf20a\") " Oct 02 02:06:43 crc kubenswrapper[4885]: I1002 02:06:43.610746 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pm989\" (UniqueName: \"kubernetes.io/projected/7617cdd5-023a-427d-8365-52dbfc6cf20a-kube-api-access-pm989\") pod \"7617cdd5-023a-427d-8365-52dbfc6cf20a\" (UID: \"7617cdd5-023a-427d-8365-52dbfc6cf20a\") " Oct 02 02:06:43 crc kubenswrapper[4885]: I1002 02:06:43.610826 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7617cdd5-023a-427d-8365-52dbfc6cf20a-sg-core-conf-yaml\") pod \"7617cdd5-023a-427d-8365-52dbfc6cf20a\" (UID: \"7617cdd5-023a-427d-8365-52dbfc6cf20a\") " Oct 02 02:06:43 crc kubenswrapper[4885]: I1002 02:06:43.610851 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7617cdd5-023a-427d-8365-52dbfc6cf20a-scripts\") pod \"7617cdd5-023a-427d-8365-52dbfc6cf20a\" (UID: \"7617cdd5-023a-427d-8365-52dbfc6cf20a\") " Oct 02 02:06:43 crc kubenswrapper[4885]: I1002 02:06:43.610922 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7617cdd5-023a-427d-8365-52dbfc6cf20a-log-httpd\") pod \"7617cdd5-023a-427d-8365-52dbfc6cf20a\" (UID: \"7617cdd5-023a-427d-8365-52dbfc6cf20a\") " Oct 02 02:06:43 crc kubenswrapper[4885]: I1002 02:06:43.610959 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7617cdd5-023a-427d-8365-52dbfc6cf20a-combined-ca-bundle\") pod \"7617cdd5-023a-427d-8365-52dbfc6cf20a\" (UID: \"7617cdd5-023a-427d-8365-52dbfc6cf20a\") " Oct 02 02:06:43 crc kubenswrapper[4885]: I1002 02:06:43.611032 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7617cdd5-023a-427d-8365-52dbfc6cf20a-run-httpd\") pod \"7617cdd5-023a-427d-8365-52dbfc6cf20a\" (UID: \"7617cdd5-023a-427d-8365-52dbfc6cf20a\") " Oct 02 02:06:43 crc kubenswrapper[4885]: I1002 02:06:43.611409 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7617cdd5-023a-427d-8365-52dbfc6cf20a-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "7617cdd5-023a-427d-8365-52dbfc6cf20a" (UID: "7617cdd5-023a-427d-8365-52dbfc6cf20a"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:06:43 crc kubenswrapper[4885]: I1002 02:06:43.611668 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7617cdd5-023a-427d-8365-52dbfc6cf20a-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "7617cdd5-023a-427d-8365-52dbfc6cf20a" (UID: "7617cdd5-023a-427d-8365-52dbfc6cf20a"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:06:43 crc kubenswrapper[4885]: I1002 02:06:43.620888 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7617cdd5-023a-427d-8365-52dbfc6cf20a-kube-api-access-pm989" (OuterVolumeSpecName: "kube-api-access-pm989") pod "7617cdd5-023a-427d-8365-52dbfc6cf20a" (UID: "7617cdd5-023a-427d-8365-52dbfc6cf20a"). InnerVolumeSpecName "kube-api-access-pm989". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:06:43 crc kubenswrapper[4885]: I1002 02:06:43.639547 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7617cdd5-023a-427d-8365-52dbfc6cf20a-scripts" (OuterVolumeSpecName: "scripts") pod "7617cdd5-023a-427d-8365-52dbfc6cf20a" (UID: "7617cdd5-023a-427d-8365-52dbfc6cf20a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:06:43 crc kubenswrapper[4885]: I1002 02:06:43.681100 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7617cdd5-023a-427d-8365-52dbfc6cf20a-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "7617cdd5-023a-427d-8365-52dbfc6cf20a" (UID: "7617cdd5-023a-427d-8365-52dbfc6cf20a"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:06:43 crc kubenswrapper[4885]: I1002 02:06:43.712747 4885 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7617cdd5-023a-427d-8365-52dbfc6cf20a-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 02:06:43 crc kubenswrapper[4885]: I1002 02:06:43.712780 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pm989\" (UniqueName: \"kubernetes.io/projected/7617cdd5-023a-427d-8365-52dbfc6cf20a-kube-api-access-pm989\") on node \"crc\" DevicePath \"\"" Oct 02 02:06:43 crc kubenswrapper[4885]: I1002 02:06:43.712793 4885 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7617cdd5-023a-427d-8365-52dbfc6cf20a-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 02 02:06:43 crc kubenswrapper[4885]: I1002 02:06:43.712813 4885 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7617cdd5-023a-427d-8365-52dbfc6cf20a-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 02:06:43 crc kubenswrapper[4885]: I1002 02:06:43.712822 4885 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7617cdd5-023a-427d-8365-52dbfc6cf20a-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 02:06:43 crc kubenswrapper[4885]: I1002 02:06:43.746523 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7617cdd5-023a-427d-8365-52dbfc6cf20a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7617cdd5-023a-427d-8365-52dbfc6cf20a" (UID: "7617cdd5-023a-427d-8365-52dbfc6cf20a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:06:43 crc kubenswrapper[4885]: I1002 02:06:43.789431 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7617cdd5-023a-427d-8365-52dbfc6cf20a-config-data" (OuterVolumeSpecName: "config-data") pod "7617cdd5-023a-427d-8365-52dbfc6cf20a" (UID: "7617cdd5-023a-427d-8365-52dbfc6cf20a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:06:43 crc kubenswrapper[4885]: I1002 02:06:43.814147 4885 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7617cdd5-023a-427d-8365-52dbfc6cf20a-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 02:06:43 crc kubenswrapper[4885]: I1002 02:06:43.814181 4885 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7617cdd5-023a-427d-8365-52dbfc6cf20a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:06:43 crc kubenswrapper[4885]: I1002 02:06:43.978115 4885 generic.go:334] "Generic (PLEG): container finished" podID="d169c8ba-000a-4218-8e3b-4ae53035b110" containerID="b3432e8413cbb9773cfd116ef5983ce1a5243df994d7ca0829906833f1dfa903" exitCode=0 Oct 02 02:06:43 crc kubenswrapper[4885]: I1002 02:06:43.978191 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" event={"ID":"d169c8ba-000a-4218-8e3b-4ae53035b110","Type":"ContainerDied","Data":"b3432e8413cbb9773cfd116ef5983ce1a5243df994d7ca0829906833f1dfa903"} Oct 02 02:06:43 crc kubenswrapper[4885]: I1002 02:06:43.978232 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" event={"ID":"d169c8ba-000a-4218-8e3b-4ae53035b110","Type":"ContainerStarted","Data":"fd898246d6c734dabec969ae60b5e63b1aba037d46a686b38b0527500e9db0fe"} Oct 02 02:06:43 crc kubenswrapper[4885]: I1002 02:06:43.978247 4885 scope.go:117] "RemoveContainer" containerID="4708a92cba43cfdb8c289fa5ff76e06b7b83a40fb102347f523fcefc325254bd" Oct 02 02:06:43 crc kubenswrapper[4885]: I1002 02:06:43.987173 4885 generic.go:334] "Generic (PLEG): container finished" podID="7617cdd5-023a-427d-8365-52dbfc6cf20a" containerID="5467c6f206f6d2e7c0a79b0aa4ab9d08a4129ba7f16275ccee62fec86ad3bf2e" exitCode=0 Oct 02 02:06:43 crc kubenswrapper[4885]: I1002 02:06:43.987240 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7617cdd5-023a-427d-8365-52dbfc6cf20a","Type":"ContainerDied","Data":"5467c6f206f6d2e7c0a79b0aa4ab9d08a4129ba7f16275ccee62fec86ad3bf2e"} Oct 02 02:06:43 crc kubenswrapper[4885]: I1002 02:06:43.987371 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 02:06:43 crc kubenswrapper[4885]: I1002 02:06:43.987606 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7617cdd5-023a-427d-8365-52dbfc6cf20a","Type":"ContainerDied","Data":"6d9831c52506bd7eb51f0fa792de43fc2b6db5f95cba7927541cea46a4e8ee2b"} Oct 02 02:06:44 crc kubenswrapper[4885]: I1002 02:06:44.021432 4885 scope.go:117] "RemoveContainer" containerID="737b1e2d7170ba0c915e063b22c387f8c1b2dcc4d7cb99b2b87f8998b9369cbb" Oct 02 02:06:44 crc kubenswrapper[4885]: I1002 02:06:44.037112 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 02:06:44 crc kubenswrapper[4885]: I1002 02:06:44.058955 4885 scope.go:117] "RemoveContainer" containerID="4d15decee345cf96133f34e791bdd5f7eaec17200d66f3a76a057701de1be4c6" Oct 02 02:06:44 crc kubenswrapper[4885]: I1002 02:06:44.069668 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 02 02:06:44 crc kubenswrapper[4885]: I1002 02:06:44.073449 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 02 02:06:44 crc kubenswrapper[4885]: E1002 02:06:44.073896 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7617cdd5-023a-427d-8365-52dbfc6cf20a" containerName="proxy-httpd" Oct 02 02:06:44 crc kubenswrapper[4885]: I1002 02:06:44.073912 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="7617cdd5-023a-427d-8365-52dbfc6cf20a" containerName="proxy-httpd" Oct 02 02:06:44 crc kubenswrapper[4885]: E1002 02:06:44.073938 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7617cdd5-023a-427d-8365-52dbfc6cf20a" containerName="ceilometer-central-agent" Oct 02 02:06:44 crc kubenswrapper[4885]: I1002 02:06:44.073945 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="7617cdd5-023a-427d-8365-52dbfc6cf20a" containerName="ceilometer-central-agent" Oct 02 02:06:44 crc kubenswrapper[4885]: E1002 02:06:44.073954 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7617cdd5-023a-427d-8365-52dbfc6cf20a" containerName="ceilometer-notification-agent" Oct 02 02:06:44 crc kubenswrapper[4885]: I1002 02:06:44.073960 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="7617cdd5-023a-427d-8365-52dbfc6cf20a" containerName="ceilometer-notification-agent" Oct 02 02:06:44 crc kubenswrapper[4885]: E1002 02:06:44.073986 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7617cdd5-023a-427d-8365-52dbfc6cf20a" containerName="sg-core" Oct 02 02:06:44 crc kubenswrapper[4885]: I1002 02:06:44.073992 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="7617cdd5-023a-427d-8365-52dbfc6cf20a" containerName="sg-core" Oct 02 02:06:44 crc kubenswrapper[4885]: I1002 02:06:44.074202 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="7617cdd5-023a-427d-8365-52dbfc6cf20a" containerName="ceilometer-notification-agent" Oct 02 02:06:44 crc kubenswrapper[4885]: I1002 02:06:44.074213 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="7617cdd5-023a-427d-8365-52dbfc6cf20a" containerName="ceilometer-central-agent" Oct 02 02:06:44 crc kubenswrapper[4885]: I1002 02:06:44.074224 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="7617cdd5-023a-427d-8365-52dbfc6cf20a" containerName="sg-core" Oct 02 02:06:44 crc kubenswrapper[4885]: I1002 02:06:44.074233 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="7617cdd5-023a-427d-8365-52dbfc6cf20a" containerName="proxy-httpd" Oct 02 02:06:44 crc kubenswrapper[4885]: I1002 02:06:44.075859 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 02:06:44 crc kubenswrapper[4885]: I1002 02:06:44.079529 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 02 02:06:44 crc kubenswrapper[4885]: I1002 02:06:44.079634 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 02 02:06:44 crc kubenswrapper[4885]: I1002 02:06:44.079636 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 02 02:06:44 crc kubenswrapper[4885]: I1002 02:06:44.083096 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 02:06:44 crc kubenswrapper[4885]: I1002 02:06:44.089704 4885 scope.go:117] "RemoveContainer" containerID="5467c6f206f6d2e7c0a79b0aa4ab9d08a4129ba7f16275ccee62fec86ad3bf2e" Oct 02 02:06:44 crc kubenswrapper[4885]: I1002 02:06:44.114324 4885 scope.go:117] "RemoveContainer" containerID="1a275dbd70e21d0358dfdb3b1ee5eb025a0342c7445566bddcc00930eaa277f3" Oct 02 02:06:44 crc kubenswrapper[4885]: I1002 02:06:44.120993 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/251ca869-48a9-4245-b887-02c0e455500b-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"251ca869-48a9-4245-b887-02c0e455500b\") " pod="openstack/ceilometer-0" Oct 02 02:06:44 crc kubenswrapper[4885]: I1002 02:06:44.121229 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/251ca869-48a9-4245-b887-02c0e455500b-run-httpd\") pod \"ceilometer-0\" (UID: \"251ca869-48a9-4245-b887-02c0e455500b\") " pod="openstack/ceilometer-0" Oct 02 02:06:44 crc kubenswrapper[4885]: I1002 02:06:44.121371 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/251ca869-48a9-4245-b887-02c0e455500b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"251ca869-48a9-4245-b887-02c0e455500b\") " pod="openstack/ceilometer-0" Oct 02 02:06:44 crc kubenswrapper[4885]: I1002 02:06:44.121458 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/251ca869-48a9-4245-b887-02c0e455500b-log-httpd\") pod \"ceilometer-0\" (UID: \"251ca869-48a9-4245-b887-02c0e455500b\") " pod="openstack/ceilometer-0" Oct 02 02:06:44 crc kubenswrapper[4885]: I1002 02:06:44.121522 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/251ca869-48a9-4245-b887-02c0e455500b-config-data\") pod \"ceilometer-0\" (UID: \"251ca869-48a9-4245-b887-02c0e455500b\") " pod="openstack/ceilometer-0" Oct 02 02:06:44 crc kubenswrapper[4885]: I1002 02:06:44.121539 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/251ca869-48a9-4245-b887-02c0e455500b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"251ca869-48a9-4245-b887-02c0e455500b\") " pod="openstack/ceilometer-0" Oct 02 02:06:44 crc kubenswrapper[4885]: I1002 02:06:44.121559 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/251ca869-48a9-4245-b887-02c0e455500b-scripts\") pod \"ceilometer-0\" (UID: \"251ca869-48a9-4245-b887-02c0e455500b\") " pod="openstack/ceilometer-0" Oct 02 02:06:44 crc kubenswrapper[4885]: I1002 02:06:44.121590 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gt8pn\" (UniqueName: \"kubernetes.io/projected/251ca869-48a9-4245-b887-02c0e455500b-kube-api-access-gt8pn\") pod \"ceilometer-0\" (UID: \"251ca869-48a9-4245-b887-02c0e455500b\") " pod="openstack/ceilometer-0" Oct 02 02:06:44 crc kubenswrapper[4885]: I1002 02:06:44.132103 4885 scope.go:117] "RemoveContainer" containerID="737b1e2d7170ba0c915e063b22c387f8c1b2dcc4d7cb99b2b87f8998b9369cbb" Oct 02 02:06:44 crc kubenswrapper[4885]: E1002 02:06:44.132582 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"737b1e2d7170ba0c915e063b22c387f8c1b2dcc4d7cb99b2b87f8998b9369cbb\": container with ID starting with 737b1e2d7170ba0c915e063b22c387f8c1b2dcc4d7cb99b2b87f8998b9369cbb not found: ID does not exist" containerID="737b1e2d7170ba0c915e063b22c387f8c1b2dcc4d7cb99b2b87f8998b9369cbb" Oct 02 02:06:44 crc kubenswrapper[4885]: I1002 02:06:44.132620 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"737b1e2d7170ba0c915e063b22c387f8c1b2dcc4d7cb99b2b87f8998b9369cbb"} err="failed to get container status \"737b1e2d7170ba0c915e063b22c387f8c1b2dcc4d7cb99b2b87f8998b9369cbb\": rpc error: code = NotFound desc = could not find container \"737b1e2d7170ba0c915e063b22c387f8c1b2dcc4d7cb99b2b87f8998b9369cbb\": container with ID starting with 737b1e2d7170ba0c915e063b22c387f8c1b2dcc4d7cb99b2b87f8998b9369cbb not found: ID does not exist" Oct 02 02:06:44 crc kubenswrapper[4885]: I1002 02:06:44.132645 4885 scope.go:117] "RemoveContainer" containerID="4d15decee345cf96133f34e791bdd5f7eaec17200d66f3a76a057701de1be4c6" Oct 02 02:06:44 crc kubenswrapper[4885]: E1002 02:06:44.132926 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4d15decee345cf96133f34e791bdd5f7eaec17200d66f3a76a057701de1be4c6\": container with ID starting with 4d15decee345cf96133f34e791bdd5f7eaec17200d66f3a76a057701de1be4c6 not found: ID does not exist" containerID="4d15decee345cf96133f34e791bdd5f7eaec17200d66f3a76a057701de1be4c6" Oct 02 02:06:44 crc kubenswrapper[4885]: I1002 02:06:44.132946 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d15decee345cf96133f34e791bdd5f7eaec17200d66f3a76a057701de1be4c6"} err="failed to get container status \"4d15decee345cf96133f34e791bdd5f7eaec17200d66f3a76a057701de1be4c6\": rpc error: code = NotFound desc = could not find container \"4d15decee345cf96133f34e791bdd5f7eaec17200d66f3a76a057701de1be4c6\": container with ID starting with 4d15decee345cf96133f34e791bdd5f7eaec17200d66f3a76a057701de1be4c6 not found: ID does not exist" Oct 02 02:06:44 crc kubenswrapper[4885]: I1002 02:06:44.132963 4885 scope.go:117] "RemoveContainer" containerID="5467c6f206f6d2e7c0a79b0aa4ab9d08a4129ba7f16275ccee62fec86ad3bf2e" Oct 02 02:06:44 crc kubenswrapper[4885]: E1002 02:06:44.133475 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5467c6f206f6d2e7c0a79b0aa4ab9d08a4129ba7f16275ccee62fec86ad3bf2e\": container with ID starting with 5467c6f206f6d2e7c0a79b0aa4ab9d08a4129ba7f16275ccee62fec86ad3bf2e not found: ID does not exist" containerID="5467c6f206f6d2e7c0a79b0aa4ab9d08a4129ba7f16275ccee62fec86ad3bf2e" Oct 02 02:06:44 crc kubenswrapper[4885]: I1002 02:06:44.133591 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5467c6f206f6d2e7c0a79b0aa4ab9d08a4129ba7f16275ccee62fec86ad3bf2e"} err="failed to get container status \"5467c6f206f6d2e7c0a79b0aa4ab9d08a4129ba7f16275ccee62fec86ad3bf2e\": rpc error: code = NotFound desc = could not find container \"5467c6f206f6d2e7c0a79b0aa4ab9d08a4129ba7f16275ccee62fec86ad3bf2e\": container with ID starting with 5467c6f206f6d2e7c0a79b0aa4ab9d08a4129ba7f16275ccee62fec86ad3bf2e not found: ID does not exist" Oct 02 02:06:44 crc kubenswrapper[4885]: I1002 02:06:44.133638 4885 scope.go:117] "RemoveContainer" containerID="1a275dbd70e21d0358dfdb3b1ee5eb025a0342c7445566bddcc00930eaa277f3" Oct 02 02:06:44 crc kubenswrapper[4885]: E1002 02:06:44.134057 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1a275dbd70e21d0358dfdb3b1ee5eb025a0342c7445566bddcc00930eaa277f3\": container with ID starting with 1a275dbd70e21d0358dfdb3b1ee5eb025a0342c7445566bddcc00930eaa277f3 not found: ID does not exist" containerID="1a275dbd70e21d0358dfdb3b1ee5eb025a0342c7445566bddcc00930eaa277f3" Oct 02 02:06:44 crc kubenswrapper[4885]: I1002 02:06:44.134113 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a275dbd70e21d0358dfdb3b1ee5eb025a0342c7445566bddcc00930eaa277f3"} err="failed to get container status \"1a275dbd70e21d0358dfdb3b1ee5eb025a0342c7445566bddcc00930eaa277f3\": rpc error: code = NotFound desc = could not find container \"1a275dbd70e21d0358dfdb3b1ee5eb025a0342c7445566bddcc00930eaa277f3\": container with ID starting with 1a275dbd70e21d0358dfdb3b1ee5eb025a0342c7445566bddcc00930eaa277f3 not found: ID does not exist" Oct 02 02:06:44 crc kubenswrapper[4885]: I1002 02:06:44.223389 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/251ca869-48a9-4245-b887-02c0e455500b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"251ca869-48a9-4245-b887-02c0e455500b\") " pod="openstack/ceilometer-0" Oct 02 02:06:44 crc kubenswrapper[4885]: I1002 02:06:44.223462 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/251ca869-48a9-4245-b887-02c0e455500b-log-httpd\") pod \"ceilometer-0\" (UID: \"251ca869-48a9-4245-b887-02c0e455500b\") " pod="openstack/ceilometer-0" Oct 02 02:06:44 crc kubenswrapper[4885]: I1002 02:06:44.223502 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/251ca869-48a9-4245-b887-02c0e455500b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"251ca869-48a9-4245-b887-02c0e455500b\") " pod="openstack/ceilometer-0" Oct 02 02:06:44 crc kubenswrapper[4885]: I1002 02:06:44.223520 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/251ca869-48a9-4245-b887-02c0e455500b-config-data\") pod \"ceilometer-0\" (UID: \"251ca869-48a9-4245-b887-02c0e455500b\") " pod="openstack/ceilometer-0" Oct 02 02:06:44 crc kubenswrapper[4885]: I1002 02:06:44.223537 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/251ca869-48a9-4245-b887-02c0e455500b-scripts\") pod \"ceilometer-0\" (UID: \"251ca869-48a9-4245-b887-02c0e455500b\") " pod="openstack/ceilometer-0" Oct 02 02:06:44 crc kubenswrapper[4885]: I1002 02:06:44.223562 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gt8pn\" (UniqueName: \"kubernetes.io/projected/251ca869-48a9-4245-b887-02c0e455500b-kube-api-access-gt8pn\") pod \"ceilometer-0\" (UID: \"251ca869-48a9-4245-b887-02c0e455500b\") " pod="openstack/ceilometer-0" Oct 02 02:06:44 crc kubenswrapper[4885]: I1002 02:06:44.223621 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/251ca869-48a9-4245-b887-02c0e455500b-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"251ca869-48a9-4245-b887-02c0e455500b\") " pod="openstack/ceilometer-0" Oct 02 02:06:44 crc kubenswrapper[4885]: I1002 02:06:44.223661 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/251ca869-48a9-4245-b887-02c0e455500b-run-httpd\") pod \"ceilometer-0\" (UID: \"251ca869-48a9-4245-b887-02c0e455500b\") " pod="openstack/ceilometer-0" Oct 02 02:06:44 crc kubenswrapper[4885]: I1002 02:06:44.224143 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/251ca869-48a9-4245-b887-02c0e455500b-log-httpd\") pod \"ceilometer-0\" (UID: \"251ca869-48a9-4245-b887-02c0e455500b\") " pod="openstack/ceilometer-0" Oct 02 02:06:44 crc kubenswrapper[4885]: I1002 02:06:44.224616 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/251ca869-48a9-4245-b887-02c0e455500b-run-httpd\") pod \"ceilometer-0\" (UID: \"251ca869-48a9-4245-b887-02c0e455500b\") " pod="openstack/ceilometer-0" Oct 02 02:06:44 crc kubenswrapper[4885]: I1002 02:06:44.227929 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/251ca869-48a9-4245-b887-02c0e455500b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"251ca869-48a9-4245-b887-02c0e455500b\") " pod="openstack/ceilometer-0" Oct 02 02:06:44 crc kubenswrapper[4885]: I1002 02:06:44.228472 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/251ca869-48a9-4245-b887-02c0e455500b-scripts\") pod \"ceilometer-0\" (UID: \"251ca869-48a9-4245-b887-02c0e455500b\") " pod="openstack/ceilometer-0" Oct 02 02:06:44 crc kubenswrapper[4885]: I1002 02:06:44.228640 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/251ca869-48a9-4245-b887-02c0e455500b-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"251ca869-48a9-4245-b887-02c0e455500b\") " pod="openstack/ceilometer-0" Oct 02 02:06:44 crc kubenswrapper[4885]: I1002 02:06:44.238917 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/251ca869-48a9-4245-b887-02c0e455500b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"251ca869-48a9-4245-b887-02c0e455500b\") " pod="openstack/ceilometer-0" Oct 02 02:06:44 crc kubenswrapper[4885]: I1002 02:06:44.239805 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/251ca869-48a9-4245-b887-02c0e455500b-config-data\") pod \"ceilometer-0\" (UID: \"251ca869-48a9-4245-b887-02c0e455500b\") " pod="openstack/ceilometer-0" Oct 02 02:06:44 crc kubenswrapper[4885]: I1002 02:06:44.241162 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gt8pn\" (UniqueName: \"kubernetes.io/projected/251ca869-48a9-4245-b887-02c0e455500b-kube-api-access-gt8pn\") pod \"ceilometer-0\" (UID: \"251ca869-48a9-4245-b887-02c0e455500b\") " pod="openstack/ceilometer-0" Oct 02 02:06:44 crc kubenswrapper[4885]: I1002 02:06:44.331890 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 02 02:06:44 crc kubenswrapper[4885]: I1002 02:06:44.331946 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 02 02:06:44 crc kubenswrapper[4885]: I1002 02:06:44.398715 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 02:06:44 crc kubenswrapper[4885]: W1002 02:06:44.942004 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod251ca869_48a9_4245_b887_02c0e455500b.slice/crio-d440b55c2efcdd85509a17ff942a766353829c8b8ec3388e1e060b7524b2a6e4 WatchSource:0}: Error finding container d440b55c2efcdd85509a17ff942a766353829c8b8ec3388e1e060b7524b2a6e4: Status 404 returned error can't find the container with id d440b55c2efcdd85509a17ff942a766353829c8b8ec3388e1e060b7524b2a6e4 Oct 02 02:06:44 crc kubenswrapper[4885]: I1002 02:06:44.947426 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 02:06:44 crc kubenswrapper[4885]: I1002 02:06:44.996374 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"251ca869-48a9-4245-b887-02c0e455500b","Type":"ContainerStarted","Data":"d440b55c2efcdd85509a17ff942a766353829c8b8ec3388e1e060b7524b2a6e4"} Oct 02 02:06:46 crc kubenswrapper[4885]: I1002 02:06:46.012727 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"251ca869-48a9-4245-b887-02c0e455500b","Type":"ContainerStarted","Data":"77bfbb776f5312554a4ad739c0eb121c7e60dd17859a76b62523104e981db211"} Oct 02 02:06:46 crc kubenswrapper[4885]: I1002 02:06:46.059861 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7617cdd5-023a-427d-8365-52dbfc6cf20a" path="/var/lib/kubelet/pods/7617cdd5-023a-427d-8365-52dbfc6cf20a/volumes" Oct 02 02:06:46 crc kubenswrapper[4885]: I1002 02:06:46.244586 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Oct 02 02:06:48 crc kubenswrapper[4885]: I1002 02:06:48.044282 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"251ca869-48a9-4245-b887-02c0e455500b","Type":"ContainerStarted","Data":"c077ca3a91c4f46dd68daf5cec540f7a73c79d3a57c07534f16b9de1f08d03c5"} Oct 02 02:06:48 crc kubenswrapper[4885]: I1002 02:06:48.223872 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 02 02:06:48 crc kubenswrapper[4885]: I1002 02:06:48.223931 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 02 02:06:48 crc kubenswrapper[4885]: I1002 02:06:48.328773 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 02 02:06:48 crc kubenswrapper[4885]: I1002 02:06:48.359038 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 02 02:06:49 crc kubenswrapper[4885]: I1002 02:06:49.057132 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"251ca869-48a9-4245-b887-02c0e455500b","Type":"ContainerStarted","Data":"440de512958880efdcb1452d5f9efab7d8c26048ef30bea63fc94bbb5e59e13e"} Oct 02 02:06:49 crc kubenswrapper[4885]: I1002 02:06:49.090923 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 02 02:06:49 crc kubenswrapper[4885]: I1002 02:06:49.307551 4885 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="77273755-f99f-4c79-a730-6e021de17e85" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.195:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 02 02:06:49 crc kubenswrapper[4885]: I1002 02:06:49.307574 4885 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="77273755-f99f-4c79-a730-6e021de17e85" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.195:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 02 02:06:49 crc kubenswrapper[4885]: I1002 02:06:49.332159 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 02 02:06:49 crc kubenswrapper[4885]: I1002 02:06:49.332204 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 02 02:06:50 crc kubenswrapper[4885]: I1002 02:06:50.347457 4885 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="3555c77f-8b14-48c5-a9e1-42e8856dfe19" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.197:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 02 02:06:50 crc kubenswrapper[4885]: I1002 02:06:50.347498 4885 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="3555c77f-8b14-48c5-a9e1-42e8856dfe19" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.197:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 02 02:06:51 crc kubenswrapper[4885]: I1002 02:06:51.086305 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"251ca869-48a9-4245-b887-02c0e455500b","Type":"ContainerStarted","Data":"623cf9aecca2a3da78eb783755a6f609dd04a3dc96da01bfea5dc45cd733d28b"} Oct 02 02:06:51 crc kubenswrapper[4885]: I1002 02:06:51.087237 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 02 02:06:51 crc kubenswrapper[4885]: I1002 02:06:51.131637 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.4838861780000001 podStartE2EDuration="7.131611195s" podCreationTimestamp="2025-10-02 02:06:44 +0000 UTC" firstStartedPulling="2025-10-02 02:06:44.944987786 +0000 UTC m=+1193.756735185" lastFinishedPulling="2025-10-02 02:06:50.592712763 +0000 UTC m=+1199.404460202" observedRunningTime="2025-10-02 02:06:51.111470874 +0000 UTC m=+1199.923218293" watchObservedRunningTime="2025-10-02 02:06:51.131611195 +0000 UTC m=+1199.943358624" Oct 02 02:06:51 crc kubenswrapper[4885]: I1002 02:06:51.321050 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 02 02:06:58 crc kubenswrapper[4885]: I1002 02:06:58.227645 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 02 02:06:58 crc kubenswrapper[4885]: I1002 02:06:58.229469 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 02 02:06:58 crc kubenswrapper[4885]: I1002 02:06:58.229756 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 02 02:06:58 crc kubenswrapper[4885]: I1002 02:06:58.229822 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 02 02:06:58 crc kubenswrapper[4885]: I1002 02:06:58.232777 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 02 02:06:58 crc kubenswrapper[4885]: I1002 02:06:58.236687 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 02 02:06:58 crc kubenswrapper[4885]: I1002 02:06:58.468429 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-vgljb"] Oct 02 02:06:58 crc kubenswrapper[4885]: I1002 02:06:58.471076 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cf4bdb65-vgljb" Oct 02 02:06:58 crc kubenswrapper[4885]: I1002 02:06:58.478603 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-vgljb"] Oct 02 02:06:58 crc kubenswrapper[4885]: I1002 02:06:58.607187 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/05914369-d37a-4911-b64b-4ae7d03385a1-dns-svc\") pod \"dnsmasq-dns-59cf4bdb65-vgljb\" (UID: \"05914369-d37a-4911-b64b-4ae7d03385a1\") " pod="openstack/dnsmasq-dns-59cf4bdb65-vgljb" Oct 02 02:06:58 crc kubenswrapper[4885]: I1002 02:06:58.607779 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g4bth\" (UniqueName: \"kubernetes.io/projected/05914369-d37a-4911-b64b-4ae7d03385a1-kube-api-access-g4bth\") pod \"dnsmasq-dns-59cf4bdb65-vgljb\" (UID: \"05914369-d37a-4911-b64b-4ae7d03385a1\") " pod="openstack/dnsmasq-dns-59cf4bdb65-vgljb" Oct 02 02:06:58 crc kubenswrapper[4885]: I1002 02:06:58.607858 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/05914369-d37a-4911-b64b-4ae7d03385a1-ovsdbserver-nb\") pod \"dnsmasq-dns-59cf4bdb65-vgljb\" (UID: \"05914369-d37a-4911-b64b-4ae7d03385a1\") " pod="openstack/dnsmasq-dns-59cf4bdb65-vgljb" Oct 02 02:06:58 crc kubenswrapper[4885]: I1002 02:06:58.608062 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/05914369-d37a-4911-b64b-4ae7d03385a1-ovsdbserver-sb\") pod \"dnsmasq-dns-59cf4bdb65-vgljb\" (UID: \"05914369-d37a-4911-b64b-4ae7d03385a1\") " pod="openstack/dnsmasq-dns-59cf4bdb65-vgljb" Oct 02 02:06:58 crc kubenswrapper[4885]: I1002 02:06:58.608104 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/05914369-d37a-4911-b64b-4ae7d03385a1-config\") pod \"dnsmasq-dns-59cf4bdb65-vgljb\" (UID: \"05914369-d37a-4911-b64b-4ae7d03385a1\") " pod="openstack/dnsmasq-dns-59cf4bdb65-vgljb" Oct 02 02:06:58 crc kubenswrapper[4885]: I1002 02:06:58.608145 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/05914369-d37a-4911-b64b-4ae7d03385a1-dns-swift-storage-0\") pod \"dnsmasq-dns-59cf4bdb65-vgljb\" (UID: \"05914369-d37a-4911-b64b-4ae7d03385a1\") " pod="openstack/dnsmasq-dns-59cf4bdb65-vgljb" Oct 02 02:06:58 crc kubenswrapper[4885]: I1002 02:06:58.709537 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/05914369-d37a-4911-b64b-4ae7d03385a1-ovsdbserver-nb\") pod \"dnsmasq-dns-59cf4bdb65-vgljb\" (UID: \"05914369-d37a-4911-b64b-4ae7d03385a1\") " pod="openstack/dnsmasq-dns-59cf4bdb65-vgljb" Oct 02 02:06:58 crc kubenswrapper[4885]: I1002 02:06:58.709601 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/05914369-d37a-4911-b64b-4ae7d03385a1-ovsdbserver-sb\") pod \"dnsmasq-dns-59cf4bdb65-vgljb\" (UID: \"05914369-d37a-4911-b64b-4ae7d03385a1\") " pod="openstack/dnsmasq-dns-59cf4bdb65-vgljb" Oct 02 02:06:58 crc kubenswrapper[4885]: I1002 02:06:58.709643 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/05914369-d37a-4911-b64b-4ae7d03385a1-config\") pod \"dnsmasq-dns-59cf4bdb65-vgljb\" (UID: \"05914369-d37a-4911-b64b-4ae7d03385a1\") " pod="openstack/dnsmasq-dns-59cf4bdb65-vgljb" Oct 02 02:06:58 crc kubenswrapper[4885]: I1002 02:06:58.709682 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/05914369-d37a-4911-b64b-4ae7d03385a1-dns-swift-storage-0\") pod \"dnsmasq-dns-59cf4bdb65-vgljb\" (UID: \"05914369-d37a-4911-b64b-4ae7d03385a1\") " pod="openstack/dnsmasq-dns-59cf4bdb65-vgljb" Oct 02 02:06:58 crc kubenswrapper[4885]: I1002 02:06:58.709714 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/05914369-d37a-4911-b64b-4ae7d03385a1-dns-svc\") pod \"dnsmasq-dns-59cf4bdb65-vgljb\" (UID: \"05914369-d37a-4911-b64b-4ae7d03385a1\") " pod="openstack/dnsmasq-dns-59cf4bdb65-vgljb" Oct 02 02:06:58 crc kubenswrapper[4885]: I1002 02:06:58.709742 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g4bth\" (UniqueName: \"kubernetes.io/projected/05914369-d37a-4911-b64b-4ae7d03385a1-kube-api-access-g4bth\") pod \"dnsmasq-dns-59cf4bdb65-vgljb\" (UID: \"05914369-d37a-4911-b64b-4ae7d03385a1\") " pod="openstack/dnsmasq-dns-59cf4bdb65-vgljb" Oct 02 02:06:58 crc kubenswrapper[4885]: I1002 02:06:58.710498 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/05914369-d37a-4911-b64b-4ae7d03385a1-config\") pod \"dnsmasq-dns-59cf4bdb65-vgljb\" (UID: \"05914369-d37a-4911-b64b-4ae7d03385a1\") " pod="openstack/dnsmasq-dns-59cf4bdb65-vgljb" Oct 02 02:06:58 crc kubenswrapper[4885]: I1002 02:06:58.710498 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/05914369-d37a-4911-b64b-4ae7d03385a1-dns-svc\") pod \"dnsmasq-dns-59cf4bdb65-vgljb\" (UID: \"05914369-d37a-4911-b64b-4ae7d03385a1\") " pod="openstack/dnsmasq-dns-59cf4bdb65-vgljb" Oct 02 02:06:58 crc kubenswrapper[4885]: I1002 02:06:58.710498 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/05914369-d37a-4911-b64b-4ae7d03385a1-ovsdbserver-nb\") pod \"dnsmasq-dns-59cf4bdb65-vgljb\" (UID: \"05914369-d37a-4911-b64b-4ae7d03385a1\") " pod="openstack/dnsmasq-dns-59cf4bdb65-vgljb" Oct 02 02:06:58 crc kubenswrapper[4885]: I1002 02:06:58.710585 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/05914369-d37a-4911-b64b-4ae7d03385a1-ovsdbserver-sb\") pod \"dnsmasq-dns-59cf4bdb65-vgljb\" (UID: \"05914369-d37a-4911-b64b-4ae7d03385a1\") " pod="openstack/dnsmasq-dns-59cf4bdb65-vgljb" Oct 02 02:06:58 crc kubenswrapper[4885]: I1002 02:06:58.711158 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/05914369-d37a-4911-b64b-4ae7d03385a1-dns-swift-storage-0\") pod \"dnsmasq-dns-59cf4bdb65-vgljb\" (UID: \"05914369-d37a-4911-b64b-4ae7d03385a1\") " pod="openstack/dnsmasq-dns-59cf4bdb65-vgljb" Oct 02 02:06:58 crc kubenswrapper[4885]: I1002 02:06:58.730131 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g4bth\" (UniqueName: \"kubernetes.io/projected/05914369-d37a-4911-b64b-4ae7d03385a1-kube-api-access-g4bth\") pod \"dnsmasq-dns-59cf4bdb65-vgljb\" (UID: \"05914369-d37a-4911-b64b-4ae7d03385a1\") " pod="openstack/dnsmasq-dns-59cf4bdb65-vgljb" Oct 02 02:06:58 crc kubenswrapper[4885]: I1002 02:06:58.833376 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cf4bdb65-vgljb" Oct 02 02:06:59 crc kubenswrapper[4885]: I1002 02:06:59.312648 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-vgljb"] Oct 02 02:06:59 crc kubenswrapper[4885]: I1002 02:06:59.339348 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 02 02:06:59 crc kubenswrapper[4885]: I1002 02:06:59.342378 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 02 02:06:59 crc kubenswrapper[4885]: I1002 02:06:59.348739 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 02 02:07:00 crc kubenswrapper[4885]: I1002 02:07:00.210956 4885 generic.go:334] "Generic (PLEG): container finished" podID="05914369-d37a-4911-b64b-4ae7d03385a1" containerID="59d44cb4910beddea301c3811e444fac74f711d7e3aabef257b6071f7a9dc01c" exitCode=0 Oct 02 02:07:00 crc kubenswrapper[4885]: I1002 02:07:00.211114 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-vgljb" event={"ID":"05914369-d37a-4911-b64b-4ae7d03385a1","Type":"ContainerDied","Data":"59d44cb4910beddea301c3811e444fac74f711d7e3aabef257b6071f7a9dc01c"} Oct 02 02:07:00 crc kubenswrapper[4885]: I1002 02:07:00.211321 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-vgljb" event={"ID":"05914369-d37a-4911-b64b-4ae7d03385a1","Type":"ContainerStarted","Data":"51aca9c27883ab045da41f8f4a0298c170ef143c2d4cf7345be995ebb50bc380"} Oct 02 02:07:00 crc kubenswrapper[4885]: I1002 02:07:00.220345 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 02 02:07:00 crc kubenswrapper[4885]: I1002 02:07:00.427896 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 02:07:00 crc kubenswrapper[4885]: I1002 02:07:00.428193 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="251ca869-48a9-4245-b887-02c0e455500b" containerName="ceilometer-central-agent" containerID="cri-o://77bfbb776f5312554a4ad739c0eb121c7e60dd17859a76b62523104e981db211" gracePeriod=30 Oct 02 02:07:00 crc kubenswrapper[4885]: I1002 02:07:00.428280 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="251ca869-48a9-4245-b887-02c0e455500b" containerName="proxy-httpd" containerID="cri-o://623cf9aecca2a3da78eb783755a6f609dd04a3dc96da01bfea5dc45cd733d28b" gracePeriod=30 Oct 02 02:07:00 crc kubenswrapper[4885]: I1002 02:07:00.428341 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="251ca869-48a9-4245-b887-02c0e455500b" containerName="sg-core" containerID="cri-o://440de512958880efdcb1452d5f9efab7d8c26048ef30bea63fc94bbb5e59e13e" gracePeriod=30 Oct 02 02:07:00 crc kubenswrapper[4885]: I1002 02:07:00.428387 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="251ca869-48a9-4245-b887-02c0e455500b" containerName="ceilometer-notification-agent" containerID="cri-o://c077ca3a91c4f46dd68daf5cec540f7a73c79d3a57c07534f16b9de1f08d03c5" gracePeriod=30 Oct 02 02:07:00 crc kubenswrapper[4885]: I1002 02:07:00.437680 4885 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="251ca869-48a9-4245-b887-02c0e455500b" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.200:3000/\": EOF" Oct 02 02:07:00 crc kubenswrapper[4885]: I1002 02:07:00.948521 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 02 02:07:01 crc kubenswrapper[4885]: I1002 02:07:01.247843 4885 generic.go:334] "Generic (PLEG): container finished" podID="251ca869-48a9-4245-b887-02c0e455500b" containerID="623cf9aecca2a3da78eb783755a6f609dd04a3dc96da01bfea5dc45cd733d28b" exitCode=0 Oct 02 02:07:01 crc kubenswrapper[4885]: I1002 02:07:01.247885 4885 generic.go:334] "Generic (PLEG): container finished" podID="251ca869-48a9-4245-b887-02c0e455500b" containerID="440de512958880efdcb1452d5f9efab7d8c26048ef30bea63fc94bbb5e59e13e" exitCode=2 Oct 02 02:07:01 crc kubenswrapper[4885]: I1002 02:07:01.247897 4885 generic.go:334] "Generic (PLEG): container finished" podID="251ca869-48a9-4245-b887-02c0e455500b" containerID="77bfbb776f5312554a4ad739c0eb121c7e60dd17859a76b62523104e981db211" exitCode=0 Oct 02 02:07:01 crc kubenswrapper[4885]: I1002 02:07:01.247929 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"251ca869-48a9-4245-b887-02c0e455500b","Type":"ContainerDied","Data":"623cf9aecca2a3da78eb783755a6f609dd04a3dc96da01bfea5dc45cd733d28b"} Oct 02 02:07:01 crc kubenswrapper[4885]: I1002 02:07:01.247990 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"251ca869-48a9-4245-b887-02c0e455500b","Type":"ContainerDied","Data":"440de512958880efdcb1452d5f9efab7d8c26048ef30bea63fc94bbb5e59e13e"} Oct 02 02:07:01 crc kubenswrapper[4885]: I1002 02:07:01.248004 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"251ca869-48a9-4245-b887-02c0e455500b","Type":"ContainerDied","Data":"77bfbb776f5312554a4ad739c0eb121c7e60dd17859a76b62523104e981db211"} Oct 02 02:07:01 crc kubenswrapper[4885]: I1002 02:07:01.251201 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-vgljb" event={"ID":"05914369-d37a-4911-b64b-4ae7d03385a1","Type":"ContainerStarted","Data":"32b84363f04f3add71d5b400ef00ba9c07cd3c14405ec09f9ea64970416f9e78"} Oct 02 02:07:01 crc kubenswrapper[4885]: I1002 02:07:01.251498 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="77273755-f99f-4c79-a730-6e021de17e85" containerName="nova-api-log" containerID="cri-o://21b6364614c91b5470857491a02ccefcda4e2881eccde16bcdcc041a8f8fab10" gracePeriod=30 Oct 02 02:07:01 crc kubenswrapper[4885]: I1002 02:07:01.251588 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="77273755-f99f-4c79-a730-6e021de17e85" containerName="nova-api-api" containerID="cri-o://5a52eb88553b095bb2a574b5316b37218677f9a500d33dacaec88407cd140131" gracePeriod=30 Oct 02 02:07:01 crc kubenswrapper[4885]: I1002 02:07:01.251951 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-59cf4bdb65-vgljb" Oct 02 02:07:01 crc kubenswrapper[4885]: I1002 02:07:01.281838 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-59cf4bdb65-vgljb" podStartSLOduration=3.281815801 podStartE2EDuration="3.281815801s" podCreationTimestamp="2025-10-02 02:06:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:07:01.274278952 +0000 UTC m=+1210.086026361" watchObservedRunningTime="2025-10-02 02:07:01.281815801 +0000 UTC m=+1210.093563210" Oct 02 02:07:02 crc kubenswrapper[4885]: I1002 02:07:02.262783 4885 generic.go:334] "Generic (PLEG): container finished" podID="77273755-f99f-4c79-a730-6e021de17e85" containerID="21b6364614c91b5470857491a02ccefcda4e2881eccde16bcdcc041a8f8fab10" exitCode=143 Oct 02 02:07:02 crc kubenswrapper[4885]: I1002 02:07:02.262850 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"77273755-f99f-4c79-a730-6e021de17e85","Type":"ContainerDied","Data":"21b6364614c91b5470857491a02ccefcda4e2881eccde16bcdcc041a8f8fab10"} Oct 02 02:07:03 crc kubenswrapper[4885]: I1002 02:07:03.869012 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.015353 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/251ca869-48a9-4245-b887-02c0e455500b-run-httpd\") pod \"251ca869-48a9-4245-b887-02c0e455500b\" (UID: \"251ca869-48a9-4245-b887-02c0e455500b\") " Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.015391 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/251ca869-48a9-4245-b887-02c0e455500b-ceilometer-tls-certs\") pod \"251ca869-48a9-4245-b887-02c0e455500b\" (UID: \"251ca869-48a9-4245-b887-02c0e455500b\") " Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.015413 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/251ca869-48a9-4245-b887-02c0e455500b-log-httpd\") pod \"251ca869-48a9-4245-b887-02c0e455500b\" (UID: \"251ca869-48a9-4245-b887-02c0e455500b\") " Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.015503 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/251ca869-48a9-4245-b887-02c0e455500b-sg-core-conf-yaml\") pod \"251ca869-48a9-4245-b887-02c0e455500b\" (UID: \"251ca869-48a9-4245-b887-02c0e455500b\") " Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.015535 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/251ca869-48a9-4245-b887-02c0e455500b-config-data\") pod \"251ca869-48a9-4245-b887-02c0e455500b\" (UID: \"251ca869-48a9-4245-b887-02c0e455500b\") " Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.015577 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/251ca869-48a9-4245-b887-02c0e455500b-combined-ca-bundle\") pod \"251ca869-48a9-4245-b887-02c0e455500b\" (UID: \"251ca869-48a9-4245-b887-02c0e455500b\") " Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.015649 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gt8pn\" (UniqueName: \"kubernetes.io/projected/251ca869-48a9-4245-b887-02c0e455500b-kube-api-access-gt8pn\") pod \"251ca869-48a9-4245-b887-02c0e455500b\" (UID: \"251ca869-48a9-4245-b887-02c0e455500b\") " Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.015704 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/251ca869-48a9-4245-b887-02c0e455500b-scripts\") pod \"251ca869-48a9-4245-b887-02c0e455500b\" (UID: \"251ca869-48a9-4245-b887-02c0e455500b\") " Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.018012 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/251ca869-48a9-4245-b887-02c0e455500b-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "251ca869-48a9-4245-b887-02c0e455500b" (UID: "251ca869-48a9-4245-b887-02c0e455500b"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.018164 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/251ca869-48a9-4245-b887-02c0e455500b-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "251ca869-48a9-4245-b887-02c0e455500b" (UID: "251ca869-48a9-4245-b887-02c0e455500b"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.023250 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/251ca869-48a9-4245-b887-02c0e455500b-kube-api-access-gt8pn" (OuterVolumeSpecName: "kube-api-access-gt8pn") pod "251ca869-48a9-4245-b887-02c0e455500b" (UID: "251ca869-48a9-4245-b887-02c0e455500b"). InnerVolumeSpecName "kube-api-access-gt8pn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.030469 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/251ca869-48a9-4245-b887-02c0e455500b-scripts" (OuterVolumeSpecName: "scripts") pod "251ca869-48a9-4245-b887-02c0e455500b" (UID: "251ca869-48a9-4245-b887-02c0e455500b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.051093 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/251ca869-48a9-4245-b887-02c0e455500b-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "251ca869-48a9-4245-b887-02c0e455500b" (UID: "251ca869-48a9-4245-b887-02c0e455500b"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.080404 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/251ca869-48a9-4245-b887-02c0e455500b-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "251ca869-48a9-4245-b887-02c0e455500b" (UID: "251ca869-48a9-4245-b887-02c0e455500b"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.100639 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/251ca869-48a9-4245-b887-02c0e455500b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "251ca869-48a9-4245-b887-02c0e455500b" (UID: "251ca869-48a9-4245-b887-02c0e455500b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.118033 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gt8pn\" (UniqueName: \"kubernetes.io/projected/251ca869-48a9-4245-b887-02c0e455500b-kube-api-access-gt8pn\") on node \"crc\" DevicePath \"\"" Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.118063 4885 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/251ca869-48a9-4245-b887-02c0e455500b-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.118073 4885 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/251ca869-48a9-4245-b887-02c0e455500b-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.118082 4885 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/251ca869-48a9-4245-b887-02c0e455500b-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.118092 4885 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/251ca869-48a9-4245-b887-02c0e455500b-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.118099 4885 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/251ca869-48a9-4245-b887-02c0e455500b-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.118107 4885 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/251ca869-48a9-4245-b887-02c0e455500b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.126790 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.144142 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/251ca869-48a9-4245-b887-02c0e455500b-config-data" (OuterVolumeSpecName: "config-data") pod "251ca869-48a9-4245-b887-02c0e455500b" (UID: "251ca869-48a9-4245-b887-02c0e455500b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.218951 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f450788-51fa-430e-a52e-770d02dd9ff0-config-data\") pod \"3f450788-51fa-430e-a52e-770d02dd9ff0\" (UID: \"3f450788-51fa-430e-a52e-770d02dd9ff0\") " Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.219175 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f450788-51fa-430e-a52e-770d02dd9ff0-combined-ca-bundle\") pod \"3f450788-51fa-430e-a52e-770d02dd9ff0\" (UID: \"3f450788-51fa-430e-a52e-770d02dd9ff0\") " Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.219422 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mxkg6\" (UniqueName: \"kubernetes.io/projected/3f450788-51fa-430e-a52e-770d02dd9ff0-kube-api-access-mxkg6\") pod \"3f450788-51fa-430e-a52e-770d02dd9ff0\" (UID: \"3f450788-51fa-430e-a52e-770d02dd9ff0\") " Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.220005 4885 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/251ca869-48a9-4245-b887-02c0e455500b-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.228579 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3f450788-51fa-430e-a52e-770d02dd9ff0-kube-api-access-mxkg6" (OuterVolumeSpecName: "kube-api-access-mxkg6") pod "3f450788-51fa-430e-a52e-770d02dd9ff0" (UID: "3f450788-51fa-430e-a52e-770d02dd9ff0"). InnerVolumeSpecName "kube-api-access-mxkg6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.244292 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f450788-51fa-430e-a52e-770d02dd9ff0-config-data" (OuterVolumeSpecName: "config-data") pod "3f450788-51fa-430e-a52e-770d02dd9ff0" (UID: "3f450788-51fa-430e-a52e-770d02dd9ff0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.246632 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f450788-51fa-430e-a52e-770d02dd9ff0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3f450788-51fa-430e-a52e-770d02dd9ff0" (UID: "3f450788-51fa-430e-a52e-770d02dd9ff0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.281121 4885 generic.go:334] "Generic (PLEG): container finished" podID="3f450788-51fa-430e-a52e-770d02dd9ff0" containerID="62a53bb82dd9556b15940442cc66e0553ed172da19d9e4dce1714428edf18e64" exitCode=137 Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.281171 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.281199 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"3f450788-51fa-430e-a52e-770d02dd9ff0","Type":"ContainerDied","Data":"62a53bb82dd9556b15940442cc66e0553ed172da19d9e4dce1714428edf18e64"} Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.281232 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"3f450788-51fa-430e-a52e-770d02dd9ff0","Type":"ContainerDied","Data":"1d5790ea9a1dec217c85fbab4852cd3c354d8c98d42df670f4fb55766c96eb41"} Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.281254 4885 scope.go:117] "RemoveContainer" containerID="62a53bb82dd9556b15940442cc66e0553ed172da19d9e4dce1714428edf18e64" Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.284384 4885 generic.go:334] "Generic (PLEG): container finished" podID="251ca869-48a9-4245-b887-02c0e455500b" containerID="c077ca3a91c4f46dd68daf5cec540f7a73c79d3a57c07534f16b9de1f08d03c5" exitCode=0 Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.284415 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"251ca869-48a9-4245-b887-02c0e455500b","Type":"ContainerDied","Data":"c077ca3a91c4f46dd68daf5cec540f7a73c79d3a57c07534f16b9de1f08d03c5"} Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.284436 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"251ca869-48a9-4245-b887-02c0e455500b","Type":"ContainerDied","Data":"d440b55c2efcdd85509a17ff942a766353829c8b8ec3388e1e060b7524b2a6e4"} Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.284509 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.304869 4885 scope.go:117] "RemoveContainer" containerID="62a53bb82dd9556b15940442cc66e0553ed172da19d9e4dce1714428edf18e64" Oct 02 02:07:04 crc kubenswrapper[4885]: E1002 02:07:04.305567 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"62a53bb82dd9556b15940442cc66e0553ed172da19d9e4dce1714428edf18e64\": container with ID starting with 62a53bb82dd9556b15940442cc66e0553ed172da19d9e4dce1714428edf18e64 not found: ID does not exist" containerID="62a53bb82dd9556b15940442cc66e0553ed172da19d9e4dce1714428edf18e64" Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.305605 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"62a53bb82dd9556b15940442cc66e0553ed172da19d9e4dce1714428edf18e64"} err="failed to get container status \"62a53bb82dd9556b15940442cc66e0553ed172da19d9e4dce1714428edf18e64\": rpc error: code = NotFound desc = could not find container \"62a53bb82dd9556b15940442cc66e0553ed172da19d9e4dce1714428edf18e64\": container with ID starting with 62a53bb82dd9556b15940442cc66e0553ed172da19d9e4dce1714428edf18e64 not found: ID does not exist" Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.305632 4885 scope.go:117] "RemoveContainer" containerID="623cf9aecca2a3da78eb783755a6f609dd04a3dc96da01bfea5dc45cd733d28b" Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.335683 4885 scope.go:117] "RemoveContainer" containerID="440de512958880efdcb1452d5f9efab7d8c26048ef30bea63fc94bbb5e59e13e" Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.340295 4885 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f450788-51fa-430e-a52e-770d02dd9ff0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.340353 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mxkg6\" (UniqueName: \"kubernetes.io/projected/3f450788-51fa-430e-a52e-770d02dd9ff0-kube-api-access-mxkg6\") on node \"crc\" DevicePath \"\"" Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.340370 4885 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f450788-51fa-430e-a52e-770d02dd9ff0-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.374376 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.384383 4885 scope.go:117] "RemoveContainer" containerID="c077ca3a91c4f46dd68daf5cec540f7a73c79d3a57c07534f16b9de1f08d03c5" Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.392501 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.406713 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.414097 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.423244 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 02 02:07:04 crc kubenswrapper[4885]: E1002 02:07:04.423862 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="251ca869-48a9-4245-b887-02c0e455500b" containerName="ceilometer-notification-agent" Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.423878 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="251ca869-48a9-4245-b887-02c0e455500b" containerName="ceilometer-notification-agent" Oct 02 02:07:04 crc kubenswrapper[4885]: E1002 02:07:04.423907 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="251ca869-48a9-4245-b887-02c0e455500b" containerName="ceilometer-central-agent" Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.423913 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="251ca869-48a9-4245-b887-02c0e455500b" containerName="ceilometer-central-agent" Oct 02 02:07:04 crc kubenswrapper[4885]: E1002 02:07:04.423925 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="251ca869-48a9-4245-b887-02c0e455500b" containerName="sg-core" Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.423931 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="251ca869-48a9-4245-b887-02c0e455500b" containerName="sg-core" Oct 02 02:07:04 crc kubenswrapper[4885]: E1002 02:07:04.423942 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f450788-51fa-430e-a52e-770d02dd9ff0" containerName="nova-cell1-novncproxy-novncproxy" Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.423947 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f450788-51fa-430e-a52e-770d02dd9ff0" containerName="nova-cell1-novncproxy-novncproxy" Oct 02 02:07:04 crc kubenswrapper[4885]: E1002 02:07:04.423957 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="251ca869-48a9-4245-b887-02c0e455500b" containerName="proxy-httpd" Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.423962 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="251ca869-48a9-4245-b887-02c0e455500b" containerName="proxy-httpd" Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.424121 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f450788-51fa-430e-a52e-770d02dd9ff0" containerName="nova-cell1-novncproxy-novncproxy" Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.424131 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="251ca869-48a9-4245-b887-02c0e455500b" containerName="sg-core" Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.424142 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="251ca869-48a9-4245-b887-02c0e455500b" containerName="ceilometer-notification-agent" Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.424154 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="251ca869-48a9-4245-b887-02c0e455500b" containerName="proxy-httpd" Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.424165 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="251ca869-48a9-4245-b887-02c0e455500b" containerName="ceilometer-central-agent" Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.426035 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.429037 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.429362 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.429637 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.433092 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.434193 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.436271 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.436629 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.436629 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.440887 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.451547 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.502929 4885 scope.go:117] "RemoveContainer" containerID="77bfbb776f5312554a4ad739c0eb121c7e60dd17859a76b62523104e981db211" Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.525326 4885 scope.go:117] "RemoveContainer" containerID="623cf9aecca2a3da78eb783755a6f609dd04a3dc96da01bfea5dc45cd733d28b" Oct 02 02:07:04 crc kubenswrapper[4885]: E1002 02:07:04.526353 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"623cf9aecca2a3da78eb783755a6f609dd04a3dc96da01bfea5dc45cd733d28b\": container with ID starting with 623cf9aecca2a3da78eb783755a6f609dd04a3dc96da01bfea5dc45cd733d28b not found: ID does not exist" containerID="623cf9aecca2a3da78eb783755a6f609dd04a3dc96da01bfea5dc45cd733d28b" Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.526386 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"623cf9aecca2a3da78eb783755a6f609dd04a3dc96da01bfea5dc45cd733d28b"} err="failed to get container status \"623cf9aecca2a3da78eb783755a6f609dd04a3dc96da01bfea5dc45cd733d28b\": rpc error: code = NotFound desc = could not find container \"623cf9aecca2a3da78eb783755a6f609dd04a3dc96da01bfea5dc45cd733d28b\": container with ID starting with 623cf9aecca2a3da78eb783755a6f609dd04a3dc96da01bfea5dc45cd733d28b not found: ID does not exist" Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.526407 4885 scope.go:117] "RemoveContainer" containerID="440de512958880efdcb1452d5f9efab7d8c26048ef30bea63fc94bbb5e59e13e" Oct 02 02:07:04 crc kubenswrapper[4885]: E1002 02:07:04.526895 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"440de512958880efdcb1452d5f9efab7d8c26048ef30bea63fc94bbb5e59e13e\": container with ID starting with 440de512958880efdcb1452d5f9efab7d8c26048ef30bea63fc94bbb5e59e13e not found: ID does not exist" containerID="440de512958880efdcb1452d5f9efab7d8c26048ef30bea63fc94bbb5e59e13e" Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.526946 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"440de512958880efdcb1452d5f9efab7d8c26048ef30bea63fc94bbb5e59e13e"} err="failed to get container status \"440de512958880efdcb1452d5f9efab7d8c26048ef30bea63fc94bbb5e59e13e\": rpc error: code = NotFound desc = could not find container \"440de512958880efdcb1452d5f9efab7d8c26048ef30bea63fc94bbb5e59e13e\": container with ID starting with 440de512958880efdcb1452d5f9efab7d8c26048ef30bea63fc94bbb5e59e13e not found: ID does not exist" Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.526975 4885 scope.go:117] "RemoveContainer" containerID="c077ca3a91c4f46dd68daf5cec540f7a73c79d3a57c07534f16b9de1f08d03c5" Oct 02 02:07:04 crc kubenswrapper[4885]: E1002 02:07:04.528889 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c077ca3a91c4f46dd68daf5cec540f7a73c79d3a57c07534f16b9de1f08d03c5\": container with ID starting with c077ca3a91c4f46dd68daf5cec540f7a73c79d3a57c07534f16b9de1f08d03c5 not found: ID does not exist" containerID="c077ca3a91c4f46dd68daf5cec540f7a73c79d3a57c07534f16b9de1f08d03c5" Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.528923 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c077ca3a91c4f46dd68daf5cec540f7a73c79d3a57c07534f16b9de1f08d03c5"} err="failed to get container status \"c077ca3a91c4f46dd68daf5cec540f7a73c79d3a57c07534f16b9de1f08d03c5\": rpc error: code = NotFound desc = could not find container \"c077ca3a91c4f46dd68daf5cec540f7a73c79d3a57c07534f16b9de1f08d03c5\": container with ID starting with c077ca3a91c4f46dd68daf5cec540f7a73c79d3a57c07534f16b9de1f08d03c5 not found: ID does not exist" Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.528939 4885 scope.go:117] "RemoveContainer" containerID="77bfbb776f5312554a4ad739c0eb121c7e60dd17859a76b62523104e981db211" Oct 02 02:07:04 crc kubenswrapper[4885]: E1002 02:07:04.529247 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"77bfbb776f5312554a4ad739c0eb121c7e60dd17859a76b62523104e981db211\": container with ID starting with 77bfbb776f5312554a4ad739c0eb121c7e60dd17859a76b62523104e981db211 not found: ID does not exist" containerID="77bfbb776f5312554a4ad739c0eb121c7e60dd17859a76b62523104e981db211" Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.529300 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"77bfbb776f5312554a4ad739c0eb121c7e60dd17859a76b62523104e981db211"} err="failed to get container status \"77bfbb776f5312554a4ad739c0eb121c7e60dd17859a76b62523104e981db211\": rpc error: code = NotFound desc = could not find container \"77bfbb776f5312554a4ad739c0eb121c7e60dd17859a76b62523104e981db211\": container with ID starting with 77bfbb776f5312554a4ad739c0eb121c7e60dd17859a76b62523104e981db211 not found: ID does not exist" Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.546073 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gs9xz\" (UniqueName: \"kubernetes.io/projected/0fab80f5-cc23-4ed6-8c0b-1604d9809341-kube-api-access-gs9xz\") pod \"ceilometer-0\" (UID: \"0fab80f5-cc23-4ed6-8c0b-1604d9809341\") " pod="openstack/ceilometer-0" Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.546117 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0fab80f5-cc23-4ed6-8c0b-1604d9809341-scripts\") pod \"ceilometer-0\" (UID: \"0fab80f5-cc23-4ed6-8c0b-1604d9809341\") " pod="openstack/ceilometer-0" Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.546161 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/f6249fa5-32a1-4353-a521-67d1fde191d4-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"f6249fa5-32a1-4353-a521-67d1fde191d4\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.546179 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0fab80f5-cc23-4ed6-8c0b-1604d9809341-run-httpd\") pod \"ceilometer-0\" (UID: \"0fab80f5-cc23-4ed6-8c0b-1604d9809341\") " pod="openstack/ceilometer-0" Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.546378 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fab80f5-cc23-4ed6-8c0b-1604d9809341-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0fab80f5-cc23-4ed6-8c0b-1604d9809341\") " pod="openstack/ceilometer-0" Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.546476 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8rs2w\" (UniqueName: \"kubernetes.io/projected/f6249fa5-32a1-4353-a521-67d1fde191d4-kube-api-access-8rs2w\") pod \"nova-cell1-novncproxy-0\" (UID: \"f6249fa5-32a1-4353-a521-67d1fde191d4\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.546552 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6249fa5-32a1-4353-a521-67d1fde191d4-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"f6249fa5-32a1-4353-a521-67d1fde191d4\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.546575 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0fab80f5-cc23-4ed6-8c0b-1604d9809341-log-httpd\") pod \"ceilometer-0\" (UID: \"0fab80f5-cc23-4ed6-8c0b-1604d9809341\") " pod="openstack/ceilometer-0" Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.546593 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6249fa5-32a1-4353-a521-67d1fde191d4-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"f6249fa5-32a1-4353-a521-67d1fde191d4\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.546631 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fab80f5-cc23-4ed6-8c0b-1604d9809341-config-data\") pod \"ceilometer-0\" (UID: \"0fab80f5-cc23-4ed6-8c0b-1604d9809341\") " pod="openstack/ceilometer-0" Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.546687 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0fab80f5-cc23-4ed6-8c0b-1604d9809341-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0fab80f5-cc23-4ed6-8c0b-1604d9809341\") " pod="openstack/ceilometer-0" Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.546730 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0fab80f5-cc23-4ed6-8c0b-1604d9809341-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"0fab80f5-cc23-4ed6-8c0b-1604d9809341\") " pod="openstack/ceilometer-0" Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.546766 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/f6249fa5-32a1-4353-a521-67d1fde191d4-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"f6249fa5-32a1-4353-a521-67d1fde191d4\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.648455 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gs9xz\" (UniqueName: \"kubernetes.io/projected/0fab80f5-cc23-4ed6-8c0b-1604d9809341-kube-api-access-gs9xz\") pod \"ceilometer-0\" (UID: \"0fab80f5-cc23-4ed6-8c0b-1604d9809341\") " pod="openstack/ceilometer-0" Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.648511 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0fab80f5-cc23-4ed6-8c0b-1604d9809341-scripts\") pod \"ceilometer-0\" (UID: \"0fab80f5-cc23-4ed6-8c0b-1604d9809341\") " pod="openstack/ceilometer-0" Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.648568 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/f6249fa5-32a1-4353-a521-67d1fde191d4-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"f6249fa5-32a1-4353-a521-67d1fde191d4\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.648590 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0fab80f5-cc23-4ed6-8c0b-1604d9809341-run-httpd\") pod \"ceilometer-0\" (UID: \"0fab80f5-cc23-4ed6-8c0b-1604d9809341\") " pod="openstack/ceilometer-0" Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.648653 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fab80f5-cc23-4ed6-8c0b-1604d9809341-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0fab80f5-cc23-4ed6-8c0b-1604d9809341\") " pod="openstack/ceilometer-0" Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.648702 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8rs2w\" (UniqueName: \"kubernetes.io/projected/f6249fa5-32a1-4353-a521-67d1fde191d4-kube-api-access-8rs2w\") pod \"nova-cell1-novncproxy-0\" (UID: \"f6249fa5-32a1-4353-a521-67d1fde191d4\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.648742 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6249fa5-32a1-4353-a521-67d1fde191d4-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"f6249fa5-32a1-4353-a521-67d1fde191d4\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.648764 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0fab80f5-cc23-4ed6-8c0b-1604d9809341-log-httpd\") pod \"ceilometer-0\" (UID: \"0fab80f5-cc23-4ed6-8c0b-1604d9809341\") " pod="openstack/ceilometer-0" Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.648783 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6249fa5-32a1-4353-a521-67d1fde191d4-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"f6249fa5-32a1-4353-a521-67d1fde191d4\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.648812 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fab80f5-cc23-4ed6-8c0b-1604d9809341-config-data\") pod \"ceilometer-0\" (UID: \"0fab80f5-cc23-4ed6-8c0b-1604d9809341\") " pod="openstack/ceilometer-0" Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.648848 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0fab80f5-cc23-4ed6-8c0b-1604d9809341-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0fab80f5-cc23-4ed6-8c0b-1604d9809341\") " pod="openstack/ceilometer-0" Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.648881 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0fab80f5-cc23-4ed6-8c0b-1604d9809341-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"0fab80f5-cc23-4ed6-8c0b-1604d9809341\") " pod="openstack/ceilometer-0" Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.648910 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/f6249fa5-32a1-4353-a521-67d1fde191d4-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"f6249fa5-32a1-4353-a521-67d1fde191d4\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.650810 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0fab80f5-cc23-4ed6-8c0b-1604d9809341-log-httpd\") pod \"ceilometer-0\" (UID: \"0fab80f5-cc23-4ed6-8c0b-1604d9809341\") " pod="openstack/ceilometer-0" Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.651002 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0fab80f5-cc23-4ed6-8c0b-1604d9809341-run-httpd\") pod \"ceilometer-0\" (UID: \"0fab80f5-cc23-4ed6-8c0b-1604d9809341\") " pod="openstack/ceilometer-0" Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.654794 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/f6249fa5-32a1-4353-a521-67d1fde191d4-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"f6249fa5-32a1-4353-a521-67d1fde191d4\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.655069 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0fab80f5-cc23-4ed6-8c0b-1604d9809341-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0fab80f5-cc23-4ed6-8c0b-1604d9809341\") " pod="openstack/ceilometer-0" Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.656520 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6249fa5-32a1-4353-a521-67d1fde191d4-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"f6249fa5-32a1-4353-a521-67d1fde191d4\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.657469 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6249fa5-32a1-4353-a521-67d1fde191d4-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"f6249fa5-32a1-4353-a521-67d1fde191d4\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.657872 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0fab80f5-cc23-4ed6-8c0b-1604d9809341-scripts\") pod \"ceilometer-0\" (UID: \"0fab80f5-cc23-4ed6-8c0b-1604d9809341\") " pod="openstack/ceilometer-0" Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.660062 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fab80f5-cc23-4ed6-8c0b-1604d9809341-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0fab80f5-cc23-4ed6-8c0b-1604d9809341\") " pod="openstack/ceilometer-0" Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.660903 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fab80f5-cc23-4ed6-8c0b-1604d9809341-config-data\") pod \"ceilometer-0\" (UID: \"0fab80f5-cc23-4ed6-8c0b-1604d9809341\") " pod="openstack/ceilometer-0" Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.666872 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0fab80f5-cc23-4ed6-8c0b-1604d9809341-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"0fab80f5-cc23-4ed6-8c0b-1604d9809341\") " pod="openstack/ceilometer-0" Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.676755 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/f6249fa5-32a1-4353-a521-67d1fde191d4-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"f6249fa5-32a1-4353-a521-67d1fde191d4\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.681322 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gs9xz\" (UniqueName: \"kubernetes.io/projected/0fab80f5-cc23-4ed6-8c0b-1604d9809341-kube-api-access-gs9xz\") pod \"ceilometer-0\" (UID: \"0fab80f5-cc23-4ed6-8c0b-1604d9809341\") " pod="openstack/ceilometer-0" Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.682030 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8rs2w\" (UniqueName: \"kubernetes.io/projected/f6249fa5-32a1-4353-a521-67d1fde191d4-kube-api-access-8rs2w\") pod \"nova-cell1-novncproxy-0\" (UID: \"f6249fa5-32a1-4353-a521-67d1fde191d4\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.822461 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 02:07:04 crc kubenswrapper[4885]: I1002 02:07:04.832278 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 02 02:07:05 crc kubenswrapper[4885]: I1002 02:07:05.298852 4885 generic.go:334] "Generic (PLEG): container finished" podID="77273755-f99f-4c79-a730-6e021de17e85" containerID="5a52eb88553b095bb2a574b5316b37218677f9a500d33dacaec88407cd140131" exitCode=0 Oct 02 02:07:05 crc kubenswrapper[4885]: I1002 02:07:05.298945 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"77273755-f99f-4c79-a730-6e021de17e85","Type":"ContainerDied","Data":"5a52eb88553b095bb2a574b5316b37218677f9a500d33dacaec88407cd140131"} Oct 02 02:07:05 crc kubenswrapper[4885]: I1002 02:07:05.391037 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 02:07:05 crc kubenswrapper[4885]: I1002 02:07:05.453298 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 02:07:05 crc kubenswrapper[4885]: W1002 02:07:05.470339 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0fab80f5_cc23_4ed6_8c0b_1604d9809341.slice/crio-c1555e5d0ae20644bb915da43a82aebeba7140687c9b09a3ec770feffa0c45db WatchSource:0}: Error finding container c1555e5d0ae20644bb915da43a82aebeba7140687c9b09a3ec770feffa0c45db: Status 404 returned error can't find the container with id c1555e5d0ae20644bb915da43a82aebeba7140687c9b09a3ec770feffa0c45db Oct 02 02:07:05 crc kubenswrapper[4885]: I1002 02:07:05.558286 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 02:07:06 crc kubenswrapper[4885]: I1002 02:07:05.666406 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/77273755-f99f-4c79-a730-6e021de17e85-logs\") pod \"77273755-f99f-4c79-a730-6e021de17e85\" (UID: \"77273755-f99f-4c79-a730-6e021de17e85\") " Oct 02 02:07:06 crc kubenswrapper[4885]: I1002 02:07:05.666472 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77273755-f99f-4c79-a730-6e021de17e85-config-data\") pod \"77273755-f99f-4c79-a730-6e021de17e85\" (UID: \"77273755-f99f-4c79-a730-6e021de17e85\") " Oct 02 02:07:06 crc kubenswrapper[4885]: I1002 02:07:05.666569 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h4jp2\" (UniqueName: \"kubernetes.io/projected/77273755-f99f-4c79-a730-6e021de17e85-kube-api-access-h4jp2\") pod \"77273755-f99f-4c79-a730-6e021de17e85\" (UID: \"77273755-f99f-4c79-a730-6e021de17e85\") " Oct 02 02:07:06 crc kubenswrapper[4885]: I1002 02:07:05.666653 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77273755-f99f-4c79-a730-6e021de17e85-combined-ca-bundle\") pod \"77273755-f99f-4c79-a730-6e021de17e85\" (UID: \"77273755-f99f-4c79-a730-6e021de17e85\") " Oct 02 02:07:06 crc kubenswrapper[4885]: I1002 02:07:05.667296 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/77273755-f99f-4c79-a730-6e021de17e85-logs" (OuterVolumeSpecName: "logs") pod "77273755-f99f-4c79-a730-6e021de17e85" (UID: "77273755-f99f-4c79-a730-6e021de17e85"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:07:06 crc kubenswrapper[4885]: I1002 02:07:05.673792 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/77273755-f99f-4c79-a730-6e021de17e85-kube-api-access-h4jp2" (OuterVolumeSpecName: "kube-api-access-h4jp2") pod "77273755-f99f-4c79-a730-6e021de17e85" (UID: "77273755-f99f-4c79-a730-6e021de17e85"). InnerVolumeSpecName "kube-api-access-h4jp2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:07:06 crc kubenswrapper[4885]: I1002 02:07:05.706875 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77273755-f99f-4c79-a730-6e021de17e85-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "77273755-f99f-4c79-a730-6e021de17e85" (UID: "77273755-f99f-4c79-a730-6e021de17e85"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:07:06 crc kubenswrapper[4885]: I1002 02:07:05.710127 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77273755-f99f-4c79-a730-6e021de17e85-config-data" (OuterVolumeSpecName: "config-data") pod "77273755-f99f-4c79-a730-6e021de17e85" (UID: "77273755-f99f-4c79-a730-6e021de17e85"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:07:06 crc kubenswrapper[4885]: I1002 02:07:05.769663 4885 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/77273755-f99f-4c79-a730-6e021de17e85-logs\") on node \"crc\" DevicePath \"\"" Oct 02 02:07:06 crc kubenswrapper[4885]: I1002 02:07:05.769705 4885 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77273755-f99f-4c79-a730-6e021de17e85-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 02:07:06 crc kubenswrapper[4885]: I1002 02:07:05.769718 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h4jp2\" (UniqueName: \"kubernetes.io/projected/77273755-f99f-4c79-a730-6e021de17e85-kube-api-access-h4jp2\") on node \"crc\" DevicePath \"\"" Oct 02 02:07:06 crc kubenswrapper[4885]: I1002 02:07:05.769732 4885 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77273755-f99f-4c79-a730-6e021de17e85-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:07:06 crc kubenswrapper[4885]: I1002 02:07:06.065917 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="251ca869-48a9-4245-b887-02c0e455500b" path="/var/lib/kubelet/pods/251ca869-48a9-4245-b887-02c0e455500b/volumes" Oct 02 02:07:06 crc kubenswrapper[4885]: I1002 02:07:06.066835 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3f450788-51fa-430e-a52e-770d02dd9ff0" path="/var/lib/kubelet/pods/3f450788-51fa-430e-a52e-770d02dd9ff0/volumes" Oct 02 02:07:06 crc kubenswrapper[4885]: I1002 02:07:06.314518 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"f6249fa5-32a1-4353-a521-67d1fde191d4","Type":"ContainerStarted","Data":"4ee75b0d7d182729d04b85b58aab62392d2b744de5076c919fb99554379b9038"} Oct 02 02:07:06 crc kubenswrapper[4885]: I1002 02:07:06.314944 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"f6249fa5-32a1-4353-a521-67d1fde191d4","Type":"ContainerStarted","Data":"be6622cc1cd21fb358c62a15a619561101b111c41063e30a569effc3b39b3779"} Oct 02 02:07:06 crc kubenswrapper[4885]: I1002 02:07:06.317111 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"77273755-f99f-4c79-a730-6e021de17e85","Type":"ContainerDied","Data":"d8e6af61c85a1aadfbea122b344513b580c0ea16fd4bb38865bf1e1f3f0e0333"} Oct 02 02:07:06 crc kubenswrapper[4885]: I1002 02:07:06.317444 4885 scope.go:117] "RemoveContainer" containerID="5a52eb88553b095bb2a574b5316b37218677f9a500d33dacaec88407cd140131" Oct 02 02:07:06 crc kubenswrapper[4885]: I1002 02:07:06.318108 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 02:07:06 crc kubenswrapper[4885]: I1002 02:07:06.319454 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0fab80f5-cc23-4ed6-8c0b-1604d9809341","Type":"ContainerStarted","Data":"62c73e75ae3bffcc4a8746403f7fc2dcc3e7e542d790d718fc1741dadc8d1bcd"} Oct 02 02:07:06 crc kubenswrapper[4885]: I1002 02:07:06.320487 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0fab80f5-cc23-4ed6-8c0b-1604d9809341","Type":"ContainerStarted","Data":"c1555e5d0ae20644bb915da43a82aebeba7140687c9b09a3ec770feffa0c45db"} Oct 02 02:07:06 crc kubenswrapper[4885]: I1002 02:07:06.340485 4885 scope.go:117] "RemoveContainer" containerID="21b6364614c91b5470857491a02ccefcda4e2881eccde16bcdcc041a8f8fab10" Oct 02 02:07:06 crc kubenswrapper[4885]: I1002 02:07:06.347382 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.347360587 podStartE2EDuration="2.347360587s" podCreationTimestamp="2025-10-02 02:07:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:07:06.340815469 +0000 UTC m=+1215.152562868" watchObservedRunningTime="2025-10-02 02:07:06.347360587 +0000 UTC m=+1215.159108006" Oct 02 02:07:06 crc kubenswrapper[4885]: I1002 02:07:06.369933 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 02 02:07:06 crc kubenswrapper[4885]: I1002 02:07:06.407195 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 02 02:07:06 crc kubenswrapper[4885]: I1002 02:07:06.419333 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 02 02:07:06 crc kubenswrapper[4885]: E1002 02:07:06.419738 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77273755-f99f-4c79-a730-6e021de17e85" containerName="nova-api-log" Oct 02 02:07:06 crc kubenswrapper[4885]: I1002 02:07:06.419755 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="77273755-f99f-4c79-a730-6e021de17e85" containerName="nova-api-log" Oct 02 02:07:06 crc kubenswrapper[4885]: E1002 02:07:06.419783 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77273755-f99f-4c79-a730-6e021de17e85" containerName="nova-api-api" Oct 02 02:07:06 crc kubenswrapper[4885]: I1002 02:07:06.419789 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="77273755-f99f-4c79-a730-6e021de17e85" containerName="nova-api-api" Oct 02 02:07:06 crc kubenswrapper[4885]: I1002 02:07:06.419969 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="77273755-f99f-4c79-a730-6e021de17e85" containerName="nova-api-api" Oct 02 02:07:06 crc kubenswrapper[4885]: I1002 02:07:06.419998 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="77273755-f99f-4c79-a730-6e021de17e85" containerName="nova-api-log" Oct 02 02:07:06 crc kubenswrapper[4885]: I1002 02:07:06.420919 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 02:07:06 crc kubenswrapper[4885]: I1002 02:07:06.424776 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 02 02:07:06 crc kubenswrapper[4885]: I1002 02:07:06.424952 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 02 02:07:06 crc kubenswrapper[4885]: I1002 02:07:06.425090 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 02 02:07:06 crc kubenswrapper[4885]: I1002 02:07:06.430088 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 02 02:07:06 crc kubenswrapper[4885]: I1002 02:07:06.591681 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4b8ef3c7-8dd5-4929-82b4-f0623ba96273-public-tls-certs\") pod \"nova-api-0\" (UID: \"4b8ef3c7-8dd5-4929-82b4-f0623ba96273\") " pod="openstack/nova-api-0" Oct 02 02:07:06 crc kubenswrapper[4885]: I1002 02:07:06.591755 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b8ef3c7-8dd5-4929-82b4-f0623ba96273-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"4b8ef3c7-8dd5-4929-82b4-f0623ba96273\") " pod="openstack/nova-api-0" Oct 02 02:07:06 crc kubenswrapper[4885]: I1002 02:07:06.591865 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ms77x\" (UniqueName: \"kubernetes.io/projected/4b8ef3c7-8dd5-4929-82b4-f0623ba96273-kube-api-access-ms77x\") pod \"nova-api-0\" (UID: \"4b8ef3c7-8dd5-4929-82b4-f0623ba96273\") " pod="openstack/nova-api-0" Oct 02 02:07:06 crc kubenswrapper[4885]: I1002 02:07:06.592121 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4b8ef3c7-8dd5-4929-82b4-f0623ba96273-logs\") pod \"nova-api-0\" (UID: \"4b8ef3c7-8dd5-4929-82b4-f0623ba96273\") " pod="openstack/nova-api-0" Oct 02 02:07:06 crc kubenswrapper[4885]: I1002 02:07:06.592233 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b8ef3c7-8dd5-4929-82b4-f0623ba96273-config-data\") pod \"nova-api-0\" (UID: \"4b8ef3c7-8dd5-4929-82b4-f0623ba96273\") " pod="openstack/nova-api-0" Oct 02 02:07:06 crc kubenswrapper[4885]: I1002 02:07:06.592410 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4b8ef3c7-8dd5-4929-82b4-f0623ba96273-internal-tls-certs\") pod \"nova-api-0\" (UID: \"4b8ef3c7-8dd5-4929-82b4-f0623ba96273\") " pod="openstack/nova-api-0" Oct 02 02:07:06 crc kubenswrapper[4885]: I1002 02:07:06.700973 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4b8ef3c7-8dd5-4929-82b4-f0623ba96273-logs\") pod \"nova-api-0\" (UID: \"4b8ef3c7-8dd5-4929-82b4-f0623ba96273\") " pod="openstack/nova-api-0" Oct 02 02:07:06 crc kubenswrapper[4885]: I1002 02:07:06.701093 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b8ef3c7-8dd5-4929-82b4-f0623ba96273-config-data\") pod \"nova-api-0\" (UID: \"4b8ef3c7-8dd5-4929-82b4-f0623ba96273\") " pod="openstack/nova-api-0" Oct 02 02:07:06 crc kubenswrapper[4885]: I1002 02:07:06.701686 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4b8ef3c7-8dd5-4929-82b4-f0623ba96273-logs\") pod \"nova-api-0\" (UID: \"4b8ef3c7-8dd5-4929-82b4-f0623ba96273\") " pod="openstack/nova-api-0" Oct 02 02:07:06 crc kubenswrapper[4885]: I1002 02:07:06.703909 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4b8ef3c7-8dd5-4929-82b4-f0623ba96273-internal-tls-certs\") pod \"nova-api-0\" (UID: \"4b8ef3c7-8dd5-4929-82b4-f0623ba96273\") " pod="openstack/nova-api-0" Oct 02 02:07:06 crc kubenswrapper[4885]: I1002 02:07:06.704941 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4b8ef3c7-8dd5-4929-82b4-f0623ba96273-public-tls-certs\") pod \"nova-api-0\" (UID: \"4b8ef3c7-8dd5-4929-82b4-f0623ba96273\") " pod="openstack/nova-api-0" Oct 02 02:07:06 crc kubenswrapper[4885]: I1002 02:07:06.705029 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b8ef3c7-8dd5-4929-82b4-f0623ba96273-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"4b8ef3c7-8dd5-4929-82b4-f0623ba96273\") " pod="openstack/nova-api-0" Oct 02 02:07:06 crc kubenswrapper[4885]: I1002 02:07:06.705079 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ms77x\" (UniqueName: \"kubernetes.io/projected/4b8ef3c7-8dd5-4929-82b4-f0623ba96273-kube-api-access-ms77x\") pod \"nova-api-0\" (UID: \"4b8ef3c7-8dd5-4929-82b4-f0623ba96273\") " pod="openstack/nova-api-0" Oct 02 02:07:06 crc kubenswrapper[4885]: I1002 02:07:06.711194 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b8ef3c7-8dd5-4929-82b4-f0623ba96273-config-data\") pod \"nova-api-0\" (UID: \"4b8ef3c7-8dd5-4929-82b4-f0623ba96273\") " pod="openstack/nova-api-0" Oct 02 02:07:06 crc kubenswrapper[4885]: I1002 02:07:06.711357 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b8ef3c7-8dd5-4929-82b4-f0623ba96273-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"4b8ef3c7-8dd5-4929-82b4-f0623ba96273\") " pod="openstack/nova-api-0" Oct 02 02:07:06 crc kubenswrapper[4885]: I1002 02:07:06.711935 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4b8ef3c7-8dd5-4929-82b4-f0623ba96273-public-tls-certs\") pod \"nova-api-0\" (UID: \"4b8ef3c7-8dd5-4929-82b4-f0623ba96273\") " pod="openstack/nova-api-0" Oct 02 02:07:06 crc kubenswrapper[4885]: I1002 02:07:06.712226 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4b8ef3c7-8dd5-4929-82b4-f0623ba96273-internal-tls-certs\") pod \"nova-api-0\" (UID: \"4b8ef3c7-8dd5-4929-82b4-f0623ba96273\") " pod="openstack/nova-api-0" Oct 02 02:07:06 crc kubenswrapper[4885]: I1002 02:07:06.736835 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ms77x\" (UniqueName: \"kubernetes.io/projected/4b8ef3c7-8dd5-4929-82b4-f0623ba96273-kube-api-access-ms77x\") pod \"nova-api-0\" (UID: \"4b8ef3c7-8dd5-4929-82b4-f0623ba96273\") " pod="openstack/nova-api-0" Oct 02 02:07:06 crc kubenswrapper[4885]: I1002 02:07:06.750203 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 02:07:07 crc kubenswrapper[4885]: I1002 02:07:07.304151 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 02 02:07:07 crc kubenswrapper[4885]: I1002 02:07:07.348317 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4b8ef3c7-8dd5-4929-82b4-f0623ba96273","Type":"ContainerStarted","Data":"2194d75fe5f5ff3826fc38b1dc38606fe378b96d13e4ef1441230bd6f51ea060"} Oct 02 02:07:08 crc kubenswrapper[4885]: I1002 02:07:08.080366 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="77273755-f99f-4c79-a730-6e021de17e85" path="/var/lib/kubelet/pods/77273755-f99f-4c79-a730-6e021de17e85/volumes" Oct 02 02:07:08 crc kubenswrapper[4885]: I1002 02:07:08.363301 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4b8ef3c7-8dd5-4929-82b4-f0623ba96273","Type":"ContainerStarted","Data":"7bdb98fe4014f2819fdf39a14ebc95521fe001f0c806a90ca817e21997584d5e"} Oct 02 02:07:08 crc kubenswrapper[4885]: I1002 02:07:08.363346 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4b8ef3c7-8dd5-4929-82b4-f0623ba96273","Type":"ContainerStarted","Data":"d85e1506d9190c03bc565532abfe29da275335b9157af3a6ab6900591a8c752d"} Oct 02 02:07:08 crc kubenswrapper[4885]: I1002 02:07:08.367054 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0fab80f5-cc23-4ed6-8c0b-1604d9809341","Type":"ContainerStarted","Data":"7e1133bf827fdbfa3be397bad83aef2da8b1b1f48da8ee6b6782e1d7e1f3aa4d"} Oct 02 02:07:08 crc kubenswrapper[4885]: I1002 02:07:08.367090 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0fab80f5-cc23-4ed6-8c0b-1604d9809341","Type":"ContainerStarted","Data":"66ad6fd9e881790d3eeb29d1e69680a065c872bec42eff6402ccb60af5689389"} Oct 02 02:07:08 crc kubenswrapper[4885]: I1002 02:07:08.835518 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-59cf4bdb65-vgljb" Oct 02 02:07:08 crc kubenswrapper[4885]: I1002 02:07:08.858529 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.85851387 podStartE2EDuration="2.85851387s" podCreationTimestamp="2025-10-02 02:07:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:07:08.397881626 +0000 UTC m=+1217.209629055" watchObservedRunningTime="2025-10-02 02:07:08.85851387 +0000 UTC m=+1217.670261269" Oct 02 02:07:08 crc kubenswrapper[4885]: I1002 02:07:08.941484 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-xm97j"] Oct 02 02:07:08 crc kubenswrapper[4885]: I1002 02:07:08.941774 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-845d6d6f59-xm97j" podUID="e8d78e70-43db-4890-9351-f23f53116625" containerName="dnsmasq-dns" containerID="cri-o://cd44bce807ff26ccb11e5f121e3962a161f00ef43d5350fb2324ef439f2d8371" gracePeriod=10 Oct 02 02:07:09 crc kubenswrapper[4885]: I1002 02:07:09.385560 4885 generic.go:334] "Generic (PLEG): container finished" podID="e8d78e70-43db-4890-9351-f23f53116625" containerID="cd44bce807ff26ccb11e5f121e3962a161f00ef43d5350fb2324ef439f2d8371" exitCode=0 Oct 02 02:07:09 crc kubenswrapper[4885]: I1002 02:07:09.386314 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-xm97j" event={"ID":"e8d78e70-43db-4890-9351-f23f53116625","Type":"ContainerDied","Data":"cd44bce807ff26ccb11e5f121e3962a161f00ef43d5350fb2324ef439f2d8371"} Oct 02 02:07:09 crc kubenswrapper[4885]: I1002 02:07:09.629459 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-xm97j" Oct 02 02:07:09 crc kubenswrapper[4885]: I1002 02:07:09.684637 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8d78e70-43db-4890-9351-f23f53116625-config\") pod \"e8d78e70-43db-4890-9351-f23f53116625\" (UID: \"e8d78e70-43db-4890-9351-f23f53116625\") " Oct 02 02:07:09 crc kubenswrapper[4885]: I1002 02:07:09.684742 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e8d78e70-43db-4890-9351-f23f53116625-dns-svc\") pod \"e8d78e70-43db-4890-9351-f23f53116625\" (UID: \"e8d78e70-43db-4890-9351-f23f53116625\") " Oct 02 02:07:09 crc kubenswrapper[4885]: I1002 02:07:09.685650 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e8d78e70-43db-4890-9351-f23f53116625-ovsdbserver-nb\") pod \"e8d78e70-43db-4890-9351-f23f53116625\" (UID: \"e8d78e70-43db-4890-9351-f23f53116625\") " Oct 02 02:07:09 crc kubenswrapper[4885]: I1002 02:07:09.685683 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dwd7r\" (UniqueName: \"kubernetes.io/projected/e8d78e70-43db-4890-9351-f23f53116625-kube-api-access-dwd7r\") pod \"e8d78e70-43db-4890-9351-f23f53116625\" (UID: \"e8d78e70-43db-4890-9351-f23f53116625\") " Oct 02 02:07:09 crc kubenswrapper[4885]: I1002 02:07:09.685716 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e8d78e70-43db-4890-9351-f23f53116625-ovsdbserver-sb\") pod \"e8d78e70-43db-4890-9351-f23f53116625\" (UID: \"e8d78e70-43db-4890-9351-f23f53116625\") " Oct 02 02:07:09 crc kubenswrapper[4885]: I1002 02:07:09.685774 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e8d78e70-43db-4890-9351-f23f53116625-dns-swift-storage-0\") pod \"e8d78e70-43db-4890-9351-f23f53116625\" (UID: \"e8d78e70-43db-4890-9351-f23f53116625\") " Oct 02 02:07:09 crc kubenswrapper[4885]: I1002 02:07:09.699588 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8d78e70-43db-4890-9351-f23f53116625-kube-api-access-dwd7r" (OuterVolumeSpecName: "kube-api-access-dwd7r") pod "e8d78e70-43db-4890-9351-f23f53116625" (UID: "e8d78e70-43db-4890-9351-f23f53116625"). InnerVolumeSpecName "kube-api-access-dwd7r". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:07:09 crc kubenswrapper[4885]: I1002 02:07:09.732069 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e8d78e70-43db-4890-9351-f23f53116625-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e8d78e70-43db-4890-9351-f23f53116625" (UID: "e8d78e70-43db-4890-9351-f23f53116625"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:07:09 crc kubenswrapper[4885]: I1002 02:07:09.749978 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e8d78e70-43db-4890-9351-f23f53116625-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "e8d78e70-43db-4890-9351-f23f53116625" (UID: "e8d78e70-43db-4890-9351-f23f53116625"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:07:09 crc kubenswrapper[4885]: I1002 02:07:09.764802 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e8d78e70-43db-4890-9351-f23f53116625-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "e8d78e70-43db-4890-9351-f23f53116625" (UID: "e8d78e70-43db-4890-9351-f23f53116625"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:07:09 crc kubenswrapper[4885]: I1002 02:07:09.776277 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e8d78e70-43db-4890-9351-f23f53116625-config" (OuterVolumeSpecName: "config") pod "e8d78e70-43db-4890-9351-f23f53116625" (UID: "e8d78e70-43db-4890-9351-f23f53116625"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:07:09 crc kubenswrapper[4885]: I1002 02:07:09.783767 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e8d78e70-43db-4890-9351-f23f53116625-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "e8d78e70-43db-4890-9351-f23f53116625" (UID: "e8d78e70-43db-4890-9351-f23f53116625"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:07:09 crc kubenswrapper[4885]: I1002 02:07:09.790821 4885 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e8d78e70-43db-4890-9351-f23f53116625-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 02:07:09 crc kubenswrapper[4885]: I1002 02:07:09.791008 4885 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e8d78e70-43db-4890-9351-f23f53116625-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 02 02:07:09 crc kubenswrapper[4885]: I1002 02:07:09.791079 4885 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8d78e70-43db-4890-9351-f23f53116625-config\") on node \"crc\" DevicePath \"\"" Oct 02 02:07:09 crc kubenswrapper[4885]: I1002 02:07:09.791143 4885 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e8d78e70-43db-4890-9351-f23f53116625-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 02:07:09 crc kubenswrapper[4885]: I1002 02:07:09.791230 4885 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e8d78e70-43db-4890-9351-f23f53116625-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 02:07:09 crc kubenswrapper[4885]: I1002 02:07:09.791322 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dwd7r\" (UniqueName: \"kubernetes.io/projected/e8d78e70-43db-4890-9351-f23f53116625-kube-api-access-dwd7r\") on node \"crc\" DevicePath \"\"" Oct 02 02:07:09 crc kubenswrapper[4885]: I1002 02:07:09.833074 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 02 02:07:10 crc kubenswrapper[4885]: I1002 02:07:10.399625 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-xm97j" event={"ID":"e8d78e70-43db-4890-9351-f23f53116625","Type":"ContainerDied","Data":"6f9a1c017ff51b7b46bfb07a9aa3e4fab503a93b6488053c8a64f6f0e0d9468c"} Oct 02 02:07:10 crc kubenswrapper[4885]: I1002 02:07:10.399671 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-xm97j" Oct 02 02:07:10 crc kubenswrapper[4885]: I1002 02:07:10.399906 4885 scope.go:117] "RemoveContainer" containerID="cd44bce807ff26ccb11e5f121e3962a161f00ef43d5350fb2324ef439f2d8371" Oct 02 02:07:10 crc kubenswrapper[4885]: I1002 02:07:10.428544 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0fab80f5-cc23-4ed6-8c0b-1604d9809341","Type":"ContainerStarted","Data":"cccf75c4f90bb8a4514e2667ba4ddd1f0e97935dc08d90bbb4b4c5fc7591e251"} Oct 02 02:07:10 crc kubenswrapper[4885]: I1002 02:07:10.428750 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 02 02:07:10 crc kubenswrapper[4885]: I1002 02:07:10.443248 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-xm97j"] Oct 02 02:07:10 crc kubenswrapper[4885]: I1002 02:07:10.455934 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-xm97j"] Oct 02 02:07:10 crc kubenswrapper[4885]: I1002 02:07:10.457647 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.097663607 podStartE2EDuration="6.457626764s" podCreationTimestamp="2025-10-02 02:07:04 +0000 UTC" firstStartedPulling="2025-10-02 02:07:05.472202932 +0000 UTC m=+1214.283950331" lastFinishedPulling="2025-10-02 02:07:09.832166099 +0000 UTC m=+1218.643913488" observedRunningTime="2025-10-02 02:07:10.44893133 +0000 UTC m=+1219.260678729" watchObservedRunningTime="2025-10-02 02:07:10.457626764 +0000 UTC m=+1219.269374173" Oct 02 02:07:10 crc kubenswrapper[4885]: I1002 02:07:10.458324 4885 scope.go:117] "RemoveContainer" containerID="d68a7424c831e1a782715dc1bd7924fc00e70ad464803683bd68ab2294f92457" Oct 02 02:07:12 crc kubenswrapper[4885]: I1002 02:07:12.068509 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e8d78e70-43db-4890-9351-f23f53116625" path="/var/lib/kubelet/pods/e8d78e70-43db-4890-9351-f23f53116625/volumes" Oct 02 02:07:14 crc kubenswrapper[4885]: I1002 02:07:14.833471 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Oct 02 02:07:14 crc kubenswrapper[4885]: I1002 02:07:14.868775 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Oct 02 02:07:15 crc kubenswrapper[4885]: I1002 02:07:15.533866 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Oct 02 02:07:15 crc kubenswrapper[4885]: I1002 02:07:15.746344 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-fsn9x"] Oct 02 02:07:15 crc kubenswrapper[4885]: E1002 02:07:15.747029 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8d78e70-43db-4890-9351-f23f53116625" containerName="dnsmasq-dns" Oct 02 02:07:15 crc kubenswrapper[4885]: I1002 02:07:15.747047 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8d78e70-43db-4890-9351-f23f53116625" containerName="dnsmasq-dns" Oct 02 02:07:15 crc kubenswrapper[4885]: E1002 02:07:15.747057 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8d78e70-43db-4890-9351-f23f53116625" containerName="init" Oct 02 02:07:15 crc kubenswrapper[4885]: I1002 02:07:15.747065 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8d78e70-43db-4890-9351-f23f53116625" containerName="init" Oct 02 02:07:15 crc kubenswrapper[4885]: I1002 02:07:15.747364 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8d78e70-43db-4890-9351-f23f53116625" containerName="dnsmasq-dns" Oct 02 02:07:15 crc kubenswrapper[4885]: I1002 02:07:15.748071 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-fsn9x" Oct 02 02:07:15 crc kubenswrapper[4885]: I1002 02:07:15.750572 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Oct 02 02:07:15 crc kubenswrapper[4885]: I1002 02:07:15.751317 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Oct 02 02:07:15 crc kubenswrapper[4885]: I1002 02:07:15.769439 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-fsn9x"] Oct 02 02:07:15 crc kubenswrapper[4885]: I1002 02:07:15.861582 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5tlbc\" (UniqueName: \"kubernetes.io/projected/a91d000d-543e-42b4-b815-12243df15ea9-kube-api-access-5tlbc\") pod \"nova-cell1-cell-mapping-fsn9x\" (UID: \"a91d000d-543e-42b4-b815-12243df15ea9\") " pod="openstack/nova-cell1-cell-mapping-fsn9x" Oct 02 02:07:15 crc kubenswrapper[4885]: I1002 02:07:15.861742 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a91d000d-543e-42b4-b815-12243df15ea9-scripts\") pod \"nova-cell1-cell-mapping-fsn9x\" (UID: \"a91d000d-543e-42b4-b815-12243df15ea9\") " pod="openstack/nova-cell1-cell-mapping-fsn9x" Oct 02 02:07:15 crc kubenswrapper[4885]: I1002 02:07:15.861857 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a91d000d-543e-42b4-b815-12243df15ea9-config-data\") pod \"nova-cell1-cell-mapping-fsn9x\" (UID: \"a91d000d-543e-42b4-b815-12243df15ea9\") " pod="openstack/nova-cell1-cell-mapping-fsn9x" Oct 02 02:07:15 crc kubenswrapper[4885]: I1002 02:07:15.862315 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a91d000d-543e-42b4-b815-12243df15ea9-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-fsn9x\" (UID: \"a91d000d-543e-42b4-b815-12243df15ea9\") " pod="openstack/nova-cell1-cell-mapping-fsn9x" Oct 02 02:07:15 crc kubenswrapper[4885]: I1002 02:07:15.965051 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a91d000d-543e-42b4-b815-12243df15ea9-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-fsn9x\" (UID: \"a91d000d-543e-42b4-b815-12243df15ea9\") " pod="openstack/nova-cell1-cell-mapping-fsn9x" Oct 02 02:07:15 crc kubenswrapper[4885]: I1002 02:07:15.965318 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5tlbc\" (UniqueName: \"kubernetes.io/projected/a91d000d-543e-42b4-b815-12243df15ea9-kube-api-access-5tlbc\") pod \"nova-cell1-cell-mapping-fsn9x\" (UID: \"a91d000d-543e-42b4-b815-12243df15ea9\") " pod="openstack/nova-cell1-cell-mapping-fsn9x" Oct 02 02:07:15 crc kubenswrapper[4885]: I1002 02:07:15.965950 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a91d000d-543e-42b4-b815-12243df15ea9-scripts\") pod \"nova-cell1-cell-mapping-fsn9x\" (UID: \"a91d000d-543e-42b4-b815-12243df15ea9\") " pod="openstack/nova-cell1-cell-mapping-fsn9x" Oct 02 02:07:15 crc kubenswrapper[4885]: I1002 02:07:15.966020 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a91d000d-543e-42b4-b815-12243df15ea9-config-data\") pod \"nova-cell1-cell-mapping-fsn9x\" (UID: \"a91d000d-543e-42b4-b815-12243df15ea9\") " pod="openstack/nova-cell1-cell-mapping-fsn9x" Oct 02 02:07:15 crc kubenswrapper[4885]: I1002 02:07:15.977641 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a91d000d-543e-42b4-b815-12243df15ea9-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-fsn9x\" (UID: \"a91d000d-543e-42b4-b815-12243df15ea9\") " pod="openstack/nova-cell1-cell-mapping-fsn9x" Oct 02 02:07:15 crc kubenswrapper[4885]: I1002 02:07:15.977798 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a91d000d-543e-42b4-b815-12243df15ea9-config-data\") pod \"nova-cell1-cell-mapping-fsn9x\" (UID: \"a91d000d-543e-42b4-b815-12243df15ea9\") " pod="openstack/nova-cell1-cell-mapping-fsn9x" Oct 02 02:07:15 crc kubenswrapper[4885]: I1002 02:07:15.978404 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a91d000d-543e-42b4-b815-12243df15ea9-scripts\") pod \"nova-cell1-cell-mapping-fsn9x\" (UID: \"a91d000d-543e-42b4-b815-12243df15ea9\") " pod="openstack/nova-cell1-cell-mapping-fsn9x" Oct 02 02:07:15 crc kubenswrapper[4885]: I1002 02:07:15.998634 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5tlbc\" (UniqueName: \"kubernetes.io/projected/a91d000d-543e-42b4-b815-12243df15ea9-kube-api-access-5tlbc\") pod \"nova-cell1-cell-mapping-fsn9x\" (UID: \"a91d000d-543e-42b4-b815-12243df15ea9\") " pod="openstack/nova-cell1-cell-mapping-fsn9x" Oct 02 02:07:16 crc kubenswrapper[4885]: I1002 02:07:16.078472 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-fsn9x" Oct 02 02:07:16 crc kubenswrapper[4885]: I1002 02:07:16.642795 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-fsn9x"] Oct 02 02:07:16 crc kubenswrapper[4885]: W1002 02:07:16.648357 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda91d000d_543e_42b4_b815_12243df15ea9.slice/crio-1a0c213cd542583e4e0e0ebc3e4fca30d9617d88cc20f0d101c50d55af7073df WatchSource:0}: Error finding container 1a0c213cd542583e4e0e0ebc3e4fca30d9617d88cc20f0d101c50d55af7073df: Status 404 returned error can't find the container with id 1a0c213cd542583e4e0e0ebc3e4fca30d9617d88cc20f0d101c50d55af7073df Oct 02 02:07:16 crc kubenswrapper[4885]: I1002 02:07:16.751232 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 02 02:07:16 crc kubenswrapper[4885]: I1002 02:07:16.751402 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 02 02:07:17 crc kubenswrapper[4885]: I1002 02:07:17.533163 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-fsn9x" event={"ID":"a91d000d-543e-42b4-b815-12243df15ea9","Type":"ContainerStarted","Data":"fa1c8dfbc36fd1d1ad81f83d4fdff3abd1018ec697ea27a6ce4e03a4ed5a1ff7"} Oct 02 02:07:17 crc kubenswrapper[4885]: I1002 02:07:17.533957 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-fsn9x" event={"ID":"a91d000d-543e-42b4-b815-12243df15ea9","Type":"ContainerStarted","Data":"1a0c213cd542583e4e0e0ebc3e4fca30d9617d88cc20f0d101c50d55af7073df"} Oct 02 02:07:17 crc kubenswrapper[4885]: I1002 02:07:17.569399 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-fsn9x" podStartSLOduration=2.569374777 podStartE2EDuration="2.569374777s" podCreationTimestamp="2025-10-02 02:07:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:07:17.553951979 +0000 UTC m=+1226.365699478" watchObservedRunningTime="2025-10-02 02:07:17.569374777 +0000 UTC m=+1226.381122216" Oct 02 02:07:17 crc kubenswrapper[4885]: I1002 02:07:17.771529 4885 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="4b8ef3c7-8dd5-4929-82b4-f0623ba96273" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.204:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 02 02:07:17 crc kubenswrapper[4885]: I1002 02:07:17.771661 4885 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="4b8ef3c7-8dd5-4929-82b4-f0623ba96273" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.204:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 02 02:07:21 crc kubenswrapper[4885]: I1002 02:07:21.578071 4885 generic.go:334] "Generic (PLEG): container finished" podID="a91d000d-543e-42b4-b815-12243df15ea9" containerID="fa1c8dfbc36fd1d1ad81f83d4fdff3abd1018ec697ea27a6ce4e03a4ed5a1ff7" exitCode=0 Oct 02 02:07:21 crc kubenswrapper[4885]: I1002 02:07:21.578149 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-fsn9x" event={"ID":"a91d000d-543e-42b4-b815-12243df15ea9","Type":"ContainerDied","Data":"fa1c8dfbc36fd1d1ad81f83d4fdff3abd1018ec697ea27a6ce4e03a4ed5a1ff7"} Oct 02 02:07:23 crc kubenswrapper[4885]: I1002 02:07:23.014596 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-fsn9x" Oct 02 02:07:23 crc kubenswrapper[4885]: I1002 02:07:23.144496 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5tlbc\" (UniqueName: \"kubernetes.io/projected/a91d000d-543e-42b4-b815-12243df15ea9-kube-api-access-5tlbc\") pod \"a91d000d-543e-42b4-b815-12243df15ea9\" (UID: \"a91d000d-543e-42b4-b815-12243df15ea9\") " Oct 02 02:07:23 crc kubenswrapper[4885]: I1002 02:07:23.144656 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a91d000d-543e-42b4-b815-12243df15ea9-combined-ca-bundle\") pod \"a91d000d-543e-42b4-b815-12243df15ea9\" (UID: \"a91d000d-543e-42b4-b815-12243df15ea9\") " Oct 02 02:07:23 crc kubenswrapper[4885]: I1002 02:07:23.144694 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a91d000d-543e-42b4-b815-12243df15ea9-scripts\") pod \"a91d000d-543e-42b4-b815-12243df15ea9\" (UID: \"a91d000d-543e-42b4-b815-12243df15ea9\") " Oct 02 02:07:23 crc kubenswrapper[4885]: I1002 02:07:23.144911 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a91d000d-543e-42b4-b815-12243df15ea9-config-data\") pod \"a91d000d-543e-42b4-b815-12243df15ea9\" (UID: \"a91d000d-543e-42b4-b815-12243df15ea9\") " Oct 02 02:07:23 crc kubenswrapper[4885]: I1002 02:07:23.150071 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a91d000d-543e-42b4-b815-12243df15ea9-scripts" (OuterVolumeSpecName: "scripts") pod "a91d000d-543e-42b4-b815-12243df15ea9" (UID: "a91d000d-543e-42b4-b815-12243df15ea9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:07:23 crc kubenswrapper[4885]: I1002 02:07:23.150705 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a91d000d-543e-42b4-b815-12243df15ea9-kube-api-access-5tlbc" (OuterVolumeSpecName: "kube-api-access-5tlbc") pod "a91d000d-543e-42b4-b815-12243df15ea9" (UID: "a91d000d-543e-42b4-b815-12243df15ea9"). InnerVolumeSpecName "kube-api-access-5tlbc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:07:23 crc kubenswrapper[4885]: I1002 02:07:23.171245 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a91d000d-543e-42b4-b815-12243df15ea9-config-data" (OuterVolumeSpecName: "config-data") pod "a91d000d-543e-42b4-b815-12243df15ea9" (UID: "a91d000d-543e-42b4-b815-12243df15ea9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:07:23 crc kubenswrapper[4885]: I1002 02:07:23.193081 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a91d000d-543e-42b4-b815-12243df15ea9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a91d000d-543e-42b4-b815-12243df15ea9" (UID: "a91d000d-543e-42b4-b815-12243df15ea9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:07:23 crc kubenswrapper[4885]: I1002 02:07:23.249309 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5tlbc\" (UniqueName: \"kubernetes.io/projected/a91d000d-543e-42b4-b815-12243df15ea9-kube-api-access-5tlbc\") on node \"crc\" DevicePath \"\"" Oct 02 02:07:23 crc kubenswrapper[4885]: I1002 02:07:23.249352 4885 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a91d000d-543e-42b4-b815-12243df15ea9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:07:23 crc kubenswrapper[4885]: I1002 02:07:23.249371 4885 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a91d000d-543e-42b4-b815-12243df15ea9-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 02:07:23 crc kubenswrapper[4885]: I1002 02:07:23.249389 4885 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a91d000d-543e-42b4-b815-12243df15ea9-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 02:07:23 crc kubenswrapper[4885]: I1002 02:07:23.640157 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-fsn9x" event={"ID":"a91d000d-543e-42b4-b815-12243df15ea9","Type":"ContainerDied","Data":"1a0c213cd542583e4e0e0ebc3e4fca30d9617d88cc20f0d101c50d55af7073df"} Oct 02 02:07:23 crc kubenswrapper[4885]: I1002 02:07:23.640236 4885 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1a0c213cd542583e4e0e0ebc3e4fca30d9617d88cc20f0d101c50d55af7073df" Oct 02 02:07:23 crc kubenswrapper[4885]: I1002 02:07:23.640290 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-fsn9x" Oct 02 02:07:23 crc kubenswrapper[4885]: I1002 02:07:23.832397 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 02 02:07:23 crc kubenswrapper[4885]: I1002 02:07:23.832856 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="4b8ef3c7-8dd5-4929-82b4-f0623ba96273" containerName="nova-api-log" containerID="cri-o://d85e1506d9190c03bc565532abfe29da275335b9157af3a6ab6900591a8c752d" gracePeriod=30 Oct 02 02:07:23 crc kubenswrapper[4885]: I1002 02:07:23.832936 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="4b8ef3c7-8dd5-4929-82b4-f0623ba96273" containerName="nova-api-api" containerID="cri-o://7bdb98fe4014f2819fdf39a14ebc95521fe001f0c806a90ca817e21997584d5e" gracePeriod=30 Oct 02 02:07:23 crc kubenswrapper[4885]: I1002 02:07:23.906777 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 02:07:23 crc kubenswrapper[4885]: I1002 02:07:23.907213 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="6895f52a-cc09-4eac-9f57-32f91106afb9" containerName="nova-scheduler-scheduler" containerID="cri-o://77022366cfd5951f2c9ecb6edf329467419834685af3580f17b8fdf85b2921cb" gracePeriod=30 Oct 02 02:07:23 crc kubenswrapper[4885]: I1002 02:07:23.924428 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 02:07:23 crc kubenswrapper[4885]: I1002 02:07:23.924733 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="3555c77f-8b14-48c5-a9e1-42e8856dfe19" containerName="nova-metadata-metadata" containerID="cri-o://ba1f877c8c7a3dc96a1a7cbad1d575537e1a8afaf77c80ebce637df8113780d6" gracePeriod=30 Oct 02 02:07:23 crc kubenswrapper[4885]: I1002 02:07:23.927702 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="3555c77f-8b14-48c5-a9e1-42e8856dfe19" containerName="nova-metadata-log" containerID="cri-o://f52ccc636851ade8d152837715411fd00cf473e5b713d8ff129c166e1d595b01" gracePeriod=30 Oct 02 02:07:24 crc kubenswrapper[4885]: I1002 02:07:24.649982 4885 generic.go:334] "Generic (PLEG): container finished" podID="3555c77f-8b14-48c5-a9e1-42e8856dfe19" containerID="f52ccc636851ade8d152837715411fd00cf473e5b713d8ff129c166e1d595b01" exitCode=143 Oct 02 02:07:24 crc kubenswrapper[4885]: I1002 02:07:24.650094 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3555c77f-8b14-48c5-a9e1-42e8856dfe19","Type":"ContainerDied","Data":"f52ccc636851ade8d152837715411fd00cf473e5b713d8ff129c166e1d595b01"} Oct 02 02:07:24 crc kubenswrapper[4885]: I1002 02:07:24.653851 4885 generic.go:334] "Generic (PLEG): container finished" podID="4b8ef3c7-8dd5-4929-82b4-f0623ba96273" containerID="d85e1506d9190c03bc565532abfe29da275335b9157af3a6ab6900591a8c752d" exitCode=143 Oct 02 02:07:24 crc kubenswrapper[4885]: I1002 02:07:24.653898 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4b8ef3c7-8dd5-4929-82b4-f0623ba96273","Type":"ContainerDied","Data":"d85e1506d9190c03bc565532abfe29da275335b9157af3a6ab6900591a8c752d"} Oct 02 02:07:27 crc kubenswrapper[4885]: I1002 02:07:27.065957 4885 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="3555c77f-8b14-48c5-a9e1-42e8856dfe19" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.197:8775/\": read tcp 10.217.0.2:36352->10.217.0.197:8775: read: connection reset by peer" Oct 02 02:07:27 crc kubenswrapper[4885]: I1002 02:07:27.066684 4885 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="3555c77f-8b14-48c5-a9e1-42e8856dfe19" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.197:8775/\": read tcp 10.217.0.2:36364->10.217.0.197:8775: read: connection reset by peer" Oct 02 02:07:27 crc kubenswrapper[4885]: I1002 02:07:27.543632 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 02:07:27 crc kubenswrapper[4885]: I1002 02:07:27.605475 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 02:07:27 crc kubenswrapper[4885]: I1002 02:07:27.641980 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3555c77f-8b14-48c5-a9e1-42e8856dfe19-logs\") pod \"3555c77f-8b14-48c5-a9e1-42e8856dfe19\" (UID: \"3555c77f-8b14-48c5-a9e1-42e8856dfe19\") " Oct 02 02:07:27 crc kubenswrapper[4885]: I1002 02:07:27.642056 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3555c77f-8b14-48c5-a9e1-42e8856dfe19-combined-ca-bundle\") pod \"3555c77f-8b14-48c5-a9e1-42e8856dfe19\" (UID: \"3555c77f-8b14-48c5-a9e1-42e8856dfe19\") " Oct 02 02:07:27 crc kubenswrapper[4885]: I1002 02:07:27.642119 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3555c77f-8b14-48c5-a9e1-42e8856dfe19-config-data\") pod \"3555c77f-8b14-48c5-a9e1-42e8856dfe19\" (UID: \"3555c77f-8b14-48c5-a9e1-42e8856dfe19\") " Oct 02 02:07:27 crc kubenswrapper[4885]: I1002 02:07:27.642241 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/3555c77f-8b14-48c5-a9e1-42e8856dfe19-nova-metadata-tls-certs\") pod \"3555c77f-8b14-48c5-a9e1-42e8856dfe19\" (UID: \"3555c77f-8b14-48c5-a9e1-42e8856dfe19\") " Oct 02 02:07:27 crc kubenswrapper[4885]: I1002 02:07:27.642785 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3555c77f-8b14-48c5-a9e1-42e8856dfe19-logs" (OuterVolumeSpecName: "logs") pod "3555c77f-8b14-48c5-a9e1-42e8856dfe19" (UID: "3555c77f-8b14-48c5-a9e1-42e8856dfe19"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:07:27 crc kubenswrapper[4885]: I1002 02:07:27.643012 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2h5c6\" (UniqueName: \"kubernetes.io/projected/3555c77f-8b14-48c5-a9e1-42e8856dfe19-kube-api-access-2h5c6\") pod \"3555c77f-8b14-48c5-a9e1-42e8856dfe19\" (UID: \"3555c77f-8b14-48c5-a9e1-42e8856dfe19\") " Oct 02 02:07:27 crc kubenswrapper[4885]: I1002 02:07:27.643491 4885 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3555c77f-8b14-48c5-a9e1-42e8856dfe19-logs\") on node \"crc\" DevicePath \"\"" Oct 02 02:07:27 crc kubenswrapper[4885]: I1002 02:07:27.653988 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3555c77f-8b14-48c5-a9e1-42e8856dfe19-kube-api-access-2h5c6" (OuterVolumeSpecName: "kube-api-access-2h5c6") pod "3555c77f-8b14-48c5-a9e1-42e8856dfe19" (UID: "3555c77f-8b14-48c5-a9e1-42e8856dfe19"). InnerVolumeSpecName "kube-api-access-2h5c6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:07:27 crc kubenswrapper[4885]: I1002 02:07:27.675127 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3555c77f-8b14-48c5-a9e1-42e8856dfe19-config-data" (OuterVolumeSpecName: "config-data") pod "3555c77f-8b14-48c5-a9e1-42e8856dfe19" (UID: "3555c77f-8b14-48c5-a9e1-42e8856dfe19"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:07:27 crc kubenswrapper[4885]: I1002 02:07:27.676017 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 02:07:27 crc kubenswrapper[4885]: I1002 02:07:27.684744 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3555c77f-8b14-48c5-a9e1-42e8856dfe19-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3555c77f-8b14-48c5-a9e1-42e8856dfe19" (UID: "3555c77f-8b14-48c5-a9e1-42e8856dfe19"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:07:27 crc kubenswrapper[4885]: I1002 02:07:27.689235 4885 generic.go:334] "Generic (PLEG): container finished" podID="3555c77f-8b14-48c5-a9e1-42e8856dfe19" containerID="ba1f877c8c7a3dc96a1a7cbad1d575537e1a8afaf77c80ebce637df8113780d6" exitCode=0 Oct 02 02:07:27 crc kubenswrapper[4885]: I1002 02:07:27.689399 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 02:07:27 crc kubenswrapper[4885]: I1002 02:07:27.690226 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3555c77f-8b14-48c5-a9e1-42e8856dfe19","Type":"ContainerDied","Data":"ba1f877c8c7a3dc96a1a7cbad1d575537e1a8afaf77c80ebce637df8113780d6"} Oct 02 02:07:27 crc kubenswrapper[4885]: I1002 02:07:27.690287 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3555c77f-8b14-48c5-a9e1-42e8856dfe19","Type":"ContainerDied","Data":"2e53ca06a2d7abcfee83265ec4e450bf329550e5d651058c1dfbd19240513dc5"} Oct 02 02:07:27 crc kubenswrapper[4885]: I1002 02:07:27.690312 4885 scope.go:117] "RemoveContainer" containerID="ba1f877c8c7a3dc96a1a7cbad1d575537e1a8afaf77c80ebce637df8113780d6" Oct 02 02:07:27 crc kubenswrapper[4885]: I1002 02:07:27.693355 4885 generic.go:334] "Generic (PLEG): container finished" podID="4b8ef3c7-8dd5-4929-82b4-f0623ba96273" containerID="7bdb98fe4014f2819fdf39a14ebc95521fe001f0c806a90ca817e21997584d5e" exitCode=0 Oct 02 02:07:27 crc kubenswrapper[4885]: I1002 02:07:27.693427 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4b8ef3c7-8dd5-4929-82b4-f0623ba96273","Type":"ContainerDied","Data":"7bdb98fe4014f2819fdf39a14ebc95521fe001f0c806a90ca817e21997584d5e"} Oct 02 02:07:27 crc kubenswrapper[4885]: I1002 02:07:27.693451 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4b8ef3c7-8dd5-4929-82b4-f0623ba96273","Type":"ContainerDied","Data":"2194d75fe5f5ff3826fc38b1dc38606fe378b96d13e4ef1441230bd6f51ea060"} Oct 02 02:07:27 crc kubenswrapper[4885]: I1002 02:07:27.693539 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 02:07:27 crc kubenswrapper[4885]: I1002 02:07:27.694748 4885 generic.go:334] "Generic (PLEG): container finished" podID="6895f52a-cc09-4eac-9f57-32f91106afb9" containerID="77022366cfd5951f2c9ecb6edf329467419834685af3580f17b8fdf85b2921cb" exitCode=0 Oct 02 02:07:27 crc kubenswrapper[4885]: I1002 02:07:27.694778 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"6895f52a-cc09-4eac-9f57-32f91106afb9","Type":"ContainerDied","Data":"77022366cfd5951f2c9ecb6edf329467419834685af3580f17b8fdf85b2921cb"} Oct 02 02:07:27 crc kubenswrapper[4885]: I1002 02:07:27.694798 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"6895f52a-cc09-4eac-9f57-32f91106afb9","Type":"ContainerDied","Data":"9adaa4894606bb5b8256c9368f407b32063c570f722a4768c232c7f5fd03469a"} Oct 02 02:07:27 crc kubenswrapper[4885]: I1002 02:07:27.694843 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 02:07:27 crc kubenswrapper[4885]: I1002 02:07:27.724659 4885 scope.go:117] "RemoveContainer" containerID="f52ccc636851ade8d152837715411fd00cf473e5b713d8ff129c166e1d595b01" Oct 02 02:07:27 crc kubenswrapper[4885]: I1002 02:07:27.730712 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3555c77f-8b14-48c5-a9e1-42e8856dfe19-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "3555c77f-8b14-48c5-a9e1-42e8856dfe19" (UID: "3555c77f-8b14-48c5-a9e1-42e8856dfe19"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:07:27 crc kubenswrapper[4885]: I1002 02:07:27.741554 4885 scope.go:117] "RemoveContainer" containerID="ba1f877c8c7a3dc96a1a7cbad1d575537e1a8afaf77c80ebce637df8113780d6" Oct 02 02:07:27 crc kubenswrapper[4885]: I1002 02:07:27.744274 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6895f52a-cc09-4eac-9f57-32f91106afb9-combined-ca-bundle\") pod \"6895f52a-cc09-4eac-9f57-32f91106afb9\" (UID: \"6895f52a-cc09-4eac-9f57-32f91106afb9\") " Oct 02 02:07:27 crc kubenswrapper[4885]: I1002 02:07:27.744325 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6895f52a-cc09-4eac-9f57-32f91106afb9-config-data\") pod \"6895f52a-cc09-4eac-9f57-32f91106afb9\" (UID: \"6895f52a-cc09-4eac-9f57-32f91106afb9\") " Oct 02 02:07:27 crc kubenswrapper[4885]: I1002 02:07:27.744380 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gzs6m\" (UniqueName: \"kubernetes.io/projected/6895f52a-cc09-4eac-9f57-32f91106afb9-kube-api-access-gzs6m\") pod \"6895f52a-cc09-4eac-9f57-32f91106afb9\" (UID: \"6895f52a-cc09-4eac-9f57-32f91106afb9\") " Oct 02 02:07:27 crc kubenswrapper[4885]: I1002 02:07:27.744783 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2h5c6\" (UniqueName: \"kubernetes.io/projected/3555c77f-8b14-48c5-a9e1-42e8856dfe19-kube-api-access-2h5c6\") on node \"crc\" DevicePath \"\"" Oct 02 02:07:27 crc kubenswrapper[4885]: I1002 02:07:27.744799 4885 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3555c77f-8b14-48c5-a9e1-42e8856dfe19-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:07:27 crc kubenswrapper[4885]: I1002 02:07:27.744809 4885 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3555c77f-8b14-48c5-a9e1-42e8856dfe19-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 02:07:27 crc kubenswrapper[4885]: I1002 02:07:27.744817 4885 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/3555c77f-8b14-48c5-a9e1-42e8856dfe19-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 02:07:27 crc kubenswrapper[4885]: E1002 02:07:27.745523 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ba1f877c8c7a3dc96a1a7cbad1d575537e1a8afaf77c80ebce637df8113780d6\": container with ID starting with ba1f877c8c7a3dc96a1a7cbad1d575537e1a8afaf77c80ebce637df8113780d6 not found: ID does not exist" containerID="ba1f877c8c7a3dc96a1a7cbad1d575537e1a8afaf77c80ebce637df8113780d6" Oct 02 02:07:27 crc kubenswrapper[4885]: I1002 02:07:27.745573 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba1f877c8c7a3dc96a1a7cbad1d575537e1a8afaf77c80ebce637df8113780d6"} err="failed to get container status \"ba1f877c8c7a3dc96a1a7cbad1d575537e1a8afaf77c80ebce637df8113780d6\": rpc error: code = NotFound desc = could not find container \"ba1f877c8c7a3dc96a1a7cbad1d575537e1a8afaf77c80ebce637df8113780d6\": container with ID starting with ba1f877c8c7a3dc96a1a7cbad1d575537e1a8afaf77c80ebce637df8113780d6 not found: ID does not exist" Oct 02 02:07:27 crc kubenswrapper[4885]: I1002 02:07:27.745600 4885 scope.go:117] "RemoveContainer" containerID="f52ccc636851ade8d152837715411fd00cf473e5b713d8ff129c166e1d595b01" Oct 02 02:07:27 crc kubenswrapper[4885]: E1002 02:07:27.745986 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f52ccc636851ade8d152837715411fd00cf473e5b713d8ff129c166e1d595b01\": container with ID starting with f52ccc636851ade8d152837715411fd00cf473e5b713d8ff129c166e1d595b01 not found: ID does not exist" containerID="f52ccc636851ade8d152837715411fd00cf473e5b713d8ff129c166e1d595b01" Oct 02 02:07:27 crc kubenswrapper[4885]: I1002 02:07:27.746036 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f52ccc636851ade8d152837715411fd00cf473e5b713d8ff129c166e1d595b01"} err="failed to get container status \"f52ccc636851ade8d152837715411fd00cf473e5b713d8ff129c166e1d595b01\": rpc error: code = NotFound desc = could not find container \"f52ccc636851ade8d152837715411fd00cf473e5b713d8ff129c166e1d595b01\": container with ID starting with f52ccc636851ade8d152837715411fd00cf473e5b713d8ff129c166e1d595b01 not found: ID does not exist" Oct 02 02:07:27 crc kubenswrapper[4885]: I1002 02:07:27.746071 4885 scope.go:117] "RemoveContainer" containerID="7bdb98fe4014f2819fdf39a14ebc95521fe001f0c806a90ca817e21997584d5e" Oct 02 02:07:27 crc kubenswrapper[4885]: I1002 02:07:27.749850 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6895f52a-cc09-4eac-9f57-32f91106afb9-kube-api-access-gzs6m" (OuterVolumeSpecName: "kube-api-access-gzs6m") pod "6895f52a-cc09-4eac-9f57-32f91106afb9" (UID: "6895f52a-cc09-4eac-9f57-32f91106afb9"). InnerVolumeSpecName "kube-api-access-gzs6m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:07:27 crc kubenswrapper[4885]: I1002 02:07:27.771697 4885 scope.go:117] "RemoveContainer" containerID="d85e1506d9190c03bc565532abfe29da275335b9157af3a6ab6900591a8c752d" Oct 02 02:07:27 crc kubenswrapper[4885]: I1002 02:07:27.777852 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6895f52a-cc09-4eac-9f57-32f91106afb9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6895f52a-cc09-4eac-9f57-32f91106afb9" (UID: "6895f52a-cc09-4eac-9f57-32f91106afb9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:07:27 crc kubenswrapper[4885]: I1002 02:07:27.784063 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6895f52a-cc09-4eac-9f57-32f91106afb9-config-data" (OuterVolumeSpecName: "config-data") pod "6895f52a-cc09-4eac-9f57-32f91106afb9" (UID: "6895f52a-cc09-4eac-9f57-32f91106afb9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:07:27 crc kubenswrapper[4885]: I1002 02:07:27.791563 4885 scope.go:117] "RemoveContainer" containerID="7bdb98fe4014f2819fdf39a14ebc95521fe001f0c806a90ca817e21997584d5e" Oct 02 02:07:27 crc kubenswrapper[4885]: E1002 02:07:27.791922 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7bdb98fe4014f2819fdf39a14ebc95521fe001f0c806a90ca817e21997584d5e\": container with ID starting with 7bdb98fe4014f2819fdf39a14ebc95521fe001f0c806a90ca817e21997584d5e not found: ID does not exist" containerID="7bdb98fe4014f2819fdf39a14ebc95521fe001f0c806a90ca817e21997584d5e" Oct 02 02:07:27 crc kubenswrapper[4885]: I1002 02:07:27.791958 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7bdb98fe4014f2819fdf39a14ebc95521fe001f0c806a90ca817e21997584d5e"} err="failed to get container status \"7bdb98fe4014f2819fdf39a14ebc95521fe001f0c806a90ca817e21997584d5e\": rpc error: code = NotFound desc = could not find container \"7bdb98fe4014f2819fdf39a14ebc95521fe001f0c806a90ca817e21997584d5e\": container with ID starting with 7bdb98fe4014f2819fdf39a14ebc95521fe001f0c806a90ca817e21997584d5e not found: ID does not exist" Oct 02 02:07:27 crc kubenswrapper[4885]: I1002 02:07:27.791990 4885 scope.go:117] "RemoveContainer" containerID="d85e1506d9190c03bc565532abfe29da275335b9157af3a6ab6900591a8c752d" Oct 02 02:07:27 crc kubenswrapper[4885]: E1002 02:07:27.792362 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d85e1506d9190c03bc565532abfe29da275335b9157af3a6ab6900591a8c752d\": container with ID starting with d85e1506d9190c03bc565532abfe29da275335b9157af3a6ab6900591a8c752d not found: ID does not exist" containerID="d85e1506d9190c03bc565532abfe29da275335b9157af3a6ab6900591a8c752d" Oct 02 02:07:27 crc kubenswrapper[4885]: I1002 02:07:27.792416 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d85e1506d9190c03bc565532abfe29da275335b9157af3a6ab6900591a8c752d"} err="failed to get container status \"d85e1506d9190c03bc565532abfe29da275335b9157af3a6ab6900591a8c752d\": rpc error: code = NotFound desc = could not find container \"d85e1506d9190c03bc565532abfe29da275335b9157af3a6ab6900591a8c752d\": container with ID starting with d85e1506d9190c03bc565532abfe29da275335b9157af3a6ab6900591a8c752d not found: ID does not exist" Oct 02 02:07:27 crc kubenswrapper[4885]: I1002 02:07:27.792455 4885 scope.go:117] "RemoveContainer" containerID="77022366cfd5951f2c9ecb6edf329467419834685af3580f17b8fdf85b2921cb" Oct 02 02:07:27 crc kubenswrapper[4885]: I1002 02:07:27.831811 4885 scope.go:117] "RemoveContainer" containerID="77022366cfd5951f2c9ecb6edf329467419834685af3580f17b8fdf85b2921cb" Oct 02 02:07:27 crc kubenswrapper[4885]: E1002 02:07:27.832594 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"77022366cfd5951f2c9ecb6edf329467419834685af3580f17b8fdf85b2921cb\": container with ID starting with 77022366cfd5951f2c9ecb6edf329467419834685af3580f17b8fdf85b2921cb not found: ID does not exist" containerID="77022366cfd5951f2c9ecb6edf329467419834685af3580f17b8fdf85b2921cb" Oct 02 02:07:27 crc kubenswrapper[4885]: I1002 02:07:27.832665 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"77022366cfd5951f2c9ecb6edf329467419834685af3580f17b8fdf85b2921cb"} err="failed to get container status \"77022366cfd5951f2c9ecb6edf329467419834685af3580f17b8fdf85b2921cb\": rpc error: code = NotFound desc = could not find container \"77022366cfd5951f2c9ecb6edf329467419834685af3580f17b8fdf85b2921cb\": container with ID starting with 77022366cfd5951f2c9ecb6edf329467419834685af3580f17b8fdf85b2921cb not found: ID does not exist" Oct 02 02:07:27 crc kubenswrapper[4885]: I1002 02:07:27.846941 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4b8ef3c7-8dd5-4929-82b4-f0623ba96273-public-tls-certs\") pod \"4b8ef3c7-8dd5-4929-82b4-f0623ba96273\" (UID: \"4b8ef3c7-8dd5-4929-82b4-f0623ba96273\") " Oct 02 02:07:27 crc kubenswrapper[4885]: I1002 02:07:27.847350 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4b8ef3c7-8dd5-4929-82b4-f0623ba96273-internal-tls-certs\") pod \"4b8ef3c7-8dd5-4929-82b4-f0623ba96273\" (UID: \"4b8ef3c7-8dd5-4929-82b4-f0623ba96273\") " Oct 02 02:07:27 crc kubenswrapper[4885]: I1002 02:07:27.847451 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b8ef3c7-8dd5-4929-82b4-f0623ba96273-config-data\") pod \"4b8ef3c7-8dd5-4929-82b4-f0623ba96273\" (UID: \"4b8ef3c7-8dd5-4929-82b4-f0623ba96273\") " Oct 02 02:07:27 crc kubenswrapper[4885]: I1002 02:07:27.847708 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ms77x\" (UniqueName: \"kubernetes.io/projected/4b8ef3c7-8dd5-4929-82b4-f0623ba96273-kube-api-access-ms77x\") pod \"4b8ef3c7-8dd5-4929-82b4-f0623ba96273\" (UID: \"4b8ef3c7-8dd5-4929-82b4-f0623ba96273\") " Oct 02 02:07:27 crc kubenswrapper[4885]: I1002 02:07:27.847826 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b8ef3c7-8dd5-4929-82b4-f0623ba96273-combined-ca-bundle\") pod \"4b8ef3c7-8dd5-4929-82b4-f0623ba96273\" (UID: \"4b8ef3c7-8dd5-4929-82b4-f0623ba96273\") " Oct 02 02:07:27 crc kubenswrapper[4885]: I1002 02:07:27.848026 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4b8ef3c7-8dd5-4929-82b4-f0623ba96273-logs\") pod \"4b8ef3c7-8dd5-4929-82b4-f0623ba96273\" (UID: \"4b8ef3c7-8dd5-4929-82b4-f0623ba96273\") " Oct 02 02:07:27 crc kubenswrapper[4885]: I1002 02:07:27.848851 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4b8ef3c7-8dd5-4929-82b4-f0623ba96273-logs" (OuterVolumeSpecName: "logs") pod "4b8ef3c7-8dd5-4929-82b4-f0623ba96273" (UID: "4b8ef3c7-8dd5-4929-82b4-f0623ba96273"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:07:27 crc kubenswrapper[4885]: I1002 02:07:27.849730 4885 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4b8ef3c7-8dd5-4929-82b4-f0623ba96273-logs\") on node \"crc\" DevicePath \"\"" Oct 02 02:07:27 crc kubenswrapper[4885]: I1002 02:07:27.849770 4885 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6895f52a-cc09-4eac-9f57-32f91106afb9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:07:27 crc kubenswrapper[4885]: I1002 02:07:27.849794 4885 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6895f52a-cc09-4eac-9f57-32f91106afb9-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 02:07:27 crc kubenswrapper[4885]: I1002 02:07:27.849813 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gzs6m\" (UniqueName: \"kubernetes.io/projected/6895f52a-cc09-4eac-9f57-32f91106afb9-kube-api-access-gzs6m\") on node \"crc\" DevicePath \"\"" Oct 02 02:07:27 crc kubenswrapper[4885]: I1002 02:07:27.851772 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b8ef3c7-8dd5-4929-82b4-f0623ba96273-kube-api-access-ms77x" (OuterVolumeSpecName: "kube-api-access-ms77x") pod "4b8ef3c7-8dd5-4929-82b4-f0623ba96273" (UID: "4b8ef3c7-8dd5-4929-82b4-f0623ba96273"). InnerVolumeSpecName "kube-api-access-ms77x". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:07:27 crc kubenswrapper[4885]: I1002 02:07:27.876300 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b8ef3c7-8dd5-4929-82b4-f0623ba96273-config-data" (OuterVolumeSpecName: "config-data") pod "4b8ef3c7-8dd5-4929-82b4-f0623ba96273" (UID: "4b8ef3c7-8dd5-4929-82b4-f0623ba96273"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:07:27 crc kubenswrapper[4885]: I1002 02:07:27.876777 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b8ef3c7-8dd5-4929-82b4-f0623ba96273-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4b8ef3c7-8dd5-4929-82b4-f0623ba96273" (UID: "4b8ef3c7-8dd5-4929-82b4-f0623ba96273"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:07:27 crc kubenswrapper[4885]: I1002 02:07:27.902601 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b8ef3c7-8dd5-4929-82b4-f0623ba96273-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "4b8ef3c7-8dd5-4929-82b4-f0623ba96273" (UID: "4b8ef3c7-8dd5-4929-82b4-f0623ba96273"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:07:27 crc kubenswrapper[4885]: I1002 02:07:27.903990 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b8ef3c7-8dd5-4929-82b4-f0623ba96273-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "4b8ef3c7-8dd5-4929-82b4-f0623ba96273" (UID: "4b8ef3c7-8dd5-4929-82b4-f0623ba96273"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:07:27 crc kubenswrapper[4885]: I1002 02:07:27.951043 4885 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4b8ef3c7-8dd5-4929-82b4-f0623ba96273-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 02:07:27 crc kubenswrapper[4885]: I1002 02:07:27.951079 4885 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b8ef3c7-8dd5-4929-82b4-f0623ba96273-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 02:07:27 crc kubenswrapper[4885]: I1002 02:07:27.951088 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ms77x\" (UniqueName: \"kubernetes.io/projected/4b8ef3c7-8dd5-4929-82b4-f0623ba96273-kube-api-access-ms77x\") on node \"crc\" DevicePath \"\"" Oct 02 02:07:27 crc kubenswrapper[4885]: I1002 02:07:27.951100 4885 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b8ef3c7-8dd5-4929-82b4-f0623ba96273-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:07:27 crc kubenswrapper[4885]: I1002 02:07:27.951110 4885 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4b8ef3c7-8dd5-4929-82b4-f0623ba96273-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 02:07:28 crc kubenswrapper[4885]: I1002 02:07:28.032301 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 02 02:07:28 crc kubenswrapper[4885]: I1002 02:07:28.064494 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 02 02:07:28 crc kubenswrapper[4885]: I1002 02:07:28.065934 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 02:07:28 crc kubenswrapper[4885]: I1002 02:07:28.082890 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 02:07:28 crc kubenswrapper[4885]: I1002 02:07:28.089410 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 02 02:07:28 crc kubenswrapper[4885]: E1002 02:07:28.089911 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6895f52a-cc09-4eac-9f57-32f91106afb9" containerName="nova-scheduler-scheduler" Oct 02 02:07:28 crc kubenswrapper[4885]: I1002 02:07:28.089931 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="6895f52a-cc09-4eac-9f57-32f91106afb9" containerName="nova-scheduler-scheduler" Oct 02 02:07:28 crc kubenswrapper[4885]: E1002 02:07:28.089963 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b8ef3c7-8dd5-4929-82b4-f0623ba96273" containerName="nova-api-api" Oct 02 02:07:28 crc kubenswrapper[4885]: I1002 02:07:28.089970 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b8ef3c7-8dd5-4929-82b4-f0623ba96273" containerName="nova-api-api" Oct 02 02:07:28 crc kubenswrapper[4885]: E1002 02:07:28.089980 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3555c77f-8b14-48c5-a9e1-42e8856dfe19" containerName="nova-metadata-metadata" Oct 02 02:07:28 crc kubenswrapper[4885]: I1002 02:07:28.089986 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="3555c77f-8b14-48c5-a9e1-42e8856dfe19" containerName="nova-metadata-metadata" Oct 02 02:07:28 crc kubenswrapper[4885]: E1002 02:07:28.089993 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b8ef3c7-8dd5-4929-82b4-f0623ba96273" containerName="nova-api-log" Oct 02 02:07:28 crc kubenswrapper[4885]: I1002 02:07:28.090000 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b8ef3c7-8dd5-4929-82b4-f0623ba96273" containerName="nova-api-log" Oct 02 02:07:28 crc kubenswrapper[4885]: E1002 02:07:28.090015 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3555c77f-8b14-48c5-a9e1-42e8856dfe19" containerName="nova-metadata-log" Oct 02 02:07:28 crc kubenswrapper[4885]: I1002 02:07:28.090022 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="3555c77f-8b14-48c5-a9e1-42e8856dfe19" containerName="nova-metadata-log" Oct 02 02:07:28 crc kubenswrapper[4885]: E1002 02:07:28.090031 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a91d000d-543e-42b4-b815-12243df15ea9" containerName="nova-manage" Oct 02 02:07:28 crc kubenswrapper[4885]: I1002 02:07:28.090036 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="a91d000d-543e-42b4-b815-12243df15ea9" containerName="nova-manage" Oct 02 02:07:28 crc kubenswrapper[4885]: I1002 02:07:28.090249 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="a91d000d-543e-42b4-b815-12243df15ea9" containerName="nova-manage" Oct 02 02:07:28 crc kubenswrapper[4885]: I1002 02:07:28.090276 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="6895f52a-cc09-4eac-9f57-32f91106afb9" containerName="nova-scheduler-scheduler" Oct 02 02:07:28 crc kubenswrapper[4885]: I1002 02:07:28.090288 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b8ef3c7-8dd5-4929-82b4-f0623ba96273" containerName="nova-api-log" Oct 02 02:07:28 crc kubenswrapper[4885]: I1002 02:07:28.090296 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="3555c77f-8b14-48c5-a9e1-42e8856dfe19" containerName="nova-metadata-metadata" Oct 02 02:07:28 crc kubenswrapper[4885]: I1002 02:07:28.090310 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b8ef3c7-8dd5-4929-82b4-f0623ba96273" containerName="nova-api-api" Oct 02 02:07:28 crc kubenswrapper[4885]: I1002 02:07:28.090324 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="3555c77f-8b14-48c5-a9e1-42e8856dfe19" containerName="nova-metadata-log" Oct 02 02:07:28 crc kubenswrapper[4885]: I1002 02:07:28.091371 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 02:07:28 crc kubenswrapper[4885]: I1002 02:07:28.093443 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 02 02:07:28 crc kubenswrapper[4885]: I1002 02:07:28.093726 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 02 02:07:28 crc kubenswrapper[4885]: I1002 02:07:28.098789 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 02 02:07:28 crc kubenswrapper[4885]: I1002 02:07:28.108112 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 02:07:28 crc kubenswrapper[4885]: I1002 02:07:28.115400 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 02:07:28 crc kubenswrapper[4885]: I1002 02:07:28.136004 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 02:07:28 crc kubenswrapper[4885]: I1002 02:07:28.137216 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 02:07:28 crc kubenswrapper[4885]: I1002 02:07:28.139832 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 02 02:07:28 crc kubenswrapper[4885]: I1002 02:07:28.166110 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 02 02:07:28 crc kubenswrapper[4885]: I1002 02:07:28.180580 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 02:07:28 crc kubenswrapper[4885]: I1002 02:07:28.188960 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 02 02:07:28 crc kubenswrapper[4885]: I1002 02:07:28.191924 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 02:07:28 crc kubenswrapper[4885]: I1002 02:07:28.194603 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 02 02:07:28 crc kubenswrapper[4885]: I1002 02:07:28.194918 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 02 02:07:28 crc kubenswrapper[4885]: I1002 02:07:28.208847 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 02:07:28 crc kubenswrapper[4885]: I1002 02:07:28.257562 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50278f9c-b6fc-437c-ba3a-972b857493e3-config-data\") pod \"nova-scheduler-0\" (UID: \"50278f9c-b6fc-437c-ba3a-972b857493e3\") " pod="openstack/nova-scheduler-0" Oct 02 02:07:28 crc kubenswrapper[4885]: I1002 02:07:28.257654 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lnjnv\" (UniqueName: \"kubernetes.io/projected/50278f9c-b6fc-437c-ba3a-972b857493e3-kube-api-access-lnjnv\") pod \"nova-scheduler-0\" (UID: \"50278f9c-b6fc-437c-ba3a-972b857493e3\") " pod="openstack/nova-scheduler-0" Oct 02 02:07:28 crc kubenswrapper[4885]: I1002 02:07:28.257694 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71959149-27b9-4b03-aa11-1c2164677d5c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"71959149-27b9-4b03-aa11-1c2164677d5c\") " pod="openstack/nova-api-0" Oct 02 02:07:28 crc kubenswrapper[4885]: I1002 02:07:28.257826 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50278f9c-b6fc-437c-ba3a-972b857493e3-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"50278f9c-b6fc-437c-ba3a-972b857493e3\") " pod="openstack/nova-scheduler-0" Oct 02 02:07:28 crc kubenswrapper[4885]: I1002 02:07:28.257940 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71959149-27b9-4b03-aa11-1c2164677d5c-config-data\") pod \"nova-api-0\" (UID: \"71959149-27b9-4b03-aa11-1c2164677d5c\") " pod="openstack/nova-api-0" Oct 02 02:07:28 crc kubenswrapper[4885]: I1002 02:07:28.257979 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kxcsn\" (UniqueName: \"kubernetes.io/projected/71959149-27b9-4b03-aa11-1c2164677d5c-kube-api-access-kxcsn\") pod \"nova-api-0\" (UID: \"71959149-27b9-4b03-aa11-1c2164677d5c\") " pod="openstack/nova-api-0" Oct 02 02:07:28 crc kubenswrapper[4885]: I1002 02:07:28.258119 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/71959149-27b9-4b03-aa11-1c2164677d5c-logs\") pod \"nova-api-0\" (UID: \"71959149-27b9-4b03-aa11-1c2164677d5c\") " pod="openstack/nova-api-0" Oct 02 02:07:28 crc kubenswrapper[4885]: I1002 02:07:28.258234 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/71959149-27b9-4b03-aa11-1c2164677d5c-public-tls-certs\") pod \"nova-api-0\" (UID: \"71959149-27b9-4b03-aa11-1c2164677d5c\") " pod="openstack/nova-api-0" Oct 02 02:07:28 crc kubenswrapper[4885]: I1002 02:07:28.258292 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/71959149-27b9-4b03-aa11-1c2164677d5c-internal-tls-certs\") pod \"nova-api-0\" (UID: \"71959149-27b9-4b03-aa11-1c2164677d5c\") " pod="openstack/nova-api-0" Oct 02 02:07:28 crc kubenswrapper[4885]: I1002 02:07:28.360806 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50278f9c-b6fc-437c-ba3a-972b857493e3-config-data\") pod \"nova-scheduler-0\" (UID: \"50278f9c-b6fc-437c-ba3a-972b857493e3\") " pod="openstack/nova-scheduler-0" Oct 02 02:07:28 crc kubenswrapper[4885]: I1002 02:07:28.361219 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lnjnv\" (UniqueName: \"kubernetes.io/projected/50278f9c-b6fc-437c-ba3a-972b857493e3-kube-api-access-lnjnv\") pod \"nova-scheduler-0\" (UID: \"50278f9c-b6fc-437c-ba3a-972b857493e3\") " pod="openstack/nova-scheduler-0" Oct 02 02:07:28 crc kubenswrapper[4885]: I1002 02:07:28.361273 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qm54k\" (UniqueName: \"kubernetes.io/projected/179284e1-8a2f-4987-a0b4-7b24b6e3cf94-kube-api-access-qm54k\") pod \"nova-metadata-0\" (UID: \"179284e1-8a2f-4987-a0b4-7b24b6e3cf94\") " pod="openstack/nova-metadata-0" Oct 02 02:07:28 crc kubenswrapper[4885]: I1002 02:07:28.361312 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71959149-27b9-4b03-aa11-1c2164677d5c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"71959149-27b9-4b03-aa11-1c2164677d5c\") " pod="openstack/nova-api-0" Oct 02 02:07:28 crc kubenswrapper[4885]: I1002 02:07:28.361347 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50278f9c-b6fc-437c-ba3a-972b857493e3-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"50278f9c-b6fc-437c-ba3a-972b857493e3\") " pod="openstack/nova-scheduler-0" Oct 02 02:07:28 crc kubenswrapper[4885]: I1002 02:07:28.361377 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71959149-27b9-4b03-aa11-1c2164677d5c-config-data\") pod \"nova-api-0\" (UID: \"71959149-27b9-4b03-aa11-1c2164677d5c\") " pod="openstack/nova-api-0" Oct 02 02:07:28 crc kubenswrapper[4885]: I1002 02:07:28.361398 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kxcsn\" (UniqueName: \"kubernetes.io/projected/71959149-27b9-4b03-aa11-1c2164677d5c-kube-api-access-kxcsn\") pod \"nova-api-0\" (UID: \"71959149-27b9-4b03-aa11-1c2164677d5c\") " pod="openstack/nova-api-0" Oct 02 02:07:28 crc kubenswrapper[4885]: I1002 02:07:28.361424 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/179284e1-8a2f-4987-a0b4-7b24b6e3cf94-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"179284e1-8a2f-4987-a0b4-7b24b6e3cf94\") " pod="openstack/nova-metadata-0" Oct 02 02:07:28 crc kubenswrapper[4885]: I1002 02:07:28.361468 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/71959149-27b9-4b03-aa11-1c2164677d5c-logs\") pod \"nova-api-0\" (UID: \"71959149-27b9-4b03-aa11-1c2164677d5c\") " pod="openstack/nova-api-0" Oct 02 02:07:28 crc kubenswrapper[4885]: I1002 02:07:28.361494 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/179284e1-8a2f-4987-a0b4-7b24b6e3cf94-logs\") pod \"nova-metadata-0\" (UID: \"179284e1-8a2f-4987-a0b4-7b24b6e3cf94\") " pod="openstack/nova-metadata-0" Oct 02 02:07:28 crc kubenswrapper[4885]: I1002 02:07:28.361516 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/179284e1-8a2f-4987-a0b4-7b24b6e3cf94-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"179284e1-8a2f-4987-a0b4-7b24b6e3cf94\") " pod="openstack/nova-metadata-0" Oct 02 02:07:28 crc kubenswrapper[4885]: I1002 02:07:28.361550 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/71959149-27b9-4b03-aa11-1c2164677d5c-public-tls-certs\") pod \"nova-api-0\" (UID: \"71959149-27b9-4b03-aa11-1c2164677d5c\") " pod="openstack/nova-api-0" Oct 02 02:07:28 crc kubenswrapper[4885]: I1002 02:07:28.361571 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/71959149-27b9-4b03-aa11-1c2164677d5c-internal-tls-certs\") pod \"nova-api-0\" (UID: \"71959149-27b9-4b03-aa11-1c2164677d5c\") " pod="openstack/nova-api-0" Oct 02 02:07:28 crc kubenswrapper[4885]: I1002 02:07:28.362234 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/71959149-27b9-4b03-aa11-1c2164677d5c-logs\") pod \"nova-api-0\" (UID: \"71959149-27b9-4b03-aa11-1c2164677d5c\") " pod="openstack/nova-api-0" Oct 02 02:07:28 crc kubenswrapper[4885]: I1002 02:07:28.362338 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/179284e1-8a2f-4987-a0b4-7b24b6e3cf94-config-data\") pod \"nova-metadata-0\" (UID: \"179284e1-8a2f-4987-a0b4-7b24b6e3cf94\") " pod="openstack/nova-metadata-0" Oct 02 02:07:28 crc kubenswrapper[4885]: I1002 02:07:28.367629 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/71959149-27b9-4b03-aa11-1c2164677d5c-internal-tls-certs\") pod \"nova-api-0\" (UID: \"71959149-27b9-4b03-aa11-1c2164677d5c\") " pod="openstack/nova-api-0" Oct 02 02:07:28 crc kubenswrapper[4885]: I1002 02:07:28.367726 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50278f9c-b6fc-437c-ba3a-972b857493e3-config-data\") pod \"nova-scheduler-0\" (UID: \"50278f9c-b6fc-437c-ba3a-972b857493e3\") " pod="openstack/nova-scheduler-0" Oct 02 02:07:28 crc kubenswrapper[4885]: I1002 02:07:28.367743 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71959149-27b9-4b03-aa11-1c2164677d5c-config-data\") pod \"nova-api-0\" (UID: \"71959149-27b9-4b03-aa11-1c2164677d5c\") " pod="openstack/nova-api-0" Oct 02 02:07:28 crc kubenswrapper[4885]: I1002 02:07:28.368413 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71959149-27b9-4b03-aa11-1c2164677d5c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"71959149-27b9-4b03-aa11-1c2164677d5c\") " pod="openstack/nova-api-0" Oct 02 02:07:28 crc kubenswrapper[4885]: I1002 02:07:28.372864 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50278f9c-b6fc-437c-ba3a-972b857493e3-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"50278f9c-b6fc-437c-ba3a-972b857493e3\") " pod="openstack/nova-scheduler-0" Oct 02 02:07:28 crc kubenswrapper[4885]: I1002 02:07:28.373045 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/71959149-27b9-4b03-aa11-1c2164677d5c-public-tls-certs\") pod \"nova-api-0\" (UID: \"71959149-27b9-4b03-aa11-1c2164677d5c\") " pod="openstack/nova-api-0" Oct 02 02:07:28 crc kubenswrapper[4885]: I1002 02:07:28.379142 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lnjnv\" (UniqueName: \"kubernetes.io/projected/50278f9c-b6fc-437c-ba3a-972b857493e3-kube-api-access-lnjnv\") pod \"nova-scheduler-0\" (UID: \"50278f9c-b6fc-437c-ba3a-972b857493e3\") " pod="openstack/nova-scheduler-0" Oct 02 02:07:28 crc kubenswrapper[4885]: I1002 02:07:28.387701 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kxcsn\" (UniqueName: \"kubernetes.io/projected/71959149-27b9-4b03-aa11-1c2164677d5c-kube-api-access-kxcsn\") pod \"nova-api-0\" (UID: \"71959149-27b9-4b03-aa11-1c2164677d5c\") " pod="openstack/nova-api-0" Oct 02 02:07:28 crc kubenswrapper[4885]: I1002 02:07:28.407975 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 02:07:28 crc kubenswrapper[4885]: I1002 02:07:28.464440 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qm54k\" (UniqueName: \"kubernetes.io/projected/179284e1-8a2f-4987-a0b4-7b24b6e3cf94-kube-api-access-qm54k\") pod \"nova-metadata-0\" (UID: \"179284e1-8a2f-4987-a0b4-7b24b6e3cf94\") " pod="openstack/nova-metadata-0" Oct 02 02:07:28 crc kubenswrapper[4885]: I1002 02:07:28.464516 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/179284e1-8a2f-4987-a0b4-7b24b6e3cf94-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"179284e1-8a2f-4987-a0b4-7b24b6e3cf94\") " pod="openstack/nova-metadata-0" Oct 02 02:07:28 crc kubenswrapper[4885]: I1002 02:07:28.464597 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/179284e1-8a2f-4987-a0b4-7b24b6e3cf94-logs\") pod \"nova-metadata-0\" (UID: \"179284e1-8a2f-4987-a0b4-7b24b6e3cf94\") " pod="openstack/nova-metadata-0" Oct 02 02:07:28 crc kubenswrapper[4885]: I1002 02:07:28.464615 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/179284e1-8a2f-4987-a0b4-7b24b6e3cf94-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"179284e1-8a2f-4987-a0b4-7b24b6e3cf94\") " pod="openstack/nova-metadata-0" Oct 02 02:07:28 crc kubenswrapper[4885]: I1002 02:07:28.464672 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/179284e1-8a2f-4987-a0b4-7b24b6e3cf94-config-data\") pod \"nova-metadata-0\" (UID: \"179284e1-8a2f-4987-a0b4-7b24b6e3cf94\") " pod="openstack/nova-metadata-0" Oct 02 02:07:28 crc kubenswrapper[4885]: I1002 02:07:28.465958 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/179284e1-8a2f-4987-a0b4-7b24b6e3cf94-logs\") pod \"nova-metadata-0\" (UID: \"179284e1-8a2f-4987-a0b4-7b24b6e3cf94\") " pod="openstack/nova-metadata-0" Oct 02 02:07:28 crc kubenswrapper[4885]: I1002 02:07:28.469369 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/179284e1-8a2f-4987-a0b4-7b24b6e3cf94-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"179284e1-8a2f-4987-a0b4-7b24b6e3cf94\") " pod="openstack/nova-metadata-0" Oct 02 02:07:28 crc kubenswrapper[4885]: I1002 02:07:28.469530 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/179284e1-8a2f-4987-a0b4-7b24b6e3cf94-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"179284e1-8a2f-4987-a0b4-7b24b6e3cf94\") " pod="openstack/nova-metadata-0" Oct 02 02:07:28 crc kubenswrapper[4885]: I1002 02:07:28.471728 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/179284e1-8a2f-4987-a0b4-7b24b6e3cf94-config-data\") pod \"nova-metadata-0\" (UID: \"179284e1-8a2f-4987-a0b4-7b24b6e3cf94\") " pod="openstack/nova-metadata-0" Oct 02 02:07:28 crc kubenswrapper[4885]: I1002 02:07:28.487946 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qm54k\" (UniqueName: \"kubernetes.io/projected/179284e1-8a2f-4987-a0b4-7b24b6e3cf94-kube-api-access-qm54k\") pod \"nova-metadata-0\" (UID: \"179284e1-8a2f-4987-a0b4-7b24b6e3cf94\") " pod="openstack/nova-metadata-0" Oct 02 02:07:28 crc kubenswrapper[4885]: I1002 02:07:28.522992 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 02:07:28 crc kubenswrapper[4885]: I1002 02:07:28.536504 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 02:07:28 crc kubenswrapper[4885]: I1002 02:07:28.953541 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 02 02:07:28 crc kubenswrapper[4885]: W1002 02:07:28.959432 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod71959149_27b9_4b03_aa11_1c2164677d5c.slice/crio-d66928981d1a48e2291e37ba0b3ad4eb677265052b2a35e2396f4d25989220de WatchSource:0}: Error finding container d66928981d1a48e2291e37ba0b3ad4eb677265052b2a35e2396f4d25989220de: Status 404 returned error can't find the container with id d66928981d1a48e2291e37ba0b3ad4eb677265052b2a35e2396f4d25989220de Oct 02 02:07:29 crc kubenswrapper[4885]: I1002 02:07:29.054339 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 02:07:29 crc kubenswrapper[4885]: I1002 02:07:29.062678 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 02:07:29 crc kubenswrapper[4885]: W1002 02:07:29.062980 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod50278f9c_b6fc_437c_ba3a_972b857493e3.slice/crio-d2406e4e0be51f2589701ef7e62c31da8b870fe44b634e63413e994fb7b3427f WatchSource:0}: Error finding container d2406e4e0be51f2589701ef7e62c31da8b870fe44b634e63413e994fb7b3427f: Status 404 returned error can't find the container with id d2406e4e0be51f2589701ef7e62c31da8b870fe44b634e63413e994fb7b3427f Oct 02 02:07:29 crc kubenswrapper[4885]: W1002 02:07:29.063928 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod179284e1_8a2f_4987_a0b4_7b24b6e3cf94.slice/crio-859a78cc7bb0c611c560cdef63422200826ff76e430dfbc0061ac2799ec30b52 WatchSource:0}: Error finding container 859a78cc7bb0c611c560cdef63422200826ff76e430dfbc0061ac2799ec30b52: Status 404 returned error can't find the container with id 859a78cc7bb0c611c560cdef63422200826ff76e430dfbc0061ac2799ec30b52 Oct 02 02:07:29 crc kubenswrapper[4885]: I1002 02:07:29.727473 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"179284e1-8a2f-4987-a0b4-7b24b6e3cf94","Type":"ContainerStarted","Data":"06a94aa97bc100e67762c75190385cda7987288c50d7abed84ffefdbbd9c8a56"} Oct 02 02:07:29 crc kubenswrapper[4885]: I1002 02:07:29.727748 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"179284e1-8a2f-4987-a0b4-7b24b6e3cf94","Type":"ContainerStarted","Data":"34272a9db18aae4c8885eb2f38c650f4b8fe223f9fdfc6aa1379335e20a3d742"} Oct 02 02:07:29 crc kubenswrapper[4885]: I1002 02:07:29.727764 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"179284e1-8a2f-4987-a0b4-7b24b6e3cf94","Type":"ContainerStarted","Data":"859a78cc7bb0c611c560cdef63422200826ff76e430dfbc0061ac2799ec30b52"} Oct 02 02:07:29 crc kubenswrapper[4885]: I1002 02:07:29.731597 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"71959149-27b9-4b03-aa11-1c2164677d5c","Type":"ContainerStarted","Data":"a54e8bbc4793abd8808e566772ac02669d47f997e4eddee1313af0a7075c9c06"} Oct 02 02:07:29 crc kubenswrapper[4885]: I1002 02:07:29.731671 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"71959149-27b9-4b03-aa11-1c2164677d5c","Type":"ContainerStarted","Data":"35a399c83c1890a100d9e1e211acb7cefd9d3be1745b9a327ef66ea9ef517b1b"} Oct 02 02:07:29 crc kubenswrapper[4885]: I1002 02:07:29.731691 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"71959149-27b9-4b03-aa11-1c2164677d5c","Type":"ContainerStarted","Data":"d66928981d1a48e2291e37ba0b3ad4eb677265052b2a35e2396f4d25989220de"} Oct 02 02:07:29 crc kubenswrapper[4885]: I1002 02:07:29.733896 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"50278f9c-b6fc-437c-ba3a-972b857493e3","Type":"ContainerStarted","Data":"3ca0a204c3fe86fe1c2199568a14849df44a16e66058d4ad0fa460de34a1f548"} Oct 02 02:07:29 crc kubenswrapper[4885]: I1002 02:07:29.733934 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"50278f9c-b6fc-437c-ba3a-972b857493e3","Type":"ContainerStarted","Data":"d2406e4e0be51f2589701ef7e62c31da8b870fe44b634e63413e994fb7b3427f"} Oct 02 02:07:29 crc kubenswrapper[4885]: I1002 02:07:29.750850 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=1.7508319 podStartE2EDuration="1.7508319s" podCreationTimestamp="2025-10-02 02:07:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:07:29.744142087 +0000 UTC m=+1238.555889506" watchObservedRunningTime="2025-10-02 02:07:29.7508319 +0000 UTC m=+1238.562579299" Oct 02 02:07:29 crc kubenswrapper[4885]: I1002 02:07:29.769549 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=1.769532947 podStartE2EDuration="1.769532947s" podCreationTimestamp="2025-10-02 02:07:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:07:29.763632718 +0000 UTC m=+1238.575380127" watchObservedRunningTime="2025-10-02 02:07:29.769532947 +0000 UTC m=+1238.581280346" Oct 02 02:07:29 crc kubenswrapper[4885]: I1002 02:07:29.793236 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=1.793216157 podStartE2EDuration="1.793216157s" podCreationTimestamp="2025-10-02 02:07:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:07:29.78707108 +0000 UTC m=+1238.598818479" watchObservedRunningTime="2025-10-02 02:07:29.793216157 +0000 UTC m=+1238.604963576" Oct 02 02:07:30 crc kubenswrapper[4885]: I1002 02:07:30.067786 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3555c77f-8b14-48c5-a9e1-42e8856dfe19" path="/var/lib/kubelet/pods/3555c77f-8b14-48c5-a9e1-42e8856dfe19/volumes" Oct 02 02:07:30 crc kubenswrapper[4885]: I1002 02:07:30.069657 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4b8ef3c7-8dd5-4929-82b4-f0623ba96273" path="/var/lib/kubelet/pods/4b8ef3c7-8dd5-4929-82b4-f0623ba96273/volumes" Oct 02 02:07:30 crc kubenswrapper[4885]: I1002 02:07:30.077813 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6895f52a-cc09-4eac-9f57-32f91106afb9" path="/var/lib/kubelet/pods/6895f52a-cc09-4eac-9f57-32f91106afb9/volumes" Oct 02 02:07:33 crc kubenswrapper[4885]: I1002 02:07:33.523624 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 02 02:07:33 crc kubenswrapper[4885]: I1002 02:07:33.536850 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 02 02:07:33 crc kubenswrapper[4885]: I1002 02:07:33.536951 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 02 02:07:34 crc kubenswrapper[4885]: I1002 02:07:34.847502 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 02 02:07:38 crc kubenswrapper[4885]: I1002 02:07:38.408977 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 02 02:07:38 crc kubenswrapper[4885]: I1002 02:07:38.409955 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 02 02:07:38 crc kubenswrapper[4885]: I1002 02:07:38.523722 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 02 02:07:38 crc kubenswrapper[4885]: I1002 02:07:38.537891 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 02 02:07:38 crc kubenswrapper[4885]: I1002 02:07:38.537962 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 02 02:07:38 crc kubenswrapper[4885]: I1002 02:07:38.549404 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 02 02:07:38 crc kubenswrapper[4885]: I1002 02:07:38.880893 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 02 02:07:39 crc kubenswrapper[4885]: I1002 02:07:39.418453 4885 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="71959149-27b9-4b03-aa11-1c2164677d5c" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.206:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 02 02:07:39 crc kubenswrapper[4885]: I1002 02:07:39.418487 4885 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="71959149-27b9-4b03-aa11-1c2164677d5c" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.206:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 02 02:07:39 crc kubenswrapper[4885]: I1002 02:07:39.554445 4885 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="179284e1-8a2f-4987-a0b4-7b24b6e3cf94" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.208:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 02 02:07:39 crc kubenswrapper[4885]: I1002 02:07:39.554486 4885 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="179284e1-8a2f-4987-a0b4-7b24b6e3cf94" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.208:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 02 02:07:48 crc kubenswrapper[4885]: I1002 02:07:48.422628 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 02 02:07:48 crc kubenswrapper[4885]: I1002 02:07:48.423379 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 02 02:07:48 crc kubenswrapper[4885]: I1002 02:07:48.423955 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 02 02:07:48 crc kubenswrapper[4885]: I1002 02:07:48.424206 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 02 02:07:48 crc kubenswrapper[4885]: I1002 02:07:48.432801 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 02 02:07:48 crc kubenswrapper[4885]: I1002 02:07:48.434200 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 02 02:07:48 crc kubenswrapper[4885]: I1002 02:07:48.550634 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 02 02:07:48 crc kubenswrapper[4885]: I1002 02:07:48.553464 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 02 02:07:48 crc kubenswrapper[4885]: I1002 02:07:48.578710 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 02 02:07:48 crc kubenswrapper[4885]: I1002 02:07:48.959713 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 02 02:07:56 crc kubenswrapper[4885]: I1002 02:07:56.935509 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 02 02:07:58 crc kubenswrapper[4885]: I1002 02:07:58.508454 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 02 02:08:01 crc kubenswrapper[4885]: I1002 02:08:01.395536 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="6692f52c-5ee9-4cd6-8674-0fb3f5fd6679" containerName="rabbitmq" containerID="cri-o://18584fe73de75abdffba2d36424556d76845cfba20bc3152c3df90230f533362" gracePeriod=604796 Oct 02 02:08:03 crc kubenswrapper[4885]: I1002 02:08:03.055894 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="eea66006-67ba-45e3-bdf9-9d144d772386" containerName="rabbitmq" containerID="cri-o://6d97c5a8d103c2c2b9ce70a5d8c6ef3e51104fe5a791fe6d1e52bdf3ad332dff" gracePeriod=604796 Oct 02 02:08:04 crc kubenswrapper[4885]: I1002 02:08:04.142878 4885 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="eea66006-67ba-45e3-bdf9-9d144d772386" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.101:5671: connect: connection refused" Oct 02 02:08:04 crc kubenswrapper[4885]: I1002 02:08:04.364575 4885 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="6692f52c-5ee9-4cd6-8674-0fb3f5fd6679" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.102:5671: connect: connection refused" Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.031397 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.153168 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/6692f52c-5ee9-4cd6-8674-0fb3f5fd6679-plugins-conf\") pod \"6692f52c-5ee9-4cd6-8674-0fb3f5fd6679\" (UID: \"6692f52c-5ee9-4cd6-8674-0fb3f5fd6679\") " Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.153223 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6692f52c-5ee9-4cd6-8674-0fb3f5fd6679-config-data\") pod \"6692f52c-5ee9-4cd6-8674-0fb3f5fd6679\" (UID: \"6692f52c-5ee9-4cd6-8674-0fb3f5fd6679\") " Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.153301 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/6692f52c-5ee9-4cd6-8674-0fb3f5fd6679-erlang-cookie-secret\") pod \"6692f52c-5ee9-4cd6-8674-0fb3f5fd6679\" (UID: \"6692f52c-5ee9-4cd6-8674-0fb3f5fd6679\") " Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.153336 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/6692f52c-5ee9-4cd6-8674-0fb3f5fd6679-rabbitmq-confd\") pod \"6692f52c-5ee9-4cd6-8674-0fb3f5fd6679\" (UID: \"6692f52c-5ee9-4cd6-8674-0fb3f5fd6679\") " Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.153393 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/6692f52c-5ee9-4cd6-8674-0fb3f5fd6679-rabbitmq-erlang-cookie\") pod \"6692f52c-5ee9-4cd6-8674-0fb3f5fd6679\" (UID: \"6692f52c-5ee9-4cd6-8674-0fb3f5fd6679\") " Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.153430 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/6692f52c-5ee9-4cd6-8674-0fb3f5fd6679-rabbitmq-plugins\") pod \"6692f52c-5ee9-4cd6-8674-0fb3f5fd6679\" (UID: \"6692f52c-5ee9-4cd6-8674-0fb3f5fd6679\") " Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.153503 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tlkrh\" (UniqueName: \"kubernetes.io/projected/6692f52c-5ee9-4cd6-8674-0fb3f5fd6679-kube-api-access-tlkrh\") pod \"6692f52c-5ee9-4cd6-8674-0fb3f5fd6679\" (UID: \"6692f52c-5ee9-4cd6-8674-0fb3f5fd6679\") " Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.153548 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/6692f52c-5ee9-4cd6-8674-0fb3f5fd6679-server-conf\") pod \"6692f52c-5ee9-4cd6-8674-0fb3f5fd6679\" (UID: \"6692f52c-5ee9-4cd6-8674-0fb3f5fd6679\") " Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.153600 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"6692f52c-5ee9-4cd6-8674-0fb3f5fd6679\" (UID: \"6692f52c-5ee9-4cd6-8674-0fb3f5fd6679\") " Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.153627 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/6692f52c-5ee9-4cd6-8674-0fb3f5fd6679-rabbitmq-tls\") pod \"6692f52c-5ee9-4cd6-8674-0fb3f5fd6679\" (UID: \"6692f52c-5ee9-4cd6-8674-0fb3f5fd6679\") " Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.153697 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/6692f52c-5ee9-4cd6-8674-0fb3f5fd6679-pod-info\") pod \"6692f52c-5ee9-4cd6-8674-0fb3f5fd6679\" (UID: \"6692f52c-5ee9-4cd6-8674-0fb3f5fd6679\") " Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.154193 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6692f52c-5ee9-4cd6-8674-0fb3f5fd6679-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "6692f52c-5ee9-4cd6-8674-0fb3f5fd6679" (UID: "6692f52c-5ee9-4cd6-8674-0fb3f5fd6679"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.160857 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6692f52c-5ee9-4cd6-8674-0fb3f5fd6679-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "6692f52c-5ee9-4cd6-8674-0fb3f5fd6679" (UID: "6692f52c-5ee9-4cd6-8674-0fb3f5fd6679"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.165248 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6692f52c-5ee9-4cd6-8674-0fb3f5fd6679-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "6692f52c-5ee9-4cd6-8674-0fb3f5fd6679" (UID: "6692f52c-5ee9-4cd6-8674-0fb3f5fd6679"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.170152 4885 generic.go:334] "Generic (PLEG): container finished" podID="6692f52c-5ee9-4cd6-8674-0fb3f5fd6679" containerID="18584fe73de75abdffba2d36424556d76845cfba20bc3152c3df90230f533362" exitCode=0 Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.170523 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"6692f52c-5ee9-4cd6-8674-0fb3f5fd6679","Type":"ContainerDied","Data":"18584fe73de75abdffba2d36424556d76845cfba20bc3152c3df90230f533362"} Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.170557 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"6692f52c-5ee9-4cd6-8674-0fb3f5fd6679","Type":"ContainerDied","Data":"212fc43e2e0d43911b7a078d43f599e155a6d9739da223da5377570980156bd5"} Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.170577 4885 scope.go:117] "RemoveContainer" containerID="18584fe73de75abdffba2d36424556d76845cfba20bc3152c3df90230f533362" Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.170743 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.171492 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6692f52c-5ee9-4cd6-8674-0fb3f5fd6679-kube-api-access-tlkrh" (OuterVolumeSpecName: "kube-api-access-tlkrh") pod "6692f52c-5ee9-4cd6-8674-0fb3f5fd6679" (UID: "6692f52c-5ee9-4cd6-8674-0fb3f5fd6679"). InnerVolumeSpecName "kube-api-access-tlkrh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.176373 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "persistence") pod "6692f52c-5ee9-4cd6-8674-0fb3f5fd6679" (UID: "6692f52c-5ee9-4cd6-8674-0fb3f5fd6679"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.183568 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/6692f52c-5ee9-4cd6-8674-0fb3f5fd6679-pod-info" (OuterVolumeSpecName: "pod-info") pod "6692f52c-5ee9-4cd6-8674-0fb3f5fd6679" (UID: "6692f52c-5ee9-4cd6-8674-0fb3f5fd6679"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.203033 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6692f52c-5ee9-4cd6-8674-0fb3f5fd6679-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "6692f52c-5ee9-4cd6-8674-0fb3f5fd6679" (UID: "6692f52c-5ee9-4cd6-8674-0fb3f5fd6679"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.240896 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6692f52c-5ee9-4cd6-8674-0fb3f5fd6679-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "6692f52c-5ee9-4cd6-8674-0fb3f5fd6679" (UID: "6692f52c-5ee9-4cd6-8674-0fb3f5fd6679"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.242707 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6692f52c-5ee9-4cd6-8674-0fb3f5fd6679-config-data" (OuterVolumeSpecName: "config-data") pod "6692f52c-5ee9-4cd6-8674-0fb3f5fd6679" (UID: "6692f52c-5ee9-4cd6-8674-0fb3f5fd6679"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.256156 4885 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/6692f52c-5ee9-4cd6-8674-0fb3f5fd6679-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.256182 4885 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6692f52c-5ee9-4cd6-8674-0fb3f5fd6679-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.256190 4885 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/6692f52c-5ee9-4cd6-8674-0fb3f5fd6679-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.256199 4885 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/6692f52c-5ee9-4cd6-8674-0fb3f5fd6679-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.256208 4885 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/6692f52c-5ee9-4cd6-8674-0fb3f5fd6679-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.256216 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tlkrh\" (UniqueName: \"kubernetes.io/projected/6692f52c-5ee9-4cd6-8674-0fb3f5fd6679-kube-api-access-tlkrh\") on node \"crc\" DevicePath \"\"" Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.256244 4885 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.256266 4885 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/6692f52c-5ee9-4cd6-8674-0fb3f5fd6679-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.256275 4885 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/6692f52c-5ee9-4cd6-8674-0fb3f5fd6679-pod-info\") on node \"crc\" DevicePath \"\"" Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.294301 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6692f52c-5ee9-4cd6-8674-0fb3f5fd6679-server-conf" (OuterVolumeSpecName: "server-conf") pod "6692f52c-5ee9-4cd6-8674-0fb3f5fd6679" (UID: "6692f52c-5ee9-4cd6-8674-0fb3f5fd6679"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.300664 4885 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.310691 4885 scope.go:117] "RemoveContainer" containerID="a52ec36756e9b2deb9a3362cfb43912af369dd3918986e7fd40070a1385f68f0" Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.311931 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6692f52c-5ee9-4cd6-8674-0fb3f5fd6679-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "6692f52c-5ee9-4cd6-8674-0fb3f5fd6679" (UID: "6692f52c-5ee9-4cd6-8674-0fb3f5fd6679"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.335527 4885 scope.go:117] "RemoveContainer" containerID="18584fe73de75abdffba2d36424556d76845cfba20bc3152c3df90230f533362" Oct 02 02:08:08 crc kubenswrapper[4885]: E1002 02:08:08.336534 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"18584fe73de75abdffba2d36424556d76845cfba20bc3152c3df90230f533362\": container with ID starting with 18584fe73de75abdffba2d36424556d76845cfba20bc3152c3df90230f533362 not found: ID does not exist" containerID="18584fe73de75abdffba2d36424556d76845cfba20bc3152c3df90230f533362" Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.336590 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"18584fe73de75abdffba2d36424556d76845cfba20bc3152c3df90230f533362"} err="failed to get container status \"18584fe73de75abdffba2d36424556d76845cfba20bc3152c3df90230f533362\": rpc error: code = NotFound desc = could not find container \"18584fe73de75abdffba2d36424556d76845cfba20bc3152c3df90230f533362\": container with ID starting with 18584fe73de75abdffba2d36424556d76845cfba20bc3152c3df90230f533362 not found: ID does not exist" Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.336623 4885 scope.go:117] "RemoveContainer" containerID="a52ec36756e9b2deb9a3362cfb43912af369dd3918986e7fd40070a1385f68f0" Oct 02 02:08:08 crc kubenswrapper[4885]: E1002 02:08:08.336922 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a52ec36756e9b2deb9a3362cfb43912af369dd3918986e7fd40070a1385f68f0\": container with ID starting with a52ec36756e9b2deb9a3362cfb43912af369dd3918986e7fd40070a1385f68f0 not found: ID does not exist" containerID="a52ec36756e9b2deb9a3362cfb43912af369dd3918986e7fd40070a1385f68f0" Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.336974 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a52ec36756e9b2deb9a3362cfb43912af369dd3918986e7fd40070a1385f68f0"} err="failed to get container status \"a52ec36756e9b2deb9a3362cfb43912af369dd3918986e7fd40070a1385f68f0\": rpc error: code = NotFound desc = could not find container \"a52ec36756e9b2deb9a3362cfb43912af369dd3918986e7fd40070a1385f68f0\": container with ID starting with a52ec36756e9b2deb9a3362cfb43912af369dd3918986e7fd40070a1385f68f0 not found: ID does not exist" Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.358124 4885 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/6692f52c-5ee9-4cd6-8674-0fb3f5fd6679-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.358161 4885 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/6692f52c-5ee9-4cd6-8674-0fb3f5fd6679-server-conf\") on node \"crc\" DevicePath \"\"" Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.358176 4885 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.501513 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.508003 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.529108 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 02 02:08:08 crc kubenswrapper[4885]: E1002 02:08:08.529488 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6692f52c-5ee9-4cd6-8674-0fb3f5fd6679" containerName="rabbitmq" Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.529506 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="6692f52c-5ee9-4cd6-8674-0fb3f5fd6679" containerName="rabbitmq" Oct 02 02:08:08 crc kubenswrapper[4885]: E1002 02:08:08.529547 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6692f52c-5ee9-4cd6-8674-0fb3f5fd6679" containerName="setup-container" Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.529554 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="6692f52c-5ee9-4cd6-8674-0fb3f5fd6679" containerName="setup-container" Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.529714 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="6692f52c-5ee9-4cd6-8674-0fb3f5fd6679" containerName="rabbitmq" Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.530722 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.535309 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.535475 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.535660 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.535698 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.535777 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.535665 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-2stht" Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.538877 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.540485 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.663843 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/18642896-a2b6-412c-beb1-8010484744a4-pod-info\") pod \"rabbitmq-server-0\" (UID: \"18642896-a2b6-412c-beb1-8010484744a4\") " pod="openstack/rabbitmq-server-0" Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.663898 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/18642896-a2b6-412c-beb1-8010484744a4-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"18642896-a2b6-412c-beb1-8010484744a4\") " pod="openstack/rabbitmq-server-0" Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.663931 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/18642896-a2b6-412c-beb1-8010484744a4-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"18642896-a2b6-412c-beb1-8010484744a4\") " pod="openstack/rabbitmq-server-0" Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.663971 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/18642896-a2b6-412c-beb1-8010484744a4-server-conf\") pod \"rabbitmq-server-0\" (UID: \"18642896-a2b6-412c-beb1-8010484744a4\") " pod="openstack/rabbitmq-server-0" Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.664007 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/18642896-a2b6-412c-beb1-8010484744a4-config-data\") pod \"rabbitmq-server-0\" (UID: \"18642896-a2b6-412c-beb1-8010484744a4\") " pod="openstack/rabbitmq-server-0" Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.664027 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"18642896-a2b6-412c-beb1-8010484744a4\") " pod="openstack/rabbitmq-server-0" Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.664078 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/18642896-a2b6-412c-beb1-8010484744a4-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"18642896-a2b6-412c-beb1-8010484744a4\") " pod="openstack/rabbitmq-server-0" Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.664112 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/18642896-a2b6-412c-beb1-8010484744a4-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"18642896-a2b6-412c-beb1-8010484744a4\") " pod="openstack/rabbitmq-server-0" Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.664135 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r2lvs\" (UniqueName: \"kubernetes.io/projected/18642896-a2b6-412c-beb1-8010484744a4-kube-api-access-r2lvs\") pod \"rabbitmq-server-0\" (UID: \"18642896-a2b6-412c-beb1-8010484744a4\") " pod="openstack/rabbitmq-server-0" Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.664160 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/18642896-a2b6-412c-beb1-8010484744a4-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"18642896-a2b6-412c-beb1-8010484744a4\") " pod="openstack/rabbitmq-server-0" Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.664214 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/18642896-a2b6-412c-beb1-8010484744a4-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"18642896-a2b6-412c-beb1-8010484744a4\") " pod="openstack/rabbitmq-server-0" Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.766681 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/18642896-a2b6-412c-beb1-8010484744a4-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"18642896-a2b6-412c-beb1-8010484744a4\") " pod="openstack/rabbitmq-server-0" Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.766745 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/18642896-a2b6-412c-beb1-8010484744a4-pod-info\") pod \"rabbitmq-server-0\" (UID: \"18642896-a2b6-412c-beb1-8010484744a4\") " pod="openstack/rabbitmq-server-0" Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.766774 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/18642896-a2b6-412c-beb1-8010484744a4-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"18642896-a2b6-412c-beb1-8010484744a4\") " pod="openstack/rabbitmq-server-0" Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.766795 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/18642896-a2b6-412c-beb1-8010484744a4-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"18642896-a2b6-412c-beb1-8010484744a4\") " pod="openstack/rabbitmq-server-0" Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.766825 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/18642896-a2b6-412c-beb1-8010484744a4-server-conf\") pod \"rabbitmq-server-0\" (UID: \"18642896-a2b6-412c-beb1-8010484744a4\") " pod="openstack/rabbitmq-server-0" Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.766853 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/18642896-a2b6-412c-beb1-8010484744a4-config-data\") pod \"rabbitmq-server-0\" (UID: \"18642896-a2b6-412c-beb1-8010484744a4\") " pod="openstack/rabbitmq-server-0" Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.766869 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"18642896-a2b6-412c-beb1-8010484744a4\") " pod="openstack/rabbitmq-server-0" Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.766905 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/18642896-a2b6-412c-beb1-8010484744a4-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"18642896-a2b6-412c-beb1-8010484744a4\") " pod="openstack/rabbitmq-server-0" Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.766930 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/18642896-a2b6-412c-beb1-8010484744a4-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"18642896-a2b6-412c-beb1-8010484744a4\") " pod="openstack/rabbitmq-server-0" Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.766951 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r2lvs\" (UniqueName: \"kubernetes.io/projected/18642896-a2b6-412c-beb1-8010484744a4-kube-api-access-r2lvs\") pod \"rabbitmq-server-0\" (UID: \"18642896-a2b6-412c-beb1-8010484744a4\") " pod="openstack/rabbitmq-server-0" Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.766971 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/18642896-a2b6-412c-beb1-8010484744a4-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"18642896-a2b6-412c-beb1-8010484744a4\") " pod="openstack/rabbitmq-server-0" Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.768219 4885 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"18642896-a2b6-412c-beb1-8010484744a4\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/rabbitmq-server-0" Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.768495 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/18642896-a2b6-412c-beb1-8010484744a4-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"18642896-a2b6-412c-beb1-8010484744a4\") " pod="openstack/rabbitmq-server-0" Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.768821 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/18642896-a2b6-412c-beb1-8010484744a4-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"18642896-a2b6-412c-beb1-8010484744a4\") " pod="openstack/rabbitmq-server-0" Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.769670 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/18642896-a2b6-412c-beb1-8010484744a4-server-conf\") pod \"rabbitmq-server-0\" (UID: \"18642896-a2b6-412c-beb1-8010484744a4\") " pod="openstack/rabbitmq-server-0" Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.771882 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/18642896-a2b6-412c-beb1-8010484744a4-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"18642896-a2b6-412c-beb1-8010484744a4\") " pod="openstack/rabbitmq-server-0" Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.772060 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/18642896-a2b6-412c-beb1-8010484744a4-config-data\") pod \"rabbitmq-server-0\" (UID: \"18642896-a2b6-412c-beb1-8010484744a4\") " pod="openstack/rabbitmq-server-0" Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.772385 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/18642896-a2b6-412c-beb1-8010484744a4-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"18642896-a2b6-412c-beb1-8010484744a4\") " pod="openstack/rabbitmq-server-0" Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.772472 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/18642896-a2b6-412c-beb1-8010484744a4-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"18642896-a2b6-412c-beb1-8010484744a4\") " pod="openstack/rabbitmq-server-0" Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.772686 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/18642896-a2b6-412c-beb1-8010484744a4-pod-info\") pod \"rabbitmq-server-0\" (UID: \"18642896-a2b6-412c-beb1-8010484744a4\") " pod="openstack/rabbitmq-server-0" Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.776894 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/18642896-a2b6-412c-beb1-8010484744a4-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"18642896-a2b6-412c-beb1-8010484744a4\") " pod="openstack/rabbitmq-server-0" Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.785400 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r2lvs\" (UniqueName: \"kubernetes.io/projected/18642896-a2b6-412c-beb1-8010484744a4-kube-api-access-r2lvs\") pod \"rabbitmq-server-0\" (UID: \"18642896-a2b6-412c-beb1-8010484744a4\") " pod="openstack/rabbitmq-server-0" Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.816459 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"18642896-a2b6-412c-beb1-8010484744a4\") " pod="openstack/rabbitmq-server-0" Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.855914 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.916872 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-tbrnt"] Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.918935 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67b789f86c-tbrnt" Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.924544 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.931155 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-tbrnt"] Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.971028 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b0452212-e578-4bf2-9199-91eefc38a3f2-dns-svc\") pod \"dnsmasq-dns-67b789f86c-tbrnt\" (UID: \"b0452212-e578-4bf2-9199-91eefc38a3f2\") " pod="openstack/dnsmasq-dns-67b789f86c-tbrnt" Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.971069 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-62d5d\" (UniqueName: \"kubernetes.io/projected/b0452212-e578-4bf2-9199-91eefc38a3f2-kube-api-access-62d5d\") pod \"dnsmasq-dns-67b789f86c-tbrnt\" (UID: \"b0452212-e578-4bf2-9199-91eefc38a3f2\") " pod="openstack/dnsmasq-dns-67b789f86c-tbrnt" Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.971113 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/b0452212-e578-4bf2-9199-91eefc38a3f2-openstack-edpm-ipam\") pod \"dnsmasq-dns-67b789f86c-tbrnt\" (UID: \"b0452212-e578-4bf2-9199-91eefc38a3f2\") " pod="openstack/dnsmasq-dns-67b789f86c-tbrnt" Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.971152 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b0452212-e578-4bf2-9199-91eefc38a3f2-dns-swift-storage-0\") pod \"dnsmasq-dns-67b789f86c-tbrnt\" (UID: \"b0452212-e578-4bf2-9199-91eefc38a3f2\") " pod="openstack/dnsmasq-dns-67b789f86c-tbrnt" Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.971176 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b0452212-e578-4bf2-9199-91eefc38a3f2-ovsdbserver-nb\") pod \"dnsmasq-dns-67b789f86c-tbrnt\" (UID: \"b0452212-e578-4bf2-9199-91eefc38a3f2\") " pod="openstack/dnsmasq-dns-67b789f86c-tbrnt" Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.971198 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b0452212-e578-4bf2-9199-91eefc38a3f2-ovsdbserver-sb\") pod \"dnsmasq-dns-67b789f86c-tbrnt\" (UID: \"b0452212-e578-4bf2-9199-91eefc38a3f2\") " pod="openstack/dnsmasq-dns-67b789f86c-tbrnt" Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.971219 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b0452212-e578-4bf2-9199-91eefc38a3f2-config\") pod \"dnsmasq-dns-67b789f86c-tbrnt\" (UID: \"b0452212-e578-4bf2-9199-91eefc38a3f2\") " pod="openstack/dnsmasq-dns-67b789f86c-tbrnt" Oct 02 02:08:08 crc kubenswrapper[4885]: I1002 02:08:08.976895 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-tbrnt"] Oct 02 02:08:08 crc kubenswrapper[4885]: E1002 02:08:08.988455 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[config dns-svc dns-swift-storage-0 kube-api-access-62d5d openstack-edpm-ipam ovsdbserver-nb ovsdbserver-sb], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/dnsmasq-dns-67b789f86c-tbrnt" podUID="b0452212-e578-4bf2-9199-91eefc38a3f2" Oct 02 02:08:09 crc kubenswrapper[4885]: I1002 02:08:09.029669 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-cb6ffcf87-46bgc"] Oct 02 02:08:09 crc kubenswrapper[4885]: I1002 02:08:09.031292 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cb6ffcf87-46bgc" Oct 02 02:08:09 crc kubenswrapper[4885]: I1002 02:08:09.050857 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cb6ffcf87-46bgc"] Oct 02 02:08:09 crc kubenswrapper[4885]: I1002 02:08:09.072228 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/b0452212-e578-4bf2-9199-91eefc38a3f2-openstack-edpm-ipam\") pod \"dnsmasq-dns-67b789f86c-tbrnt\" (UID: \"b0452212-e578-4bf2-9199-91eefc38a3f2\") " pod="openstack/dnsmasq-dns-67b789f86c-tbrnt" Oct 02 02:08:09 crc kubenswrapper[4885]: I1002 02:08:09.072316 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b0452212-e578-4bf2-9199-91eefc38a3f2-dns-swift-storage-0\") pod \"dnsmasq-dns-67b789f86c-tbrnt\" (UID: \"b0452212-e578-4bf2-9199-91eefc38a3f2\") " pod="openstack/dnsmasq-dns-67b789f86c-tbrnt" Oct 02 02:08:09 crc kubenswrapper[4885]: I1002 02:08:09.072345 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b0452212-e578-4bf2-9199-91eefc38a3f2-ovsdbserver-nb\") pod \"dnsmasq-dns-67b789f86c-tbrnt\" (UID: \"b0452212-e578-4bf2-9199-91eefc38a3f2\") " pod="openstack/dnsmasq-dns-67b789f86c-tbrnt" Oct 02 02:08:09 crc kubenswrapper[4885]: I1002 02:08:09.072369 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b0452212-e578-4bf2-9199-91eefc38a3f2-ovsdbserver-sb\") pod \"dnsmasq-dns-67b789f86c-tbrnt\" (UID: \"b0452212-e578-4bf2-9199-91eefc38a3f2\") " pod="openstack/dnsmasq-dns-67b789f86c-tbrnt" Oct 02 02:08:09 crc kubenswrapper[4885]: I1002 02:08:09.072387 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b0452212-e578-4bf2-9199-91eefc38a3f2-config\") pod \"dnsmasq-dns-67b789f86c-tbrnt\" (UID: \"b0452212-e578-4bf2-9199-91eefc38a3f2\") " pod="openstack/dnsmasq-dns-67b789f86c-tbrnt" Oct 02 02:08:09 crc kubenswrapper[4885]: I1002 02:08:09.072470 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b0452212-e578-4bf2-9199-91eefc38a3f2-dns-svc\") pod \"dnsmasq-dns-67b789f86c-tbrnt\" (UID: \"b0452212-e578-4bf2-9199-91eefc38a3f2\") " pod="openstack/dnsmasq-dns-67b789f86c-tbrnt" Oct 02 02:08:09 crc kubenswrapper[4885]: I1002 02:08:09.072492 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-62d5d\" (UniqueName: \"kubernetes.io/projected/b0452212-e578-4bf2-9199-91eefc38a3f2-kube-api-access-62d5d\") pod \"dnsmasq-dns-67b789f86c-tbrnt\" (UID: \"b0452212-e578-4bf2-9199-91eefc38a3f2\") " pod="openstack/dnsmasq-dns-67b789f86c-tbrnt" Oct 02 02:08:09 crc kubenswrapper[4885]: I1002 02:08:09.073529 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/b0452212-e578-4bf2-9199-91eefc38a3f2-openstack-edpm-ipam\") pod \"dnsmasq-dns-67b789f86c-tbrnt\" (UID: \"b0452212-e578-4bf2-9199-91eefc38a3f2\") " pod="openstack/dnsmasq-dns-67b789f86c-tbrnt" Oct 02 02:08:09 crc kubenswrapper[4885]: I1002 02:08:09.074025 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b0452212-e578-4bf2-9199-91eefc38a3f2-dns-swift-storage-0\") pod \"dnsmasq-dns-67b789f86c-tbrnt\" (UID: \"b0452212-e578-4bf2-9199-91eefc38a3f2\") " pod="openstack/dnsmasq-dns-67b789f86c-tbrnt" Oct 02 02:08:09 crc kubenswrapper[4885]: I1002 02:08:09.074576 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b0452212-e578-4bf2-9199-91eefc38a3f2-config\") pod \"dnsmasq-dns-67b789f86c-tbrnt\" (UID: \"b0452212-e578-4bf2-9199-91eefc38a3f2\") " pod="openstack/dnsmasq-dns-67b789f86c-tbrnt" Oct 02 02:08:09 crc kubenswrapper[4885]: I1002 02:08:09.075088 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b0452212-e578-4bf2-9199-91eefc38a3f2-ovsdbserver-sb\") pod \"dnsmasq-dns-67b789f86c-tbrnt\" (UID: \"b0452212-e578-4bf2-9199-91eefc38a3f2\") " pod="openstack/dnsmasq-dns-67b789f86c-tbrnt" Oct 02 02:08:09 crc kubenswrapper[4885]: I1002 02:08:09.075282 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b0452212-e578-4bf2-9199-91eefc38a3f2-ovsdbserver-nb\") pod \"dnsmasq-dns-67b789f86c-tbrnt\" (UID: \"b0452212-e578-4bf2-9199-91eefc38a3f2\") " pod="openstack/dnsmasq-dns-67b789f86c-tbrnt" Oct 02 02:08:09 crc kubenswrapper[4885]: I1002 02:08:09.075568 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b0452212-e578-4bf2-9199-91eefc38a3f2-dns-svc\") pod \"dnsmasq-dns-67b789f86c-tbrnt\" (UID: \"b0452212-e578-4bf2-9199-91eefc38a3f2\") " pod="openstack/dnsmasq-dns-67b789f86c-tbrnt" Oct 02 02:08:09 crc kubenswrapper[4885]: I1002 02:08:09.090594 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-62d5d\" (UniqueName: \"kubernetes.io/projected/b0452212-e578-4bf2-9199-91eefc38a3f2-kube-api-access-62d5d\") pod \"dnsmasq-dns-67b789f86c-tbrnt\" (UID: \"b0452212-e578-4bf2-9199-91eefc38a3f2\") " pod="openstack/dnsmasq-dns-67b789f86c-tbrnt" Oct 02 02:08:09 crc kubenswrapper[4885]: I1002 02:08:09.174907 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/58f621da-0934-482b-8f10-4273021207f0-openstack-edpm-ipam\") pod \"dnsmasq-dns-cb6ffcf87-46bgc\" (UID: \"58f621da-0934-482b-8f10-4273021207f0\") " pod="openstack/dnsmasq-dns-cb6ffcf87-46bgc" Oct 02 02:08:09 crc kubenswrapper[4885]: I1002 02:08:09.174960 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kldg2\" (UniqueName: \"kubernetes.io/projected/58f621da-0934-482b-8f10-4273021207f0-kube-api-access-kldg2\") pod \"dnsmasq-dns-cb6ffcf87-46bgc\" (UID: \"58f621da-0934-482b-8f10-4273021207f0\") " pod="openstack/dnsmasq-dns-cb6ffcf87-46bgc" Oct 02 02:08:09 crc kubenswrapper[4885]: I1002 02:08:09.174993 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/58f621da-0934-482b-8f10-4273021207f0-ovsdbserver-sb\") pod \"dnsmasq-dns-cb6ffcf87-46bgc\" (UID: \"58f621da-0934-482b-8f10-4273021207f0\") " pod="openstack/dnsmasq-dns-cb6ffcf87-46bgc" Oct 02 02:08:09 crc kubenswrapper[4885]: I1002 02:08:09.181103 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/58f621da-0934-482b-8f10-4273021207f0-dns-svc\") pod \"dnsmasq-dns-cb6ffcf87-46bgc\" (UID: \"58f621da-0934-482b-8f10-4273021207f0\") " pod="openstack/dnsmasq-dns-cb6ffcf87-46bgc" Oct 02 02:08:09 crc kubenswrapper[4885]: I1002 02:08:09.181142 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/58f621da-0934-482b-8f10-4273021207f0-ovsdbserver-nb\") pod \"dnsmasq-dns-cb6ffcf87-46bgc\" (UID: \"58f621da-0934-482b-8f10-4273021207f0\") " pod="openstack/dnsmasq-dns-cb6ffcf87-46bgc" Oct 02 02:08:09 crc kubenswrapper[4885]: I1002 02:08:09.181162 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/58f621da-0934-482b-8f10-4273021207f0-dns-swift-storage-0\") pod \"dnsmasq-dns-cb6ffcf87-46bgc\" (UID: \"58f621da-0934-482b-8f10-4273021207f0\") " pod="openstack/dnsmasq-dns-cb6ffcf87-46bgc" Oct 02 02:08:09 crc kubenswrapper[4885]: I1002 02:08:09.181207 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/58f621da-0934-482b-8f10-4273021207f0-config\") pod \"dnsmasq-dns-cb6ffcf87-46bgc\" (UID: \"58f621da-0934-482b-8f10-4273021207f0\") " pod="openstack/dnsmasq-dns-cb6ffcf87-46bgc" Oct 02 02:08:09 crc kubenswrapper[4885]: I1002 02:08:09.190201 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67b789f86c-tbrnt" Oct 02 02:08:09 crc kubenswrapper[4885]: I1002 02:08:09.203315 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67b789f86c-tbrnt" Oct 02 02:08:09 crc kubenswrapper[4885]: I1002 02:08:09.282910 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/b0452212-e578-4bf2-9199-91eefc38a3f2-openstack-edpm-ipam\") pod \"b0452212-e578-4bf2-9199-91eefc38a3f2\" (UID: \"b0452212-e578-4bf2-9199-91eefc38a3f2\") " Oct 02 02:08:09 crc kubenswrapper[4885]: I1002 02:08:09.283086 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b0452212-e578-4bf2-9199-91eefc38a3f2-dns-swift-storage-0\") pod \"b0452212-e578-4bf2-9199-91eefc38a3f2\" (UID: \"b0452212-e578-4bf2-9199-91eefc38a3f2\") " Oct 02 02:08:09 crc kubenswrapper[4885]: I1002 02:08:09.283132 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b0452212-e578-4bf2-9199-91eefc38a3f2-config\") pod \"b0452212-e578-4bf2-9199-91eefc38a3f2\" (UID: \"b0452212-e578-4bf2-9199-91eefc38a3f2\") " Oct 02 02:08:09 crc kubenswrapper[4885]: I1002 02:08:09.283159 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-62d5d\" (UniqueName: \"kubernetes.io/projected/b0452212-e578-4bf2-9199-91eefc38a3f2-kube-api-access-62d5d\") pod \"b0452212-e578-4bf2-9199-91eefc38a3f2\" (UID: \"b0452212-e578-4bf2-9199-91eefc38a3f2\") " Oct 02 02:08:09 crc kubenswrapper[4885]: I1002 02:08:09.283195 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b0452212-e578-4bf2-9199-91eefc38a3f2-ovsdbserver-sb\") pod \"b0452212-e578-4bf2-9199-91eefc38a3f2\" (UID: \"b0452212-e578-4bf2-9199-91eefc38a3f2\") " Oct 02 02:08:09 crc kubenswrapper[4885]: I1002 02:08:09.283412 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b0452212-e578-4bf2-9199-91eefc38a3f2-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "b0452212-e578-4bf2-9199-91eefc38a3f2" (UID: "b0452212-e578-4bf2-9199-91eefc38a3f2"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:08:09 crc kubenswrapper[4885]: I1002 02:08:09.283668 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b0452212-e578-4bf2-9199-91eefc38a3f2-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "b0452212-e578-4bf2-9199-91eefc38a3f2" (UID: "b0452212-e578-4bf2-9199-91eefc38a3f2"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:08:09 crc kubenswrapper[4885]: I1002 02:08:09.283773 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b0452212-e578-4bf2-9199-91eefc38a3f2-dns-svc\") pod \"b0452212-e578-4bf2-9199-91eefc38a3f2\" (UID: \"b0452212-e578-4bf2-9199-91eefc38a3f2\") " Oct 02 02:08:09 crc kubenswrapper[4885]: I1002 02:08:09.283826 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b0452212-e578-4bf2-9199-91eefc38a3f2-ovsdbserver-nb\") pod \"b0452212-e578-4bf2-9199-91eefc38a3f2\" (UID: \"b0452212-e578-4bf2-9199-91eefc38a3f2\") " Oct 02 02:08:09 crc kubenswrapper[4885]: I1002 02:08:09.283833 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b0452212-e578-4bf2-9199-91eefc38a3f2-config" (OuterVolumeSpecName: "config") pod "b0452212-e578-4bf2-9199-91eefc38a3f2" (UID: "b0452212-e578-4bf2-9199-91eefc38a3f2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:08:09 crc kubenswrapper[4885]: I1002 02:08:09.284038 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b0452212-e578-4bf2-9199-91eefc38a3f2-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b0452212-e578-4bf2-9199-91eefc38a3f2" (UID: "b0452212-e578-4bf2-9199-91eefc38a3f2"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:08:09 crc kubenswrapper[4885]: I1002 02:08:09.284119 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b0452212-e578-4bf2-9199-91eefc38a3f2-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b0452212-e578-4bf2-9199-91eefc38a3f2" (UID: "b0452212-e578-4bf2-9199-91eefc38a3f2"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:08:09 crc kubenswrapper[4885]: I1002 02:08:09.284186 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b0452212-e578-4bf2-9199-91eefc38a3f2-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b0452212-e578-4bf2-9199-91eefc38a3f2" (UID: "b0452212-e578-4bf2-9199-91eefc38a3f2"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:08:09 crc kubenswrapper[4885]: I1002 02:08:09.284469 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/58f621da-0934-482b-8f10-4273021207f0-openstack-edpm-ipam\") pod \"dnsmasq-dns-cb6ffcf87-46bgc\" (UID: \"58f621da-0934-482b-8f10-4273021207f0\") " pod="openstack/dnsmasq-dns-cb6ffcf87-46bgc" Oct 02 02:08:09 crc kubenswrapper[4885]: I1002 02:08:09.284601 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kldg2\" (UniqueName: \"kubernetes.io/projected/58f621da-0934-482b-8f10-4273021207f0-kube-api-access-kldg2\") pod \"dnsmasq-dns-cb6ffcf87-46bgc\" (UID: \"58f621da-0934-482b-8f10-4273021207f0\") " pod="openstack/dnsmasq-dns-cb6ffcf87-46bgc" Oct 02 02:08:09 crc kubenswrapper[4885]: I1002 02:08:09.284726 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/58f621da-0934-482b-8f10-4273021207f0-ovsdbserver-sb\") pod \"dnsmasq-dns-cb6ffcf87-46bgc\" (UID: \"58f621da-0934-482b-8f10-4273021207f0\") " pod="openstack/dnsmasq-dns-cb6ffcf87-46bgc" Oct 02 02:08:09 crc kubenswrapper[4885]: I1002 02:08:09.284842 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/58f621da-0934-482b-8f10-4273021207f0-dns-svc\") pod \"dnsmasq-dns-cb6ffcf87-46bgc\" (UID: \"58f621da-0934-482b-8f10-4273021207f0\") " pod="openstack/dnsmasq-dns-cb6ffcf87-46bgc" Oct 02 02:08:09 crc kubenswrapper[4885]: I1002 02:08:09.284882 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/58f621da-0934-482b-8f10-4273021207f0-ovsdbserver-nb\") pod \"dnsmasq-dns-cb6ffcf87-46bgc\" (UID: \"58f621da-0934-482b-8f10-4273021207f0\") " pod="openstack/dnsmasq-dns-cb6ffcf87-46bgc" Oct 02 02:08:09 crc kubenswrapper[4885]: I1002 02:08:09.284918 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/58f621da-0934-482b-8f10-4273021207f0-dns-swift-storage-0\") pod \"dnsmasq-dns-cb6ffcf87-46bgc\" (UID: \"58f621da-0934-482b-8f10-4273021207f0\") " pod="openstack/dnsmasq-dns-cb6ffcf87-46bgc" Oct 02 02:08:09 crc kubenswrapper[4885]: I1002 02:08:09.284981 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/58f621da-0934-482b-8f10-4273021207f0-config\") pod \"dnsmasq-dns-cb6ffcf87-46bgc\" (UID: \"58f621da-0934-482b-8f10-4273021207f0\") " pod="openstack/dnsmasq-dns-cb6ffcf87-46bgc" Oct 02 02:08:09 crc kubenswrapper[4885]: I1002 02:08:09.285118 4885 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b0452212-e578-4bf2-9199-91eefc38a3f2-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 02:08:09 crc kubenswrapper[4885]: I1002 02:08:09.285141 4885 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b0452212-e578-4bf2-9199-91eefc38a3f2-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 02:08:09 crc kubenswrapper[4885]: I1002 02:08:09.285161 4885 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b0452212-e578-4bf2-9199-91eefc38a3f2-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 02:08:09 crc kubenswrapper[4885]: I1002 02:08:09.285178 4885 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/b0452212-e578-4bf2-9199-91eefc38a3f2-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Oct 02 02:08:09 crc kubenswrapper[4885]: I1002 02:08:09.285196 4885 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b0452212-e578-4bf2-9199-91eefc38a3f2-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 02 02:08:09 crc kubenswrapper[4885]: I1002 02:08:09.285214 4885 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b0452212-e578-4bf2-9199-91eefc38a3f2-config\") on node \"crc\" DevicePath \"\"" Oct 02 02:08:09 crc kubenswrapper[4885]: I1002 02:08:09.285460 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/58f621da-0934-482b-8f10-4273021207f0-ovsdbserver-sb\") pod \"dnsmasq-dns-cb6ffcf87-46bgc\" (UID: \"58f621da-0934-482b-8f10-4273021207f0\") " pod="openstack/dnsmasq-dns-cb6ffcf87-46bgc" Oct 02 02:08:09 crc kubenswrapper[4885]: I1002 02:08:09.285561 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/58f621da-0934-482b-8f10-4273021207f0-openstack-edpm-ipam\") pod \"dnsmasq-dns-cb6ffcf87-46bgc\" (UID: \"58f621da-0934-482b-8f10-4273021207f0\") " pod="openstack/dnsmasq-dns-cb6ffcf87-46bgc" Oct 02 02:08:09 crc kubenswrapper[4885]: I1002 02:08:09.285655 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/58f621da-0934-482b-8f10-4273021207f0-dns-svc\") pod \"dnsmasq-dns-cb6ffcf87-46bgc\" (UID: \"58f621da-0934-482b-8f10-4273021207f0\") " pod="openstack/dnsmasq-dns-cb6ffcf87-46bgc" Oct 02 02:08:09 crc kubenswrapper[4885]: I1002 02:08:09.286069 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/58f621da-0934-482b-8f10-4273021207f0-ovsdbserver-nb\") pod \"dnsmasq-dns-cb6ffcf87-46bgc\" (UID: \"58f621da-0934-482b-8f10-4273021207f0\") " pod="openstack/dnsmasq-dns-cb6ffcf87-46bgc" Oct 02 02:08:09 crc kubenswrapper[4885]: I1002 02:08:09.286483 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/58f621da-0934-482b-8f10-4273021207f0-config\") pod \"dnsmasq-dns-cb6ffcf87-46bgc\" (UID: \"58f621da-0934-482b-8f10-4273021207f0\") " pod="openstack/dnsmasq-dns-cb6ffcf87-46bgc" Oct 02 02:08:09 crc kubenswrapper[4885]: I1002 02:08:09.287981 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/58f621da-0934-482b-8f10-4273021207f0-dns-swift-storage-0\") pod \"dnsmasq-dns-cb6ffcf87-46bgc\" (UID: \"58f621da-0934-482b-8f10-4273021207f0\") " pod="openstack/dnsmasq-dns-cb6ffcf87-46bgc" Oct 02 02:08:09 crc kubenswrapper[4885]: I1002 02:08:09.289125 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b0452212-e578-4bf2-9199-91eefc38a3f2-kube-api-access-62d5d" (OuterVolumeSpecName: "kube-api-access-62d5d") pod "b0452212-e578-4bf2-9199-91eefc38a3f2" (UID: "b0452212-e578-4bf2-9199-91eefc38a3f2"). InnerVolumeSpecName "kube-api-access-62d5d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:08:09 crc kubenswrapper[4885]: I1002 02:08:09.306900 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kldg2\" (UniqueName: \"kubernetes.io/projected/58f621da-0934-482b-8f10-4273021207f0-kube-api-access-kldg2\") pod \"dnsmasq-dns-cb6ffcf87-46bgc\" (UID: \"58f621da-0934-482b-8f10-4273021207f0\") " pod="openstack/dnsmasq-dns-cb6ffcf87-46bgc" Oct 02 02:08:09 crc kubenswrapper[4885]: I1002 02:08:09.362093 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cb6ffcf87-46bgc" Oct 02 02:08:09 crc kubenswrapper[4885]: I1002 02:08:09.383811 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 02 02:08:09 crc kubenswrapper[4885]: I1002 02:08:09.386574 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-62d5d\" (UniqueName: \"kubernetes.io/projected/b0452212-e578-4bf2-9199-91eefc38a3f2-kube-api-access-62d5d\") on node \"crc\" DevicePath \"\"" Oct 02 02:08:09 crc kubenswrapper[4885]: I1002 02:08:09.764214 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:08:09 crc kubenswrapper[4885]: I1002 02:08:09.896059 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/eea66006-67ba-45e3-bdf9-9d144d772386-rabbitmq-erlang-cookie\") pod \"eea66006-67ba-45e3-bdf9-9d144d772386\" (UID: \"eea66006-67ba-45e3-bdf9-9d144d772386\") " Oct 02 02:08:09 crc kubenswrapper[4885]: I1002 02:08:09.896148 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/eea66006-67ba-45e3-bdf9-9d144d772386-plugins-conf\") pod \"eea66006-67ba-45e3-bdf9-9d144d772386\" (UID: \"eea66006-67ba-45e3-bdf9-9d144d772386\") " Oct 02 02:08:09 crc kubenswrapper[4885]: I1002 02:08:09.896182 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/eea66006-67ba-45e3-bdf9-9d144d772386-pod-info\") pod \"eea66006-67ba-45e3-bdf9-9d144d772386\" (UID: \"eea66006-67ba-45e3-bdf9-9d144d772386\") " Oct 02 02:08:09 crc kubenswrapper[4885]: I1002 02:08:09.896223 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/eea66006-67ba-45e3-bdf9-9d144d772386-rabbitmq-tls\") pod \"eea66006-67ba-45e3-bdf9-9d144d772386\" (UID: \"eea66006-67ba-45e3-bdf9-9d144d772386\") " Oct 02 02:08:09 crc kubenswrapper[4885]: I1002 02:08:09.896283 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"eea66006-67ba-45e3-bdf9-9d144d772386\" (UID: \"eea66006-67ba-45e3-bdf9-9d144d772386\") " Oct 02 02:08:09 crc kubenswrapper[4885]: I1002 02:08:09.897014 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/eea66006-67ba-45e3-bdf9-9d144d772386-config-data\") pod \"eea66006-67ba-45e3-bdf9-9d144d772386\" (UID: \"eea66006-67ba-45e3-bdf9-9d144d772386\") " Oct 02 02:08:09 crc kubenswrapper[4885]: I1002 02:08:09.897080 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/eea66006-67ba-45e3-bdf9-9d144d772386-server-conf\") pod \"eea66006-67ba-45e3-bdf9-9d144d772386\" (UID: \"eea66006-67ba-45e3-bdf9-9d144d772386\") " Oct 02 02:08:09 crc kubenswrapper[4885]: I1002 02:08:09.898805 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eea66006-67ba-45e3-bdf9-9d144d772386-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "eea66006-67ba-45e3-bdf9-9d144d772386" (UID: "eea66006-67ba-45e3-bdf9-9d144d772386"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:08:09 crc kubenswrapper[4885]: I1002 02:08:09.899320 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/eea66006-67ba-45e3-bdf9-9d144d772386-erlang-cookie-secret\") pod \"eea66006-67ba-45e3-bdf9-9d144d772386\" (UID: \"eea66006-67ba-45e3-bdf9-9d144d772386\") " Oct 02 02:08:09 crc kubenswrapper[4885]: I1002 02:08:09.899407 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/eea66006-67ba-45e3-bdf9-9d144d772386-rabbitmq-confd\") pod \"eea66006-67ba-45e3-bdf9-9d144d772386\" (UID: \"eea66006-67ba-45e3-bdf9-9d144d772386\") " Oct 02 02:08:09 crc kubenswrapper[4885]: I1002 02:08:09.899444 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/eea66006-67ba-45e3-bdf9-9d144d772386-rabbitmq-plugins\") pod \"eea66006-67ba-45e3-bdf9-9d144d772386\" (UID: \"eea66006-67ba-45e3-bdf9-9d144d772386\") " Oct 02 02:08:09 crc kubenswrapper[4885]: I1002 02:08:09.899499 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9c4hd\" (UniqueName: \"kubernetes.io/projected/eea66006-67ba-45e3-bdf9-9d144d772386-kube-api-access-9c4hd\") pod \"eea66006-67ba-45e3-bdf9-9d144d772386\" (UID: \"eea66006-67ba-45e3-bdf9-9d144d772386\") " Oct 02 02:08:09 crc kubenswrapper[4885]: I1002 02:08:09.899501 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eea66006-67ba-45e3-bdf9-9d144d772386-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "eea66006-67ba-45e3-bdf9-9d144d772386" (UID: "eea66006-67ba-45e3-bdf9-9d144d772386"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:08:09 crc kubenswrapper[4885]: I1002 02:08:09.900106 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eea66006-67ba-45e3-bdf9-9d144d772386-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "eea66006-67ba-45e3-bdf9-9d144d772386" (UID: "eea66006-67ba-45e3-bdf9-9d144d772386"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:08:09 crc kubenswrapper[4885]: I1002 02:08:09.900829 4885 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/eea66006-67ba-45e3-bdf9-9d144d772386-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 02 02:08:09 crc kubenswrapper[4885]: I1002 02:08:09.900851 4885 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/eea66006-67ba-45e3-bdf9-9d144d772386-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 02 02:08:09 crc kubenswrapper[4885]: I1002 02:08:09.900864 4885 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/eea66006-67ba-45e3-bdf9-9d144d772386-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 02 02:08:09 crc kubenswrapper[4885]: I1002 02:08:09.907775 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/eea66006-67ba-45e3-bdf9-9d144d772386-pod-info" (OuterVolumeSpecName: "pod-info") pod "eea66006-67ba-45e3-bdf9-9d144d772386" (UID: "eea66006-67ba-45e3-bdf9-9d144d772386"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 02 02:08:09 crc kubenswrapper[4885]: I1002 02:08:09.914435 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eea66006-67ba-45e3-bdf9-9d144d772386-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "eea66006-67ba-45e3-bdf9-9d144d772386" (UID: "eea66006-67ba-45e3-bdf9-9d144d772386"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:08:09 crc kubenswrapper[4885]: I1002 02:08:09.914452 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eea66006-67ba-45e3-bdf9-9d144d772386-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "eea66006-67ba-45e3-bdf9-9d144d772386" (UID: "eea66006-67ba-45e3-bdf9-9d144d772386"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:08:09 crc kubenswrapper[4885]: I1002 02:08:09.918092 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "persistence") pod "eea66006-67ba-45e3-bdf9-9d144d772386" (UID: "eea66006-67ba-45e3-bdf9-9d144d772386"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 02 02:08:09 crc kubenswrapper[4885]: I1002 02:08:09.918212 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eea66006-67ba-45e3-bdf9-9d144d772386-kube-api-access-9c4hd" (OuterVolumeSpecName: "kube-api-access-9c4hd") pod "eea66006-67ba-45e3-bdf9-9d144d772386" (UID: "eea66006-67ba-45e3-bdf9-9d144d772386"). InnerVolumeSpecName "kube-api-access-9c4hd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:08:09 crc kubenswrapper[4885]: I1002 02:08:09.923618 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cb6ffcf87-46bgc"] Oct 02 02:08:09 crc kubenswrapper[4885]: I1002 02:08:09.954215 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eea66006-67ba-45e3-bdf9-9d144d772386-config-data" (OuterVolumeSpecName: "config-data") pod "eea66006-67ba-45e3-bdf9-9d144d772386" (UID: "eea66006-67ba-45e3-bdf9-9d144d772386"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:08:09 crc kubenswrapper[4885]: I1002 02:08:09.964410 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eea66006-67ba-45e3-bdf9-9d144d772386-server-conf" (OuterVolumeSpecName: "server-conf") pod "eea66006-67ba-45e3-bdf9-9d144d772386" (UID: "eea66006-67ba-45e3-bdf9-9d144d772386"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:08:10 crc kubenswrapper[4885]: I1002 02:08:10.009388 4885 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/eea66006-67ba-45e3-bdf9-9d144d772386-pod-info\") on node \"crc\" DevicePath \"\"" Oct 02 02:08:10 crc kubenswrapper[4885]: I1002 02:08:10.009424 4885 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/eea66006-67ba-45e3-bdf9-9d144d772386-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 02 02:08:10 crc kubenswrapper[4885]: I1002 02:08:10.009458 4885 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Oct 02 02:08:10 crc kubenswrapper[4885]: I1002 02:08:10.009468 4885 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/eea66006-67ba-45e3-bdf9-9d144d772386-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 02:08:10 crc kubenswrapper[4885]: I1002 02:08:10.009477 4885 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/eea66006-67ba-45e3-bdf9-9d144d772386-server-conf\") on node \"crc\" DevicePath \"\"" Oct 02 02:08:10 crc kubenswrapper[4885]: I1002 02:08:10.009485 4885 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/eea66006-67ba-45e3-bdf9-9d144d772386-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 02 02:08:10 crc kubenswrapper[4885]: I1002 02:08:10.009494 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9c4hd\" (UniqueName: \"kubernetes.io/projected/eea66006-67ba-45e3-bdf9-9d144d772386-kube-api-access-9c4hd\") on node \"crc\" DevicePath \"\"" Oct 02 02:08:10 crc kubenswrapper[4885]: I1002 02:08:10.033825 4885 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Oct 02 02:08:10 crc kubenswrapper[4885]: I1002 02:08:10.037318 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eea66006-67ba-45e3-bdf9-9d144d772386-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "eea66006-67ba-45e3-bdf9-9d144d772386" (UID: "eea66006-67ba-45e3-bdf9-9d144d772386"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:08:10 crc kubenswrapper[4885]: I1002 02:08:10.056639 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6692f52c-5ee9-4cd6-8674-0fb3f5fd6679" path="/var/lib/kubelet/pods/6692f52c-5ee9-4cd6-8674-0fb3f5fd6679/volumes" Oct 02 02:08:10 crc kubenswrapper[4885]: I1002 02:08:10.111183 4885 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Oct 02 02:08:10 crc kubenswrapper[4885]: I1002 02:08:10.111366 4885 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/eea66006-67ba-45e3-bdf9-9d144d772386-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 02 02:08:10 crc kubenswrapper[4885]: I1002 02:08:10.200632 4885 generic.go:334] "Generic (PLEG): container finished" podID="eea66006-67ba-45e3-bdf9-9d144d772386" containerID="6d97c5a8d103c2c2b9ce70a5d8c6ef3e51104fe5a791fe6d1e52bdf3ad332dff" exitCode=0 Oct 02 02:08:10 crc kubenswrapper[4885]: I1002 02:08:10.200696 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"eea66006-67ba-45e3-bdf9-9d144d772386","Type":"ContainerDied","Data":"6d97c5a8d103c2c2b9ce70a5d8c6ef3e51104fe5a791fe6d1e52bdf3ad332dff"} Oct 02 02:08:10 crc kubenswrapper[4885]: I1002 02:08:10.200912 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"eea66006-67ba-45e3-bdf9-9d144d772386","Type":"ContainerDied","Data":"58d97479733782d79442e6a3ba3ed7515c1bff74cd50689b956caf317562f0b9"} Oct 02 02:08:10 crc kubenswrapper[4885]: I1002 02:08:10.200710 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:08:10 crc kubenswrapper[4885]: I1002 02:08:10.200942 4885 scope.go:117] "RemoveContainer" containerID="6d97c5a8d103c2c2b9ce70a5d8c6ef3e51104fe5a791fe6d1e52bdf3ad332dff" Oct 02 02:08:10 crc kubenswrapper[4885]: I1002 02:08:10.202517 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"18642896-a2b6-412c-beb1-8010484744a4","Type":"ContainerStarted","Data":"5ffff29433b307fb0fde900c95a5c29ae9f5bd61377fb04e7861fae84e83670d"} Oct 02 02:08:10 crc kubenswrapper[4885]: I1002 02:08:10.205937 4885 generic.go:334] "Generic (PLEG): container finished" podID="58f621da-0934-482b-8f10-4273021207f0" containerID="52cbb92d06be28a9f8ef5a04df4ef6ad44697c7ea906d0db02a38c5cc5daf846" exitCode=0 Oct 02 02:08:10 crc kubenswrapper[4885]: I1002 02:08:10.206001 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67b789f86c-tbrnt" Oct 02 02:08:10 crc kubenswrapper[4885]: I1002 02:08:10.207134 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cb6ffcf87-46bgc" event={"ID":"58f621da-0934-482b-8f10-4273021207f0","Type":"ContainerDied","Data":"52cbb92d06be28a9f8ef5a04df4ef6ad44697c7ea906d0db02a38c5cc5daf846"} Oct 02 02:08:10 crc kubenswrapper[4885]: I1002 02:08:10.207165 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cb6ffcf87-46bgc" event={"ID":"58f621da-0934-482b-8f10-4273021207f0","Type":"ContainerStarted","Data":"62833f5b3cab6a9477ac825763339ad014425a15af808778a1a372bd497e4302"} Oct 02 02:08:10 crc kubenswrapper[4885]: I1002 02:08:10.325929 4885 scope.go:117] "RemoveContainer" containerID="6f3eaeb38ec37be6825c9165f2dca00ced63746de7714d811aaa270ddf582da9" Oct 02 02:08:10 crc kubenswrapper[4885]: I1002 02:08:10.345215 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-tbrnt"] Oct 02 02:08:10 crc kubenswrapper[4885]: I1002 02:08:10.361318 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-tbrnt"] Oct 02 02:08:10 crc kubenswrapper[4885]: I1002 02:08:10.369190 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 02 02:08:10 crc kubenswrapper[4885]: I1002 02:08:10.370553 4885 scope.go:117] "RemoveContainer" containerID="6d97c5a8d103c2c2b9ce70a5d8c6ef3e51104fe5a791fe6d1e52bdf3ad332dff" Oct 02 02:08:10 crc kubenswrapper[4885]: E1002 02:08:10.371116 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6d97c5a8d103c2c2b9ce70a5d8c6ef3e51104fe5a791fe6d1e52bdf3ad332dff\": container with ID starting with 6d97c5a8d103c2c2b9ce70a5d8c6ef3e51104fe5a791fe6d1e52bdf3ad332dff not found: ID does not exist" containerID="6d97c5a8d103c2c2b9ce70a5d8c6ef3e51104fe5a791fe6d1e52bdf3ad332dff" Oct 02 02:08:10 crc kubenswrapper[4885]: I1002 02:08:10.371150 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d97c5a8d103c2c2b9ce70a5d8c6ef3e51104fe5a791fe6d1e52bdf3ad332dff"} err="failed to get container status \"6d97c5a8d103c2c2b9ce70a5d8c6ef3e51104fe5a791fe6d1e52bdf3ad332dff\": rpc error: code = NotFound desc = could not find container \"6d97c5a8d103c2c2b9ce70a5d8c6ef3e51104fe5a791fe6d1e52bdf3ad332dff\": container with ID starting with 6d97c5a8d103c2c2b9ce70a5d8c6ef3e51104fe5a791fe6d1e52bdf3ad332dff not found: ID does not exist" Oct 02 02:08:10 crc kubenswrapper[4885]: I1002 02:08:10.371178 4885 scope.go:117] "RemoveContainer" containerID="6f3eaeb38ec37be6825c9165f2dca00ced63746de7714d811aaa270ddf582da9" Oct 02 02:08:10 crc kubenswrapper[4885]: E1002 02:08:10.372179 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6f3eaeb38ec37be6825c9165f2dca00ced63746de7714d811aaa270ddf582da9\": container with ID starting with 6f3eaeb38ec37be6825c9165f2dca00ced63746de7714d811aaa270ddf582da9 not found: ID does not exist" containerID="6f3eaeb38ec37be6825c9165f2dca00ced63746de7714d811aaa270ddf582da9" Oct 02 02:08:10 crc kubenswrapper[4885]: I1002 02:08:10.372224 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f3eaeb38ec37be6825c9165f2dca00ced63746de7714d811aaa270ddf582da9"} err="failed to get container status \"6f3eaeb38ec37be6825c9165f2dca00ced63746de7714d811aaa270ddf582da9\": rpc error: code = NotFound desc = could not find container \"6f3eaeb38ec37be6825c9165f2dca00ced63746de7714d811aaa270ddf582da9\": container with ID starting with 6f3eaeb38ec37be6825c9165f2dca00ced63746de7714d811aaa270ddf582da9 not found: ID does not exist" Oct 02 02:08:10 crc kubenswrapper[4885]: I1002 02:08:10.377043 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 02 02:08:10 crc kubenswrapper[4885]: I1002 02:08:10.394745 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 02 02:08:10 crc kubenswrapper[4885]: E1002 02:08:10.395137 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eea66006-67ba-45e3-bdf9-9d144d772386" containerName="setup-container" Oct 02 02:08:10 crc kubenswrapper[4885]: I1002 02:08:10.395150 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="eea66006-67ba-45e3-bdf9-9d144d772386" containerName="setup-container" Oct 02 02:08:10 crc kubenswrapper[4885]: E1002 02:08:10.395170 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eea66006-67ba-45e3-bdf9-9d144d772386" containerName="rabbitmq" Oct 02 02:08:10 crc kubenswrapper[4885]: I1002 02:08:10.395177 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="eea66006-67ba-45e3-bdf9-9d144d772386" containerName="rabbitmq" Oct 02 02:08:10 crc kubenswrapper[4885]: I1002 02:08:10.395435 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="eea66006-67ba-45e3-bdf9-9d144d772386" containerName="rabbitmq" Oct 02 02:08:10 crc kubenswrapper[4885]: I1002 02:08:10.398144 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:08:10 crc kubenswrapper[4885]: I1002 02:08:10.402831 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Oct 02 02:08:10 crc kubenswrapper[4885]: I1002 02:08:10.403015 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 02 02:08:10 crc kubenswrapper[4885]: I1002 02:08:10.403051 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-mghd5" Oct 02 02:08:10 crc kubenswrapper[4885]: I1002 02:08:10.402831 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 02 02:08:10 crc kubenswrapper[4885]: I1002 02:08:10.403887 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 02 02:08:10 crc kubenswrapper[4885]: I1002 02:08:10.403936 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Oct 02 02:08:10 crc kubenswrapper[4885]: I1002 02:08:10.404010 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 02 02:08:10 crc kubenswrapper[4885]: I1002 02:08:10.418948 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 02 02:08:10 crc kubenswrapper[4885]: I1002 02:08:10.534352 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/79f97e3d-d825-4982-9f76-91089df7a67b-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"79f97e3d-d825-4982-9f76-91089df7a67b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:08:10 crc kubenswrapper[4885]: I1002 02:08:10.534403 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/79f97e3d-d825-4982-9f76-91089df7a67b-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"79f97e3d-d825-4982-9f76-91089df7a67b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:08:10 crc kubenswrapper[4885]: I1002 02:08:10.534419 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/79f97e3d-d825-4982-9f76-91089df7a67b-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"79f97e3d-d825-4982-9f76-91089df7a67b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:08:10 crc kubenswrapper[4885]: I1002 02:08:10.534447 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/79f97e3d-d825-4982-9f76-91089df7a67b-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"79f97e3d-d825-4982-9f76-91089df7a67b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:08:10 crc kubenswrapper[4885]: I1002 02:08:10.534551 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/79f97e3d-d825-4982-9f76-91089df7a67b-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"79f97e3d-d825-4982-9f76-91089df7a67b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:08:10 crc kubenswrapper[4885]: I1002 02:08:10.534646 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/79f97e3d-d825-4982-9f76-91089df7a67b-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"79f97e3d-d825-4982-9f76-91089df7a67b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:08:10 crc kubenswrapper[4885]: I1002 02:08:10.534715 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/79f97e3d-d825-4982-9f76-91089df7a67b-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"79f97e3d-d825-4982-9f76-91089df7a67b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:08:10 crc kubenswrapper[4885]: I1002 02:08:10.534737 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xtfrq\" (UniqueName: \"kubernetes.io/projected/79f97e3d-d825-4982-9f76-91089df7a67b-kube-api-access-xtfrq\") pod \"rabbitmq-cell1-server-0\" (UID: \"79f97e3d-d825-4982-9f76-91089df7a67b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:08:10 crc kubenswrapper[4885]: I1002 02:08:10.534935 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/79f97e3d-d825-4982-9f76-91089df7a67b-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"79f97e3d-d825-4982-9f76-91089df7a67b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:08:10 crc kubenswrapper[4885]: I1002 02:08:10.535047 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"79f97e3d-d825-4982-9f76-91089df7a67b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:08:10 crc kubenswrapper[4885]: I1002 02:08:10.535082 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/79f97e3d-d825-4982-9f76-91089df7a67b-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"79f97e3d-d825-4982-9f76-91089df7a67b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:08:10 crc kubenswrapper[4885]: I1002 02:08:10.636700 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/79f97e3d-d825-4982-9f76-91089df7a67b-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"79f97e3d-d825-4982-9f76-91089df7a67b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:08:10 crc kubenswrapper[4885]: I1002 02:08:10.636778 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/79f97e3d-d825-4982-9f76-91089df7a67b-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"79f97e3d-d825-4982-9f76-91089df7a67b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:08:10 crc kubenswrapper[4885]: I1002 02:08:10.636801 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xtfrq\" (UniqueName: \"kubernetes.io/projected/79f97e3d-d825-4982-9f76-91089df7a67b-kube-api-access-xtfrq\") pod \"rabbitmq-cell1-server-0\" (UID: \"79f97e3d-d825-4982-9f76-91089df7a67b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:08:10 crc kubenswrapper[4885]: I1002 02:08:10.636843 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/79f97e3d-d825-4982-9f76-91089df7a67b-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"79f97e3d-d825-4982-9f76-91089df7a67b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:08:10 crc kubenswrapper[4885]: I1002 02:08:10.636873 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"79f97e3d-d825-4982-9f76-91089df7a67b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:08:10 crc kubenswrapper[4885]: I1002 02:08:10.636893 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/79f97e3d-d825-4982-9f76-91089df7a67b-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"79f97e3d-d825-4982-9f76-91089df7a67b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:08:10 crc kubenswrapper[4885]: I1002 02:08:10.636917 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/79f97e3d-d825-4982-9f76-91089df7a67b-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"79f97e3d-d825-4982-9f76-91089df7a67b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:08:10 crc kubenswrapper[4885]: I1002 02:08:10.636951 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/79f97e3d-d825-4982-9f76-91089df7a67b-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"79f97e3d-d825-4982-9f76-91089df7a67b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:08:10 crc kubenswrapper[4885]: I1002 02:08:10.636967 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/79f97e3d-d825-4982-9f76-91089df7a67b-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"79f97e3d-d825-4982-9f76-91089df7a67b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:08:10 crc kubenswrapper[4885]: I1002 02:08:10.636992 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/79f97e3d-d825-4982-9f76-91089df7a67b-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"79f97e3d-d825-4982-9f76-91089df7a67b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:08:10 crc kubenswrapper[4885]: I1002 02:08:10.637021 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/79f97e3d-d825-4982-9f76-91089df7a67b-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"79f97e3d-d825-4982-9f76-91089df7a67b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:08:10 crc kubenswrapper[4885]: I1002 02:08:10.637520 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/79f97e3d-d825-4982-9f76-91089df7a67b-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"79f97e3d-d825-4982-9f76-91089df7a67b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:08:10 crc kubenswrapper[4885]: I1002 02:08:10.637576 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/79f97e3d-d825-4982-9f76-91089df7a67b-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"79f97e3d-d825-4982-9f76-91089df7a67b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:08:10 crc kubenswrapper[4885]: I1002 02:08:10.637663 4885 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"79f97e3d-d825-4982-9f76-91089df7a67b\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:08:10 crc kubenswrapper[4885]: I1002 02:08:10.645866 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/79f97e3d-d825-4982-9f76-91089df7a67b-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"79f97e3d-d825-4982-9f76-91089df7a67b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:08:10 crc kubenswrapper[4885]: I1002 02:08:10.646111 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/79f97e3d-d825-4982-9f76-91089df7a67b-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"79f97e3d-d825-4982-9f76-91089df7a67b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:08:10 crc kubenswrapper[4885]: I1002 02:08:10.646118 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/79f97e3d-d825-4982-9f76-91089df7a67b-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"79f97e3d-d825-4982-9f76-91089df7a67b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:08:10 crc kubenswrapper[4885]: I1002 02:08:10.646724 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/79f97e3d-d825-4982-9f76-91089df7a67b-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"79f97e3d-d825-4982-9f76-91089df7a67b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:08:10 crc kubenswrapper[4885]: I1002 02:08:10.647274 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/79f97e3d-d825-4982-9f76-91089df7a67b-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"79f97e3d-d825-4982-9f76-91089df7a67b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:08:10 crc kubenswrapper[4885]: I1002 02:08:10.647302 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/79f97e3d-d825-4982-9f76-91089df7a67b-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"79f97e3d-d825-4982-9f76-91089df7a67b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:08:10 crc kubenswrapper[4885]: I1002 02:08:10.647819 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/79f97e3d-d825-4982-9f76-91089df7a67b-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"79f97e3d-d825-4982-9f76-91089df7a67b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:08:10 crc kubenswrapper[4885]: I1002 02:08:10.688341 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"79f97e3d-d825-4982-9f76-91089df7a67b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:08:10 crc kubenswrapper[4885]: I1002 02:08:10.759643 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xtfrq\" (UniqueName: \"kubernetes.io/projected/79f97e3d-d825-4982-9f76-91089df7a67b-kube-api-access-xtfrq\") pod \"rabbitmq-cell1-server-0\" (UID: \"79f97e3d-d825-4982-9f76-91089df7a67b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:08:11 crc kubenswrapper[4885]: I1002 02:08:11.015177 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:08:11 crc kubenswrapper[4885]: I1002 02:08:11.219715 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"18642896-a2b6-412c-beb1-8010484744a4","Type":"ContainerStarted","Data":"131cb899ad493e7f491e27c910b06f728dcc702fa63cd5d84243f710007b782c"} Oct 02 02:08:11 crc kubenswrapper[4885]: I1002 02:08:11.224919 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cb6ffcf87-46bgc" event={"ID":"58f621da-0934-482b-8f10-4273021207f0","Type":"ContainerStarted","Data":"193cc6840178ca290f995d0d7a9ab4d27d6df84c2543790edc7d027fd43185bd"} Oct 02 02:08:11 crc kubenswrapper[4885]: I1002 02:08:11.225430 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-cb6ffcf87-46bgc" Oct 02 02:08:11 crc kubenswrapper[4885]: I1002 02:08:11.291093 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-cb6ffcf87-46bgc" podStartSLOduration=3.291073517 podStartE2EDuration="3.291073517s" podCreationTimestamp="2025-10-02 02:08:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:08:11.279458104 +0000 UTC m=+1280.091205533" watchObservedRunningTime="2025-10-02 02:08:11.291073517 +0000 UTC m=+1280.102820926" Oct 02 02:08:11 crc kubenswrapper[4885]: I1002 02:08:11.576545 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 02 02:08:11 crc kubenswrapper[4885]: W1002 02:08:11.582410 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod79f97e3d_d825_4982_9f76_91089df7a67b.slice/crio-be621c1964e445abd294260a707478388a69a2c72524da326d3eae8485f693f7 WatchSource:0}: Error finding container be621c1964e445abd294260a707478388a69a2c72524da326d3eae8485f693f7: Status 404 returned error can't find the container with id be621c1964e445abd294260a707478388a69a2c72524da326d3eae8485f693f7 Oct 02 02:08:12 crc kubenswrapper[4885]: I1002 02:08:12.069474 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b0452212-e578-4bf2-9199-91eefc38a3f2" path="/var/lib/kubelet/pods/b0452212-e578-4bf2-9199-91eefc38a3f2/volumes" Oct 02 02:08:12 crc kubenswrapper[4885]: I1002 02:08:12.071541 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eea66006-67ba-45e3-bdf9-9d144d772386" path="/var/lib/kubelet/pods/eea66006-67ba-45e3-bdf9-9d144d772386/volumes" Oct 02 02:08:12 crc kubenswrapper[4885]: I1002 02:08:12.242374 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"79f97e3d-d825-4982-9f76-91089df7a67b","Type":"ContainerStarted","Data":"be621c1964e445abd294260a707478388a69a2c72524da326d3eae8485f693f7"} Oct 02 02:08:14 crc kubenswrapper[4885]: I1002 02:08:14.291234 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"79f97e3d-d825-4982-9f76-91089df7a67b","Type":"ContainerStarted","Data":"c6b91b20446e1d28482170d24172f54c7c84e5823f6b6f3e7d3dfd91fe4238b0"} Oct 02 02:08:19 crc kubenswrapper[4885]: I1002 02:08:19.363494 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-cb6ffcf87-46bgc" Oct 02 02:08:19 crc kubenswrapper[4885]: I1002 02:08:19.467588 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-vgljb"] Oct 02 02:08:19 crc kubenswrapper[4885]: I1002 02:08:19.468203 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-59cf4bdb65-vgljb" podUID="05914369-d37a-4911-b64b-4ae7d03385a1" containerName="dnsmasq-dns" containerID="cri-o://32b84363f04f3add71d5b400ef00ba9c07cd3c14405ec09f9ea64970416f9e78" gracePeriod=10 Oct 02 02:08:20 crc kubenswrapper[4885]: I1002 02:08:20.010048 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cf4bdb65-vgljb" Oct 02 02:08:20 crc kubenswrapper[4885]: I1002 02:08:20.122022 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/05914369-d37a-4911-b64b-4ae7d03385a1-dns-swift-storage-0\") pod \"05914369-d37a-4911-b64b-4ae7d03385a1\" (UID: \"05914369-d37a-4911-b64b-4ae7d03385a1\") " Oct 02 02:08:20 crc kubenswrapper[4885]: I1002 02:08:20.122153 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/05914369-d37a-4911-b64b-4ae7d03385a1-dns-svc\") pod \"05914369-d37a-4911-b64b-4ae7d03385a1\" (UID: \"05914369-d37a-4911-b64b-4ae7d03385a1\") " Oct 02 02:08:20 crc kubenswrapper[4885]: I1002 02:08:20.122221 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/05914369-d37a-4911-b64b-4ae7d03385a1-config\") pod \"05914369-d37a-4911-b64b-4ae7d03385a1\" (UID: \"05914369-d37a-4911-b64b-4ae7d03385a1\") " Oct 02 02:08:20 crc kubenswrapper[4885]: I1002 02:08:20.122334 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/05914369-d37a-4911-b64b-4ae7d03385a1-ovsdbserver-sb\") pod \"05914369-d37a-4911-b64b-4ae7d03385a1\" (UID: \"05914369-d37a-4911-b64b-4ae7d03385a1\") " Oct 02 02:08:20 crc kubenswrapper[4885]: I1002 02:08:20.122354 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g4bth\" (UniqueName: \"kubernetes.io/projected/05914369-d37a-4911-b64b-4ae7d03385a1-kube-api-access-g4bth\") pod \"05914369-d37a-4911-b64b-4ae7d03385a1\" (UID: \"05914369-d37a-4911-b64b-4ae7d03385a1\") " Oct 02 02:08:20 crc kubenswrapper[4885]: I1002 02:08:20.122384 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/05914369-d37a-4911-b64b-4ae7d03385a1-ovsdbserver-nb\") pod \"05914369-d37a-4911-b64b-4ae7d03385a1\" (UID: \"05914369-d37a-4911-b64b-4ae7d03385a1\") " Oct 02 02:08:20 crc kubenswrapper[4885]: I1002 02:08:20.134493 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/05914369-d37a-4911-b64b-4ae7d03385a1-kube-api-access-g4bth" (OuterVolumeSpecName: "kube-api-access-g4bth") pod "05914369-d37a-4911-b64b-4ae7d03385a1" (UID: "05914369-d37a-4911-b64b-4ae7d03385a1"). InnerVolumeSpecName "kube-api-access-g4bth". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:08:20 crc kubenswrapper[4885]: I1002 02:08:20.175202 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/05914369-d37a-4911-b64b-4ae7d03385a1-config" (OuterVolumeSpecName: "config") pod "05914369-d37a-4911-b64b-4ae7d03385a1" (UID: "05914369-d37a-4911-b64b-4ae7d03385a1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:08:20 crc kubenswrapper[4885]: I1002 02:08:20.176891 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/05914369-d37a-4911-b64b-4ae7d03385a1-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "05914369-d37a-4911-b64b-4ae7d03385a1" (UID: "05914369-d37a-4911-b64b-4ae7d03385a1"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:08:20 crc kubenswrapper[4885]: I1002 02:08:20.189089 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/05914369-d37a-4911-b64b-4ae7d03385a1-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "05914369-d37a-4911-b64b-4ae7d03385a1" (UID: "05914369-d37a-4911-b64b-4ae7d03385a1"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:08:20 crc kubenswrapper[4885]: I1002 02:08:20.198635 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/05914369-d37a-4911-b64b-4ae7d03385a1-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "05914369-d37a-4911-b64b-4ae7d03385a1" (UID: "05914369-d37a-4911-b64b-4ae7d03385a1"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:08:20 crc kubenswrapper[4885]: I1002 02:08:20.201387 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/05914369-d37a-4911-b64b-4ae7d03385a1-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "05914369-d37a-4911-b64b-4ae7d03385a1" (UID: "05914369-d37a-4911-b64b-4ae7d03385a1"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:08:20 crc kubenswrapper[4885]: I1002 02:08:20.240730 4885 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/05914369-d37a-4911-b64b-4ae7d03385a1-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 02:08:20 crc kubenswrapper[4885]: I1002 02:08:20.241054 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g4bth\" (UniqueName: \"kubernetes.io/projected/05914369-d37a-4911-b64b-4ae7d03385a1-kube-api-access-g4bth\") on node \"crc\" DevicePath \"\"" Oct 02 02:08:20 crc kubenswrapper[4885]: I1002 02:08:20.241070 4885 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/05914369-d37a-4911-b64b-4ae7d03385a1-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 02:08:20 crc kubenswrapper[4885]: I1002 02:08:20.241082 4885 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/05914369-d37a-4911-b64b-4ae7d03385a1-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 02 02:08:20 crc kubenswrapper[4885]: I1002 02:08:20.241097 4885 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/05914369-d37a-4911-b64b-4ae7d03385a1-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 02:08:20 crc kubenswrapper[4885]: I1002 02:08:20.241108 4885 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/05914369-d37a-4911-b64b-4ae7d03385a1-config\") on node \"crc\" DevicePath \"\"" Oct 02 02:08:20 crc kubenswrapper[4885]: I1002 02:08:20.370867 4885 generic.go:334] "Generic (PLEG): container finished" podID="05914369-d37a-4911-b64b-4ae7d03385a1" containerID="32b84363f04f3add71d5b400ef00ba9c07cd3c14405ec09f9ea64970416f9e78" exitCode=0 Oct 02 02:08:20 crc kubenswrapper[4885]: I1002 02:08:20.370916 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-vgljb" event={"ID":"05914369-d37a-4911-b64b-4ae7d03385a1","Type":"ContainerDied","Data":"32b84363f04f3add71d5b400ef00ba9c07cd3c14405ec09f9ea64970416f9e78"} Oct 02 02:08:20 crc kubenswrapper[4885]: I1002 02:08:20.370953 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-vgljb" event={"ID":"05914369-d37a-4911-b64b-4ae7d03385a1","Type":"ContainerDied","Data":"51aca9c27883ab045da41f8f4a0298c170ef143c2d4cf7345be995ebb50bc380"} Oct 02 02:08:20 crc kubenswrapper[4885]: I1002 02:08:20.370974 4885 scope.go:117] "RemoveContainer" containerID="32b84363f04f3add71d5b400ef00ba9c07cd3c14405ec09f9ea64970416f9e78" Oct 02 02:08:20 crc kubenswrapper[4885]: I1002 02:08:20.371042 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cf4bdb65-vgljb" Oct 02 02:08:20 crc kubenswrapper[4885]: I1002 02:08:20.408499 4885 scope.go:117] "RemoveContainer" containerID="59d44cb4910beddea301c3811e444fac74f711d7e3aabef257b6071f7a9dc01c" Oct 02 02:08:20 crc kubenswrapper[4885]: I1002 02:08:20.416907 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-vgljb"] Oct 02 02:08:20 crc kubenswrapper[4885]: I1002 02:08:20.428246 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-vgljb"] Oct 02 02:08:20 crc kubenswrapper[4885]: I1002 02:08:20.435717 4885 scope.go:117] "RemoveContainer" containerID="32b84363f04f3add71d5b400ef00ba9c07cd3c14405ec09f9ea64970416f9e78" Oct 02 02:08:20 crc kubenswrapper[4885]: E1002 02:08:20.436162 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"32b84363f04f3add71d5b400ef00ba9c07cd3c14405ec09f9ea64970416f9e78\": container with ID starting with 32b84363f04f3add71d5b400ef00ba9c07cd3c14405ec09f9ea64970416f9e78 not found: ID does not exist" containerID="32b84363f04f3add71d5b400ef00ba9c07cd3c14405ec09f9ea64970416f9e78" Oct 02 02:08:20 crc kubenswrapper[4885]: I1002 02:08:20.436202 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"32b84363f04f3add71d5b400ef00ba9c07cd3c14405ec09f9ea64970416f9e78"} err="failed to get container status \"32b84363f04f3add71d5b400ef00ba9c07cd3c14405ec09f9ea64970416f9e78\": rpc error: code = NotFound desc = could not find container \"32b84363f04f3add71d5b400ef00ba9c07cd3c14405ec09f9ea64970416f9e78\": container with ID starting with 32b84363f04f3add71d5b400ef00ba9c07cd3c14405ec09f9ea64970416f9e78 not found: ID does not exist" Oct 02 02:08:20 crc kubenswrapper[4885]: I1002 02:08:20.436227 4885 scope.go:117] "RemoveContainer" containerID="59d44cb4910beddea301c3811e444fac74f711d7e3aabef257b6071f7a9dc01c" Oct 02 02:08:20 crc kubenswrapper[4885]: E1002 02:08:20.436587 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"59d44cb4910beddea301c3811e444fac74f711d7e3aabef257b6071f7a9dc01c\": container with ID starting with 59d44cb4910beddea301c3811e444fac74f711d7e3aabef257b6071f7a9dc01c not found: ID does not exist" containerID="59d44cb4910beddea301c3811e444fac74f711d7e3aabef257b6071f7a9dc01c" Oct 02 02:08:20 crc kubenswrapper[4885]: I1002 02:08:20.436614 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"59d44cb4910beddea301c3811e444fac74f711d7e3aabef257b6071f7a9dc01c"} err="failed to get container status \"59d44cb4910beddea301c3811e444fac74f711d7e3aabef257b6071f7a9dc01c\": rpc error: code = NotFound desc = could not find container \"59d44cb4910beddea301c3811e444fac74f711d7e3aabef257b6071f7a9dc01c\": container with ID starting with 59d44cb4910beddea301c3811e444fac74f711d7e3aabef257b6071f7a9dc01c not found: ID does not exist" Oct 02 02:08:22 crc kubenswrapper[4885]: I1002 02:08:22.063525 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="05914369-d37a-4911-b64b-4ae7d03385a1" path="/var/lib/kubelet/pods/05914369-d37a-4911-b64b-4ae7d03385a1/volumes" Oct 02 02:08:28 crc kubenswrapper[4885]: I1002 02:08:28.192686 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vmf4j"] Oct 02 02:08:28 crc kubenswrapper[4885]: E1002 02:08:28.194195 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05914369-d37a-4911-b64b-4ae7d03385a1" containerName="dnsmasq-dns" Oct 02 02:08:28 crc kubenswrapper[4885]: I1002 02:08:28.194227 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="05914369-d37a-4911-b64b-4ae7d03385a1" containerName="dnsmasq-dns" Oct 02 02:08:28 crc kubenswrapper[4885]: E1002 02:08:28.194317 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05914369-d37a-4911-b64b-4ae7d03385a1" containerName="init" Oct 02 02:08:28 crc kubenswrapper[4885]: I1002 02:08:28.194336 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="05914369-d37a-4911-b64b-4ae7d03385a1" containerName="init" Oct 02 02:08:28 crc kubenswrapper[4885]: I1002 02:08:28.194822 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="05914369-d37a-4911-b64b-4ae7d03385a1" containerName="dnsmasq-dns" Oct 02 02:08:28 crc kubenswrapper[4885]: I1002 02:08:28.196230 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vmf4j" Oct 02 02:08:28 crc kubenswrapper[4885]: I1002 02:08:28.198879 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 02:08:28 crc kubenswrapper[4885]: I1002 02:08:28.198879 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 02:08:28 crc kubenswrapper[4885]: I1002 02:08:28.198875 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-f8tvb" Oct 02 02:08:28 crc kubenswrapper[4885]: I1002 02:08:28.204930 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 02:08:28 crc kubenswrapper[4885]: I1002 02:08:28.208441 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vmf4j"] Oct 02 02:08:28 crc kubenswrapper[4885]: I1002 02:08:28.216190 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6602e994-3938-456d-8bb3-e092b6350c8a-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vmf4j\" (UID: \"6602e994-3938-456d-8bb3-e092b6350c8a\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vmf4j" Oct 02 02:08:28 crc kubenswrapper[4885]: I1002 02:08:28.216249 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6602e994-3938-456d-8bb3-e092b6350c8a-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vmf4j\" (UID: \"6602e994-3938-456d-8bb3-e092b6350c8a\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vmf4j" Oct 02 02:08:28 crc kubenswrapper[4885]: I1002 02:08:28.216326 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gpwgd\" (UniqueName: \"kubernetes.io/projected/6602e994-3938-456d-8bb3-e092b6350c8a-kube-api-access-gpwgd\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vmf4j\" (UID: \"6602e994-3938-456d-8bb3-e092b6350c8a\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vmf4j" Oct 02 02:08:28 crc kubenswrapper[4885]: I1002 02:08:28.216503 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6602e994-3938-456d-8bb3-e092b6350c8a-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vmf4j\" (UID: \"6602e994-3938-456d-8bb3-e092b6350c8a\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vmf4j" Oct 02 02:08:28 crc kubenswrapper[4885]: I1002 02:08:28.318003 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gpwgd\" (UniqueName: \"kubernetes.io/projected/6602e994-3938-456d-8bb3-e092b6350c8a-kube-api-access-gpwgd\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vmf4j\" (UID: \"6602e994-3938-456d-8bb3-e092b6350c8a\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vmf4j" Oct 02 02:08:28 crc kubenswrapper[4885]: I1002 02:08:28.318390 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6602e994-3938-456d-8bb3-e092b6350c8a-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vmf4j\" (UID: \"6602e994-3938-456d-8bb3-e092b6350c8a\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vmf4j" Oct 02 02:08:28 crc kubenswrapper[4885]: I1002 02:08:28.318489 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6602e994-3938-456d-8bb3-e092b6350c8a-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vmf4j\" (UID: \"6602e994-3938-456d-8bb3-e092b6350c8a\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vmf4j" Oct 02 02:08:28 crc kubenswrapper[4885]: I1002 02:08:28.318554 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6602e994-3938-456d-8bb3-e092b6350c8a-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vmf4j\" (UID: \"6602e994-3938-456d-8bb3-e092b6350c8a\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vmf4j" Oct 02 02:08:28 crc kubenswrapper[4885]: I1002 02:08:28.325306 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6602e994-3938-456d-8bb3-e092b6350c8a-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vmf4j\" (UID: \"6602e994-3938-456d-8bb3-e092b6350c8a\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vmf4j" Oct 02 02:08:28 crc kubenswrapper[4885]: I1002 02:08:28.325308 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6602e994-3938-456d-8bb3-e092b6350c8a-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vmf4j\" (UID: \"6602e994-3938-456d-8bb3-e092b6350c8a\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vmf4j" Oct 02 02:08:28 crc kubenswrapper[4885]: I1002 02:08:28.326043 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6602e994-3938-456d-8bb3-e092b6350c8a-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vmf4j\" (UID: \"6602e994-3938-456d-8bb3-e092b6350c8a\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vmf4j" Oct 02 02:08:28 crc kubenswrapper[4885]: I1002 02:08:28.341619 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gpwgd\" (UniqueName: \"kubernetes.io/projected/6602e994-3938-456d-8bb3-e092b6350c8a-kube-api-access-gpwgd\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vmf4j\" (UID: \"6602e994-3938-456d-8bb3-e092b6350c8a\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vmf4j" Oct 02 02:08:28 crc kubenswrapper[4885]: I1002 02:08:28.518043 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vmf4j" Oct 02 02:08:29 crc kubenswrapper[4885]: I1002 02:08:29.254690 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vmf4j"] Oct 02 02:08:29 crc kubenswrapper[4885]: W1002 02:08:29.261250 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6602e994_3938_456d_8bb3_e092b6350c8a.slice/crio-3ec27d58493494821396dcbb6833b19caf30c560f51902ef50ae64cddd04c750 WatchSource:0}: Error finding container 3ec27d58493494821396dcbb6833b19caf30c560f51902ef50ae64cddd04c750: Status 404 returned error can't find the container with id 3ec27d58493494821396dcbb6833b19caf30c560f51902ef50ae64cddd04c750 Oct 02 02:08:29 crc kubenswrapper[4885]: I1002 02:08:29.482620 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vmf4j" event={"ID":"6602e994-3938-456d-8bb3-e092b6350c8a","Type":"ContainerStarted","Data":"3ec27d58493494821396dcbb6833b19caf30c560f51902ef50ae64cddd04c750"} Oct 02 02:08:38 crc kubenswrapper[4885]: I1002 02:08:38.575803 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vmf4j" event={"ID":"6602e994-3938-456d-8bb3-e092b6350c8a","Type":"ContainerStarted","Data":"f9701dfbe800df79e26f904a1f20c4e599785885948f64d73ea2294524a03550"} Oct 02 02:08:38 crc kubenswrapper[4885]: I1002 02:08:38.599415 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vmf4j" podStartSLOduration=2.31349655 podStartE2EDuration="10.599398532s" podCreationTimestamp="2025-10-02 02:08:28 +0000 UTC" firstStartedPulling="2025-10-02 02:08:29.263706032 +0000 UTC m=+1298.075453431" lastFinishedPulling="2025-10-02 02:08:37.549607974 +0000 UTC m=+1306.361355413" observedRunningTime="2025-10-02 02:08:38.597613468 +0000 UTC m=+1307.409360877" watchObservedRunningTime="2025-10-02 02:08:38.599398532 +0000 UTC m=+1307.411145931" Oct 02 02:08:43 crc kubenswrapper[4885]: I1002 02:08:43.265786 4885 patch_prober.go:28] interesting pod/machine-config-daemon-rttx8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 02:08:43 crc kubenswrapper[4885]: I1002 02:08:43.266122 4885 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 02:08:45 crc kubenswrapper[4885]: I1002 02:08:45.670584 4885 generic.go:334] "Generic (PLEG): container finished" podID="18642896-a2b6-412c-beb1-8010484744a4" containerID="131cb899ad493e7f491e27c910b06f728dcc702fa63cd5d84243f710007b782c" exitCode=0 Oct 02 02:08:45 crc kubenswrapper[4885]: I1002 02:08:45.670698 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"18642896-a2b6-412c-beb1-8010484744a4","Type":"ContainerDied","Data":"131cb899ad493e7f491e27c910b06f728dcc702fa63cd5d84243f710007b782c"} Oct 02 02:08:46 crc kubenswrapper[4885]: I1002 02:08:46.687231 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"18642896-a2b6-412c-beb1-8010484744a4","Type":"ContainerStarted","Data":"cdbe17a50ce5e33e28a8a3b626a144c134d94379929e2ab768583d047483fba8"} Oct 02 02:08:46 crc kubenswrapper[4885]: I1002 02:08:46.688685 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 02 02:08:46 crc kubenswrapper[4885]: I1002 02:08:46.737363 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=38.73732905 podStartE2EDuration="38.73732905s" podCreationTimestamp="2025-10-02 02:08:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:08:46.731613337 +0000 UTC m=+1315.543360766" watchObservedRunningTime="2025-10-02 02:08:46.73732905 +0000 UTC m=+1315.549076479" Oct 02 02:08:47 crc kubenswrapper[4885]: I1002 02:08:47.699338 4885 generic.go:334] "Generic (PLEG): container finished" podID="79f97e3d-d825-4982-9f76-91089df7a67b" containerID="c6b91b20446e1d28482170d24172f54c7c84e5823f6b6f3e7d3dfd91fe4238b0" exitCode=0 Oct 02 02:08:47 crc kubenswrapper[4885]: I1002 02:08:47.699479 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"79f97e3d-d825-4982-9f76-91089df7a67b","Type":"ContainerDied","Data":"c6b91b20446e1d28482170d24172f54c7c84e5823f6b6f3e7d3dfd91fe4238b0"} Oct 02 02:08:48 crc kubenswrapper[4885]: I1002 02:08:48.711384 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"79f97e3d-d825-4982-9f76-91089df7a67b","Type":"ContainerStarted","Data":"b18996265f4acd9e59f8cd2c82ad321d75f3c0571fd263765b5399d4d34b8be3"} Oct 02 02:08:48 crc kubenswrapper[4885]: I1002 02:08:48.712064 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:08:48 crc kubenswrapper[4885]: I1002 02:08:48.744790 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=38.744771778 podStartE2EDuration="38.744771778s" podCreationTimestamp="2025-10-02 02:08:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:08:48.735014352 +0000 UTC m=+1317.546761751" watchObservedRunningTime="2025-10-02 02:08:48.744771778 +0000 UTC m=+1317.556519167" Oct 02 02:08:49 crc kubenswrapper[4885]: I1002 02:08:49.724990 4885 generic.go:334] "Generic (PLEG): container finished" podID="6602e994-3938-456d-8bb3-e092b6350c8a" containerID="f9701dfbe800df79e26f904a1f20c4e599785885948f64d73ea2294524a03550" exitCode=0 Oct 02 02:08:49 crc kubenswrapper[4885]: I1002 02:08:49.725081 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vmf4j" event={"ID":"6602e994-3938-456d-8bb3-e092b6350c8a","Type":"ContainerDied","Data":"f9701dfbe800df79e26f904a1f20c4e599785885948f64d73ea2294524a03550"} Oct 02 02:08:51 crc kubenswrapper[4885]: I1002 02:08:51.197543 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vmf4j" Oct 02 02:08:51 crc kubenswrapper[4885]: I1002 02:08:51.252368 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6602e994-3938-456d-8bb3-e092b6350c8a-ssh-key\") pod \"6602e994-3938-456d-8bb3-e092b6350c8a\" (UID: \"6602e994-3938-456d-8bb3-e092b6350c8a\") " Oct 02 02:08:51 crc kubenswrapper[4885]: I1002 02:08:51.252478 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6602e994-3938-456d-8bb3-e092b6350c8a-inventory\") pod \"6602e994-3938-456d-8bb3-e092b6350c8a\" (UID: \"6602e994-3938-456d-8bb3-e092b6350c8a\") " Oct 02 02:08:51 crc kubenswrapper[4885]: I1002 02:08:51.252536 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6602e994-3938-456d-8bb3-e092b6350c8a-repo-setup-combined-ca-bundle\") pod \"6602e994-3938-456d-8bb3-e092b6350c8a\" (UID: \"6602e994-3938-456d-8bb3-e092b6350c8a\") " Oct 02 02:08:51 crc kubenswrapper[4885]: I1002 02:08:51.252555 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gpwgd\" (UniqueName: \"kubernetes.io/projected/6602e994-3938-456d-8bb3-e092b6350c8a-kube-api-access-gpwgd\") pod \"6602e994-3938-456d-8bb3-e092b6350c8a\" (UID: \"6602e994-3938-456d-8bb3-e092b6350c8a\") " Oct 02 02:08:51 crc kubenswrapper[4885]: I1002 02:08:51.257796 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6602e994-3938-456d-8bb3-e092b6350c8a-kube-api-access-gpwgd" (OuterVolumeSpecName: "kube-api-access-gpwgd") pod "6602e994-3938-456d-8bb3-e092b6350c8a" (UID: "6602e994-3938-456d-8bb3-e092b6350c8a"). InnerVolumeSpecName "kube-api-access-gpwgd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:08:51 crc kubenswrapper[4885]: I1002 02:08:51.258562 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6602e994-3938-456d-8bb3-e092b6350c8a-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "6602e994-3938-456d-8bb3-e092b6350c8a" (UID: "6602e994-3938-456d-8bb3-e092b6350c8a"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:08:51 crc kubenswrapper[4885]: I1002 02:08:51.287096 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6602e994-3938-456d-8bb3-e092b6350c8a-inventory" (OuterVolumeSpecName: "inventory") pod "6602e994-3938-456d-8bb3-e092b6350c8a" (UID: "6602e994-3938-456d-8bb3-e092b6350c8a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:08:51 crc kubenswrapper[4885]: I1002 02:08:51.291727 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6602e994-3938-456d-8bb3-e092b6350c8a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "6602e994-3938-456d-8bb3-e092b6350c8a" (UID: "6602e994-3938-456d-8bb3-e092b6350c8a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:08:51 crc kubenswrapper[4885]: I1002 02:08:51.354555 4885 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6602e994-3938-456d-8bb3-e092b6350c8a-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 02:08:51 crc kubenswrapper[4885]: I1002 02:08:51.354591 4885 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6602e994-3938-456d-8bb3-e092b6350c8a-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 02:08:51 crc kubenswrapper[4885]: I1002 02:08:51.354608 4885 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6602e994-3938-456d-8bb3-e092b6350c8a-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:08:51 crc kubenswrapper[4885]: I1002 02:08:51.354622 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gpwgd\" (UniqueName: \"kubernetes.io/projected/6602e994-3938-456d-8bb3-e092b6350c8a-kube-api-access-gpwgd\") on node \"crc\" DevicePath \"\"" Oct 02 02:08:51 crc kubenswrapper[4885]: I1002 02:08:51.748236 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vmf4j" event={"ID":"6602e994-3938-456d-8bb3-e092b6350c8a","Type":"ContainerDied","Data":"3ec27d58493494821396dcbb6833b19caf30c560f51902ef50ae64cddd04c750"} Oct 02 02:08:51 crc kubenswrapper[4885]: I1002 02:08:51.748299 4885 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3ec27d58493494821396dcbb6833b19caf30c560f51902ef50ae64cddd04c750" Oct 02 02:08:51 crc kubenswrapper[4885]: I1002 02:08:51.748310 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vmf4j" Oct 02 02:08:51 crc kubenswrapper[4885]: I1002 02:08:51.857955 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-6nvvk"] Oct 02 02:08:51 crc kubenswrapper[4885]: E1002 02:08:51.858500 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6602e994-3938-456d-8bb3-e092b6350c8a" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 02 02:08:51 crc kubenswrapper[4885]: I1002 02:08:51.858528 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="6602e994-3938-456d-8bb3-e092b6350c8a" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 02 02:08:51 crc kubenswrapper[4885]: I1002 02:08:51.858851 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="6602e994-3938-456d-8bb3-e092b6350c8a" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 02 02:08:51 crc kubenswrapper[4885]: I1002 02:08:51.859670 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-6nvvk" Oct 02 02:08:51 crc kubenswrapper[4885]: I1002 02:08:51.864358 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 02:08:51 crc kubenswrapper[4885]: I1002 02:08:51.864413 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 02:08:51 crc kubenswrapper[4885]: I1002 02:08:51.864527 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 02:08:51 crc kubenswrapper[4885]: I1002 02:08:51.864646 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-f8tvb" Oct 02 02:08:51 crc kubenswrapper[4885]: I1002 02:08:51.872543 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-6nvvk"] Oct 02 02:08:51 crc kubenswrapper[4885]: I1002 02:08:51.967764 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/379db8cf-874f-4d6e-9e3c-795457fdddb9-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-6nvvk\" (UID: \"379db8cf-874f-4d6e-9e3c-795457fdddb9\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-6nvvk" Oct 02 02:08:51 crc kubenswrapper[4885]: I1002 02:08:51.967901 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/379db8cf-874f-4d6e-9e3c-795457fdddb9-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-6nvvk\" (UID: \"379db8cf-874f-4d6e-9e3c-795457fdddb9\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-6nvvk" Oct 02 02:08:51 crc kubenswrapper[4885]: I1002 02:08:51.968082 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2x8s9\" (UniqueName: \"kubernetes.io/projected/379db8cf-874f-4d6e-9e3c-795457fdddb9-kube-api-access-2x8s9\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-6nvvk\" (UID: \"379db8cf-874f-4d6e-9e3c-795457fdddb9\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-6nvvk" Oct 02 02:08:52 crc kubenswrapper[4885]: I1002 02:08:52.070330 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/379db8cf-874f-4d6e-9e3c-795457fdddb9-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-6nvvk\" (UID: \"379db8cf-874f-4d6e-9e3c-795457fdddb9\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-6nvvk" Oct 02 02:08:52 crc kubenswrapper[4885]: I1002 02:08:52.070520 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2x8s9\" (UniqueName: \"kubernetes.io/projected/379db8cf-874f-4d6e-9e3c-795457fdddb9-kube-api-access-2x8s9\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-6nvvk\" (UID: \"379db8cf-874f-4d6e-9e3c-795457fdddb9\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-6nvvk" Oct 02 02:08:52 crc kubenswrapper[4885]: I1002 02:08:52.070589 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/379db8cf-874f-4d6e-9e3c-795457fdddb9-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-6nvvk\" (UID: \"379db8cf-874f-4d6e-9e3c-795457fdddb9\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-6nvvk" Oct 02 02:08:52 crc kubenswrapper[4885]: I1002 02:08:52.074419 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 02:08:52 crc kubenswrapper[4885]: I1002 02:08:52.075131 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 02:08:52 crc kubenswrapper[4885]: I1002 02:08:52.088279 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/379db8cf-874f-4d6e-9e3c-795457fdddb9-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-6nvvk\" (UID: \"379db8cf-874f-4d6e-9e3c-795457fdddb9\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-6nvvk" Oct 02 02:08:52 crc kubenswrapper[4885]: I1002 02:08:52.089346 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/379db8cf-874f-4d6e-9e3c-795457fdddb9-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-6nvvk\" (UID: \"379db8cf-874f-4d6e-9e3c-795457fdddb9\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-6nvvk" Oct 02 02:08:52 crc kubenswrapper[4885]: I1002 02:08:52.095972 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2x8s9\" (UniqueName: \"kubernetes.io/projected/379db8cf-874f-4d6e-9e3c-795457fdddb9-kube-api-access-2x8s9\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-6nvvk\" (UID: \"379db8cf-874f-4d6e-9e3c-795457fdddb9\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-6nvvk" Oct 02 02:08:52 crc kubenswrapper[4885]: I1002 02:08:52.183899 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-f8tvb" Oct 02 02:08:52 crc kubenswrapper[4885]: I1002 02:08:52.192938 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-6nvvk" Oct 02 02:08:52 crc kubenswrapper[4885]: I1002 02:08:52.784584 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-6nvvk"] Oct 02 02:08:52 crc kubenswrapper[4885]: W1002 02:08:52.804094 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod379db8cf_874f_4d6e_9e3c_795457fdddb9.slice/crio-c757bd6fc64828cd16efc54a34bc4d977fbfa90e7db137a6b0a24d0992aaeb4e WatchSource:0}: Error finding container c757bd6fc64828cd16efc54a34bc4d977fbfa90e7db137a6b0a24d0992aaeb4e: Status 404 returned error can't find the container with id c757bd6fc64828cd16efc54a34bc4d977fbfa90e7db137a6b0a24d0992aaeb4e Oct 02 02:08:53 crc kubenswrapper[4885]: I1002 02:08:53.776629 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-6nvvk" event={"ID":"379db8cf-874f-4d6e-9e3c-795457fdddb9","Type":"ContainerStarted","Data":"c757bd6fc64828cd16efc54a34bc4d977fbfa90e7db137a6b0a24d0992aaeb4e"} Oct 02 02:08:53 crc kubenswrapper[4885]: I1002 02:08:53.962420 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 02:08:54 crc kubenswrapper[4885]: I1002 02:08:54.790713 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-6nvvk" event={"ID":"379db8cf-874f-4d6e-9e3c-795457fdddb9","Type":"ContainerStarted","Data":"282e94c985e6f00f6ae84a82e6906c7c1df565fccff651a1e88ff3d30f931c96"} Oct 02 02:08:54 crc kubenswrapper[4885]: I1002 02:08:54.821886 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-6nvvk" podStartSLOduration=2.683088724 podStartE2EDuration="3.821862085s" podCreationTimestamp="2025-10-02 02:08:51 +0000 UTC" firstStartedPulling="2025-10-02 02:08:52.814770098 +0000 UTC m=+1321.626517537" lastFinishedPulling="2025-10-02 02:08:53.953543469 +0000 UTC m=+1322.765290898" observedRunningTime="2025-10-02 02:08:54.811229252 +0000 UTC m=+1323.622976651" watchObservedRunningTime="2025-10-02 02:08:54.821862085 +0000 UTC m=+1323.633609484" Oct 02 02:08:57 crc kubenswrapper[4885]: I1002 02:08:57.829011 4885 generic.go:334] "Generic (PLEG): container finished" podID="379db8cf-874f-4d6e-9e3c-795457fdddb9" containerID="282e94c985e6f00f6ae84a82e6906c7c1df565fccff651a1e88ff3d30f931c96" exitCode=0 Oct 02 02:08:57 crc kubenswrapper[4885]: I1002 02:08:57.829109 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-6nvvk" event={"ID":"379db8cf-874f-4d6e-9e3c-795457fdddb9","Type":"ContainerDied","Data":"282e94c985e6f00f6ae84a82e6906c7c1df565fccff651a1e88ff3d30f931c96"} Oct 02 02:08:58 crc kubenswrapper[4885]: I1002 02:08:58.860558 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 02 02:08:59 crc kubenswrapper[4885]: I1002 02:08:59.350319 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-6nvvk" Oct 02 02:08:59 crc kubenswrapper[4885]: I1002 02:08:59.521925 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/379db8cf-874f-4d6e-9e3c-795457fdddb9-inventory\") pod \"379db8cf-874f-4d6e-9e3c-795457fdddb9\" (UID: \"379db8cf-874f-4d6e-9e3c-795457fdddb9\") " Oct 02 02:08:59 crc kubenswrapper[4885]: I1002 02:08:59.522048 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2x8s9\" (UniqueName: \"kubernetes.io/projected/379db8cf-874f-4d6e-9e3c-795457fdddb9-kube-api-access-2x8s9\") pod \"379db8cf-874f-4d6e-9e3c-795457fdddb9\" (UID: \"379db8cf-874f-4d6e-9e3c-795457fdddb9\") " Oct 02 02:08:59 crc kubenswrapper[4885]: I1002 02:08:59.522173 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/379db8cf-874f-4d6e-9e3c-795457fdddb9-ssh-key\") pod \"379db8cf-874f-4d6e-9e3c-795457fdddb9\" (UID: \"379db8cf-874f-4d6e-9e3c-795457fdddb9\") " Oct 02 02:08:59 crc kubenswrapper[4885]: I1002 02:08:59.528480 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/379db8cf-874f-4d6e-9e3c-795457fdddb9-kube-api-access-2x8s9" (OuterVolumeSpecName: "kube-api-access-2x8s9") pod "379db8cf-874f-4d6e-9e3c-795457fdddb9" (UID: "379db8cf-874f-4d6e-9e3c-795457fdddb9"). InnerVolumeSpecName "kube-api-access-2x8s9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:08:59 crc kubenswrapper[4885]: I1002 02:08:59.550703 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/379db8cf-874f-4d6e-9e3c-795457fdddb9-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "379db8cf-874f-4d6e-9e3c-795457fdddb9" (UID: "379db8cf-874f-4d6e-9e3c-795457fdddb9"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:08:59 crc kubenswrapper[4885]: I1002 02:08:59.560660 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/379db8cf-874f-4d6e-9e3c-795457fdddb9-inventory" (OuterVolumeSpecName: "inventory") pod "379db8cf-874f-4d6e-9e3c-795457fdddb9" (UID: "379db8cf-874f-4d6e-9e3c-795457fdddb9"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:08:59 crc kubenswrapper[4885]: I1002 02:08:59.624890 4885 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/379db8cf-874f-4d6e-9e3c-795457fdddb9-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 02:08:59 crc kubenswrapper[4885]: I1002 02:08:59.624926 4885 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/379db8cf-874f-4d6e-9e3c-795457fdddb9-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 02:08:59 crc kubenswrapper[4885]: I1002 02:08:59.624940 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2x8s9\" (UniqueName: \"kubernetes.io/projected/379db8cf-874f-4d6e-9e3c-795457fdddb9-kube-api-access-2x8s9\") on node \"crc\" DevicePath \"\"" Oct 02 02:08:59 crc kubenswrapper[4885]: I1002 02:08:59.853584 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-6nvvk" event={"ID":"379db8cf-874f-4d6e-9e3c-795457fdddb9","Type":"ContainerDied","Data":"c757bd6fc64828cd16efc54a34bc4d977fbfa90e7db137a6b0a24d0992aaeb4e"} Oct 02 02:08:59 crc kubenswrapper[4885]: I1002 02:08:59.853625 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-6nvvk" Oct 02 02:08:59 crc kubenswrapper[4885]: I1002 02:08:59.853626 4885 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c757bd6fc64828cd16efc54a34bc4d977fbfa90e7db137a6b0a24d0992aaeb4e" Oct 02 02:08:59 crc kubenswrapper[4885]: I1002 02:08:59.948759 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rgbbj"] Oct 02 02:08:59 crc kubenswrapper[4885]: E1002 02:08:59.949187 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="379db8cf-874f-4d6e-9e3c-795457fdddb9" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Oct 02 02:08:59 crc kubenswrapper[4885]: I1002 02:08:59.949204 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="379db8cf-874f-4d6e-9e3c-795457fdddb9" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Oct 02 02:08:59 crc kubenswrapper[4885]: I1002 02:08:59.949523 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="379db8cf-874f-4d6e-9e3c-795457fdddb9" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Oct 02 02:08:59 crc kubenswrapper[4885]: I1002 02:08:59.950286 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rgbbj" Oct 02 02:08:59 crc kubenswrapper[4885]: I1002 02:08:59.952386 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 02:08:59 crc kubenswrapper[4885]: I1002 02:08:59.952625 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 02:08:59 crc kubenswrapper[4885]: I1002 02:08:59.952820 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-f8tvb" Oct 02 02:08:59 crc kubenswrapper[4885]: I1002 02:08:59.952839 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 02:08:59 crc kubenswrapper[4885]: I1002 02:08:59.967529 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rgbbj"] Oct 02 02:09:00 crc kubenswrapper[4885]: I1002 02:09:00.136322 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8cba1665-9cca-44ef-a790-5cd9b174e726-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-rgbbj\" (UID: \"8cba1665-9cca-44ef-a790-5cd9b174e726\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rgbbj" Oct 02 02:09:00 crc kubenswrapper[4885]: I1002 02:09:00.137839 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8cba1665-9cca-44ef-a790-5cd9b174e726-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-rgbbj\" (UID: \"8cba1665-9cca-44ef-a790-5cd9b174e726\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rgbbj" Oct 02 02:09:00 crc kubenswrapper[4885]: I1002 02:09:00.137990 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jptxh\" (UniqueName: \"kubernetes.io/projected/8cba1665-9cca-44ef-a790-5cd9b174e726-kube-api-access-jptxh\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-rgbbj\" (UID: \"8cba1665-9cca-44ef-a790-5cd9b174e726\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rgbbj" Oct 02 02:09:00 crc kubenswrapper[4885]: I1002 02:09:00.139828 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8cba1665-9cca-44ef-a790-5cd9b174e726-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-rgbbj\" (UID: \"8cba1665-9cca-44ef-a790-5cd9b174e726\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rgbbj" Oct 02 02:09:00 crc kubenswrapper[4885]: I1002 02:09:00.242044 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8cba1665-9cca-44ef-a790-5cd9b174e726-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-rgbbj\" (UID: \"8cba1665-9cca-44ef-a790-5cd9b174e726\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rgbbj" Oct 02 02:09:00 crc kubenswrapper[4885]: I1002 02:09:00.242198 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8cba1665-9cca-44ef-a790-5cd9b174e726-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-rgbbj\" (UID: \"8cba1665-9cca-44ef-a790-5cd9b174e726\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rgbbj" Oct 02 02:09:00 crc kubenswrapper[4885]: I1002 02:09:00.242250 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jptxh\" (UniqueName: \"kubernetes.io/projected/8cba1665-9cca-44ef-a790-5cd9b174e726-kube-api-access-jptxh\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-rgbbj\" (UID: \"8cba1665-9cca-44ef-a790-5cd9b174e726\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rgbbj" Oct 02 02:09:00 crc kubenswrapper[4885]: I1002 02:09:00.242415 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8cba1665-9cca-44ef-a790-5cd9b174e726-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-rgbbj\" (UID: \"8cba1665-9cca-44ef-a790-5cd9b174e726\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rgbbj" Oct 02 02:09:00 crc kubenswrapper[4885]: I1002 02:09:00.247823 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8cba1665-9cca-44ef-a790-5cd9b174e726-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-rgbbj\" (UID: \"8cba1665-9cca-44ef-a790-5cd9b174e726\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rgbbj" Oct 02 02:09:00 crc kubenswrapper[4885]: I1002 02:09:00.248667 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8cba1665-9cca-44ef-a790-5cd9b174e726-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-rgbbj\" (UID: \"8cba1665-9cca-44ef-a790-5cd9b174e726\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rgbbj" Oct 02 02:09:00 crc kubenswrapper[4885]: I1002 02:09:00.249150 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8cba1665-9cca-44ef-a790-5cd9b174e726-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-rgbbj\" (UID: \"8cba1665-9cca-44ef-a790-5cd9b174e726\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rgbbj" Oct 02 02:09:00 crc kubenswrapper[4885]: I1002 02:09:00.278333 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jptxh\" (UniqueName: \"kubernetes.io/projected/8cba1665-9cca-44ef-a790-5cd9b174e726-kube-api-access-jptxh\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-rgbbj\" (UID: \"8cba1665-9cca-44ef-a790-5cd9b174e726\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rgbbj" Oct 02 02:09:00 crc kubenswrapper[4885]: I1002 02:09:00.575578 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rgbbj" Oct 02 02:09:01 crc kubenswrapper[4885]: I1002 02:09:01.019463 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:09:01 crc kubenswrapper[4885]: I1002 02:09:01.222352 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rgbbj"] Oct 02 02:09:01 crc kubenswrapper[4885]: I1002 02:09:01.879135 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rgbbj" event={"ID":"8cba1665-9cca-44ef-a790-5cd9b174e726","Type":"ContainerStarted","Data":"2bcb2f37bef24de52ba79bcbbc8f3afad5d39f45abaea14dcee0350379cf6fcd"} Oct 02 02:09:02 crc kubenswrapper[4885]: I1002 02:09:02.950085 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rgbbj" event={"ID":"8cba1665-9cca-44ef-a790-5cd9b174e726","Type":"ContainerStarted","Data":"850533a0036969f7e2fc59ff244f40d27f7b95486b11d451e3b3b9b0a609b7d8"} Oct 02 02:09:02 crc kubenswrapper[4885]: I1002 02:09:02.973225 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rgbbj" podStartSLOduration=3.529557382 podStartE2EDuration="3.973207409s" podCreationTimestamp="2025-10-02 02:08:59 +0000 UTC" firstStartedPulling="2025-10-02 02:09:01.232343598 +0000 UTC m=+1330.044090997" lastFinishedPulling="2025-10-02 02:09:01.675993625 +0000 UTC m=+1330.487741024" observedRunningTime="2025-10-02 02:09:02.968612569 +0000 UTC m=+1331.780359978" watchObservedRunningTime="2025-10-02 02:09:02.973207409 +0000 UTC m=+1331.784954828" Oct 02 02:09:11 crc kubenswrapper[4885]: I1002 02:09:11.339403 4885 scope.go:117] "RemoveContainer" containerID="e4469a2fed694e34c2059aa68a82bd35bddc66ce6055feb2c64800dabde10945" Oct 02 02:09:11 crc kubenswrapper[4885]: I1002 02:09:11.382831 4885 scope.go:117] "RemoveContainer" containerID="6d45bac9a24e0df35ab7b6a7861c1ef92d45e7ae755cd19856e778a659e47200" Oct 02 02:09:11 crc kubenswrapper[4885]: I1002 02:09:11.431695 4885 scope.go:117] "RemoveContainer" containerID="671ef76134a5c82d3bedd4b9993e1330e7a9f078e6f02c8f3cca851b8d9ec988" Oct 02 02:09:11 crc kubenswrapper[4885]: I1002 02:09:11.503806 4885 scope.go:117] "RemoveContainer" containerID="13d632d093ec39287feac2e2cd715160da8bd3673c8eb579a58d27664223a705" Oct 02 02:09:13 crc kubenswrapper[4885]: I1002 02:09:13.266383 4885 patch_prober.go:28] interesting pod/machine-config-daemon-rttx8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 02:09:13 crc kubenswrapper[4885]: I1002 02:09:13.266844 4885 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 02:09:43 crc kubenswrapper[4885]: I1002 02:09:43.265834 4885 patch_prober.go:28] interesting pod/machine-config-daemon-rttx8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 02:09:43 crc kubenswrapper[4885]: I1002 02:09:43.266645 4885 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 02:09:43 crc kubenswrapper[4885]: I1002 02:09:43.266710 4885 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" Oct 02 02:09:43 crc kubenswrapper[4885]: I1002 02:09:43.267798 4885 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"fd898246d6c734dabec969ae60b5e63b1aba037d46a686b38b0527500e9db0fe"} pod="openshift-machine-config-operator/machine-config-daemon-rttx8" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 02:09:43 crc kubenswrapper[4885]: I1002 02:09:43.267893 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" containerName="machine-config-daemon" containerID="cri-o://fd898246d6c734dabec969ae60b5e63b1aba037d46a686b38b0527500e9db0fe" gracePeriod=600 Oct 02 02:09:43 crc kubenswrapper[4885]: I1002 02:09:43.438366 4885 generic.go:334] "Generic (PLEG): container finished" podID="d169c8ba-000a-4218-8e3b-4ae53035b110" containerID="fd898246d6c734dabec969ae60b5e63b1aba037d46a686b38b0527500e9db0fe" exitCode=0 Oct 02 02:09:43 crc kubenswrapper[4885]: I1002 02:09:43.438424 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" event={"ID":"d169c8ba-000a-4218-8e3b-4ae53035b110","Type":"ContainerDied","Data":"fd898246d6c734dabec969ae60b5e63b1aba037d46a686b38b0527500e9db0fe"} Oct 02 02:09:43 crc kubenswrapper[4885]: I1002 02:09:43.438476 4885 scope.go:117] "RemoveContainer" containerID="b3432e8413cbb9773cfd116ef5983ce1a5243df994d7ca0829906833f1dfa903" Oct 02 02:09:44 crc kubenswrapper[4885]: I1002 02:09:44.454697 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" event={"ID":"d169c8ba-000a-4218-8e3b-4ae53035b110","Type":"ContainerStarted","Data":"be0fc35a7f63d87317c3e5145a3df69700656f283328299a3b8545263cc22494"} Oct 02 02:10:11 crc kubenswrapper[4885]: I1002 02:10:11.662131 4885 scope.go:117] "RemoveContainer" containerID="ffe87ad35a1124753436f265a479f54b35ecbb206685610a84f40e705bc72cdd" Oct 02 02:10:18 crc kubenswrapper[4885]: I1002 02:10:18.752304 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-9jppq"] Oct 02 02:10:18 crc kubenswrapper[4885]: I1002 02:10:18.757837 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9jppq" Oct 02 02:10:18 crc kubenswrapper[4885]: I1002 02:10:18.799192 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9jppq"] Oct 02 02:10:18 crc kubenswrapper[4885]: I1002 02:10:18.880114 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7d3d389-84a8-4987-804d-caa438d0e7f4-utilities\") pod \"redhat-operators-9jppq\" (UID: \"a7d3d389-84a8-4987-804d-caa438d0e7f4\") " pod="openshift-marketplace/redhat-operators-9jppq" Oct 02 02:10:18 crc kubenswrapper[4885]: I1002 02:10:18.880742 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7d3d389-84a8-4987-804d-caa438d0e7f4-catalog-content\") pod \"redhat-operators-9jppq\" (UID: \"a7d3d389-84a8-4987-804d-caa438d0e7f4\") " pod="openshift-marketplace/redhat-operators-9jppq" Oct 02 02:10:18 crc kubenswrapper[4885]: I1002 02:10:18.880885 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8x7f5\" (UniqueName: \"kubernetes.io/projected/a7d3d389-84a8-4987-804d-caa438d0e7f4-kube-api-access-8x7f5\") pod \"redhat-operators-9jppq\" (UID: \"a7d3d389-84a8-4987-804d-caa438d0e7f4\") " pod="openshift-marketplace/redhat-operators-9jppq" Oct 02 02:10:18 crc kubenswrapper[4885]: I1002 02:10:18.983325 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7d3d389-84a8-4987-804d-caa438d0e7f4-utilities\") pod \"redhat-operators-9jppq\" (UID: \"a7d3d389-84a8-4987-804d-caa438d0e7f4\") " pod="openshift-marketplace/redhat-operators-9jppq" Oct 02 02:10:18 crc kubenswrapper[4885]: I1002 02:10:18.983430 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7d3d389-84a8-4987-804d-caa438d0e7f4-catalog-content\") pod \"redhat-operators-9jppq\" (UID: \"a7d3d389-84a8-4987-804d-caa438d0e7f4\") " pod="openshift-marketplace/redhat-operators-9jppq" Oct 02 02:10:18 crc kubenswrapper[4885]: I1002 02:10:18.983465 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8x7f5\" (UniqueName: \"kubernetes.io/projected/a7d3d389-84a8-4987-804d-caa438d0e7f4-kube-api-access-8x7f5\") pod \"redhat-operators-9jppq\" (UID: \"a7d3d389-84a8-4987-804d-caa438d0e7f4\") " pod="openshift-marketplace/redhat-operators-9jppq" Oct 02 02:10:18 crc kubenswrapper[4885]: I1002 02:10:18.984025 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7d3d389-84a8-4987-804d-caa438d0e7f4-utilities\") pod \"redhat-operators-9jppq\" (UID: \"a7d3d389-84a8-4987-804d-caa438d0e7f4\") " pod="openshift-marketplace/redhat-operators-9jppq" Oct 02 02:10:18 crc kubenswrapper[4885]: I1002 02:10:18.984045 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7d3d389-84a8-4987-804d-caa438d0e7f4-catalog-content\") pod \"redhat-operators-9jppq\" (UID: \"a7d3d389-84a8-4987-804d-caa438d0e7f4\") " pod="openshift-marketplace/redhat-operators-9jppq" Oct 02 02:10:19 crc kubenswrapper[4885]: I1002 02:10:19.008139 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8x7f5\" (UniqueName: \"kubernetes.io/projected/a7d3d389-84a8-4987-804d-caa438d0e7f4-kube-api-access-8x7f5\") pod \"redhat-operators-9jppq\" (UID: \"a7d3d389-84a8-4987-804d-caa438d0e7f4\") " pod="openshift-marketplace/redhat-operators-9jppq" Oct 02 02:10:19 crc kubenswrapper[4885]: I1002 02:10:19.089099 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9jppq" Oct 02 02:10:19 crc kubenswrapper[4885]: I1002 02:10:19.589555 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9jppq"] Oct 02 02:10:19 crc kubenswrapper[4885]: I1002 02:10:19.895628 4885 generic.go:334] "Generic (PLEG): container finished" podID="a7d3d389-84a8-4987-804d-caa438d0e7f4" containerID="e5484e67e70a405e6c3c2571c20ca7e9cd6ad947543551ea3036da492a8f4009" exitCode=0 Oct 02 02:10:19 crc kubenswrapper[4885]: I1002 02:10:19.895829 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9jppq" event={"ID":"a7d3d389-84a8-4987-804d-caa438d0e7f4","Type":"ContainerDied","Data":"e5484e67e70a405e6c3c2571c20ca7e9cd6ad947543551ea3036da492a8f4009"} Oct 02 02:10:19 crc kubenswrapper[4885]: I1002 02:10:19.895937 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9jppq" event={"ID":"a7d3d389-84a8-4987-804d-caa438d0e7f4","Type":"ContainerStarted","Data":"47274ae252082559e8e0d61d09e8e74d9c260a33af817d0535d358edb94c975e"} Oct 02 02:10:21 crc kubenswrapper[4885]: I1002 02:10:21.922038 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9jppq" event={"ID":"a7d3d389-84a8-4987-804d-caa438d0e7f4","Type":"ContainerStarted","Data":"e0132936e7e53a507a1f13d028c40ada1fe95e61d671f6aedb3a4441f9b838b2"} Oct 02 02:10:22 crc kubenswrapper[4885]: I1002 02:10:22.957713 4885 generic.go:334] "Generic (PLEG): container finished" podID="a7d3d389-84a8-4987-804d-caa438d0e7f4" containerID="e0132936e7e53a507a1f13d028c40ada1fe95e61d671f6aedb3a4441f9b838b2" exitCode=0 Oct 02 02:10:22 crc kubenswrapper[4885]: I1002 02:10:22.958044 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9jppq" event={"ID":"a7d3d389-84a8-4987-804d-caa438d0e7f4","Type":"ContainerDied","Data":"e0132936e7e53a507a1f13d028c40ada1fe95e61d671f6aedb3a4441f9b838b2"} Oct 02 02:10:24 crc kubenswrapper[4885]: I1002 02:10:24.982929 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9jppq" event={"ID":"a7d3d389-84a8-4987-804d-caa438d0e7f4","Type":"ContainerStarted","Data":"228f62a9ac76391aa2abe1ac750cb1186fd2242c2353f7f7b104ff3e535a58a6"} Oct 02 02:10:25 crc kubenswrapper[4885]: I1002 02:10:25.014567 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-9jppq" podStartSLOduration=2.399647826 podStartE2EDuration="7.014542815s" podCreationTimestamp="2025-10-02 02:10:18 +0000 UTC" firstStartedPulling="2025-10-02 02:10:19.897641971 +0000 UTC m=+1408.709389370" lastFinishedPulling="2025-10-02 02:10:24.51253694 +0000 UTC m=+1413.324284359" observedRunningTime="2025-10-02 02:10:25.007067257 +0000 UTC m=+1413.818814676" watchObservedRunningTime="2025-10-02 02:10:25.014542815 +0000 UTC m=+1413.826290254" Oct 02 02:10:29 crc kubenswrapper[4885]: I1002 02:10:29.089218 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-9jppq" Oct 02 02:10:29 crc kubenswrapper[4885]: I1002 02:10:29.089696 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-9jppq" Oct 02 02:10:30 crc kubenswrapper[4885]: I1002 02:10:30.167896 4885 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-9jppq" podUID="a7d3d389-84a8-4987-804d-caa438d0e7f4" containerName="registry-server" probeResult="failure" output=< Oct 02 02:10:30 crc kubenswrapper[4885]: timeout: failed to connect service ":50051" within 1s Oct 02 02:10:30 crc kubenswrapper[4885]: > Oct 02 02:10:40 crc kubenswrapper[4885]: I1002 02:10:40.159006 4885 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-9jppq" podUID="a7d3d389-84a8-4987-804d-caa438d0e7f4" containerName="registry-server" probeResult="failure" output=< Oct 02 02:10:40 crc kubenswrapper[4885]: timeout: failed to connect service ":50051" within 1s Oct 02 02:10:40 crc kubenswrapper[4885]: > Oct 02 02:10:49 crc kubenswrapper[4885]: I1002 02:10:49.171412 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-9jppq" Oct 02 02:10:49 crc kubenswrapper[4885]: I1002 02:10:49.261769 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-9jppq" Oct 02 02:10:49 crc kubenswrapper[4885]: I1002 02:10:49.949295 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9jppq"] Oct 02 02:10:50 crc kubenswrapper[4885]: I1002 02:10:50.267685 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-9jppq" podUID="a7d3d389-84a8-4987-804d-caa438d0e7f4" containerName="registry-server" containerID="cri-o://228f62a9ac76391aa2abe1ac750cb1186fd2242c2353f7f7b104ff3e535a58a6" gracePeriod=2 Oct 02 02:10:50 crc kubenswrapper[4885]: I1002 02:10:50.795298 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9jppq" Oct 02 02:10:50 crc kubenswrapper[4885]: I1002 02:10:50.902700 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7d3d389-84a8-4987-804d-caa438d0e7f4-catalog-content\") pod \"a7d3d389-84a8-4987-804d-caa438d0e7f4\" (UID: \"a7d3d389-84a8-4987-804d-caa438d0e7f4\") " Oct 02 02:10:50 crc kubenswrapper[4885]: I1002 02:10:50.902757 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8x7f5\" (UniqueName: \"kubernetes.io/projected/a7d3d389-84a8-4987-804d-caa438d0e7f4-kube-api-access-8x7f5\") pod \"a7d3d389-84a8-4987-804d-caa438d0e7f4\" (UID: \"a7d3d389-84a8-4987-804d-caa438d0e7f4\") " Oct 02 02:10:50 crc kubenswrapper[4885]: I1002 02:10:50.902881 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7d3d389-84a8-4987-804d-caa438d0e7f4-utilities\") pod \"a7d3d389-84a8-4987-804d-caa438d0e7f4\" (UID: \"a7d3d389-84a8-4987-804d-caa438d0e7f4\") " Oct 02 02:10:50 crc kubenswrapper[4885]: I1002 02:10:50.903784 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a7d3d389-84a8-4987-804d-caa438d0e7f4-utilities" (OuterVolumeSpecName: "utilities") pod "a7d3d389-84a8-4987-804d-caa438d0e7f4" (UID: "a7d3d389-84a8-4987-804d-caa438d0e7f4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:10:50 crc kubenswrapper[4885]: I1002 02:10:50.909471 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7d3d389-84a8-4987-804d-caa438d0e7f4-kube-api-access-8x7f5" (OuterVolumeSpecName: "kube-api-access-8x7f5") pod "a7d3d389-84a8-4987-804d-caa438d0e7f4" (UID: "a7d3d389-84a8-4987-804d-caa438d0e7f4"). InnerVolumeSpecName "kube-api-access-8x7f5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:10:50 crc kubenswrapper[4885]: I1002 02:10:50.998430 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a7d3d389-84a8-4987-804d-caa438d0e7f4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a7d3d389-84a8-4987-804d-caa438d0e7f4" (UID: "a7d3d389-84a8-4987-804d-caa438d0e7f4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:10:51 crc kubenswrapper[4885]: I1002 02:10:51.006045 4885 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7d3d389-84a8-4987-804d-caa438d0e7f4-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 02:10:51 crc kubenswrapper[4885]: I1002 02:10:51.006248 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8x7f5\" (UniqueName: \"kubernetes.io/projected/a7d3d389-84a8-4987-804d-caa438d0e7f4-kube-api-access-8x7f5\") on node \"crc\" DevicePath \"\"" Oct 02 02:10:51 crc kubenswrapper[4885]: I1002 02:10:51.006462 4885 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7d3d389-84a8-4987-804d-caa438d0e7f4-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 02:10:51 crc kubenswrapper[4885]: I1002 02:10:51.283672 4885 generic.go:334] "Generic (PLEG): container finished" podID="a7d3d389-84a8-4987-804d-caa438d0e7f4" containerID="228f62a9ac76391aa2abe1ac750cb1186fd2242c2353f7f7b104ff3e535a58a6" exitCode=0 Oct 02 02:10:51 crc kubenswrapper[4885]: I1002 02:10:51.283737 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9jppq" Oct 02 02:10:51 crc kubenswrapper[4885]: I1002 02:10:51.283769 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9jppq" event={"ID":"a7d3d389-84a8-4987-804d-caa438d0e7f4","Type":"ContainerDied","Data":"228f62a9ac76391aa2abe1ac750cb1186fd2242c2353f7f7b104ff3e535a58a6"} Oct 02 02:10:51 crc kubenswrapper[4885]: I1002 02:10:51.284289 4885 scope.go:117] "RemoveContainer" containerID="228f62a9ac76391aa2abe1ac750cb1186fd2242c2353f7f7b104ff3e535a58a6" Oct 02 02:10:51 crc kubenswrapper[4885]: I1002 02:10:51.284284 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9jppq" event={"ID":"a7d3d389-84a8-4987-804d-caa438d0e7f4","Type":"ContainerDied","Data":"47274ae252082559e8e0d61d09e8e74d9c260a33af817d0535d358edb94c975e"} Oct 02 02:10:51 crc kubenswrapper[4885]: I1002 02:10:51.321211 4885 scope.go:117] "RemoveContainer" containerID="e0132936e7e53a507a1f13d028c40ada1fe95e61d671f6aedb3a4441f9b838b2" Oct 02 02:10:51 crc kubenswrapper[4885]: I1002 02:10:51.332127 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9jppq"] Oct 02 02:10:51 crc kubenswrapper[4885]: I1002 02:10:51.341999 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-9jppq"] Oct 02 02:10:51 crc kubenswrapper[4885]: I1002 02:10:51.354018 4885 scope.go:117] "RemoveContainer" containerID="e5484e67e70a405e6c3c2571c20ca7e9cd6ad947543551ea3036da492a8f4009" Oct 02 02:10:51 crc kubenswrapper[4885]: I1002 02:10:51.417977 4885 scope.go:117] "RemoveContainer" containerID="228f62a9ac76391aa2abe1ac750cb1186fd2242c2353f7f7b104ff3e535a58a6" Oct 02 02:10:51 crc kubenswrapper[4885]: E1002 02:10:51.418555 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"228f62a9ac76391aa2abe1ac750cb1186fd2242c2353f7f7b104ff3e535a58a6\": container with ID starting with 228f62a9ac76391aa2abe1ac750cb1186fd2242c2353f7f7b104ff3e535a58a6 not found: ID does not exist" containerID="228f62a9ac76391aa2abe1ac750cb1186fd2242c2353f7f7b104ff3e535a58a6" Oct 02 02:10:51 crc kubenswrapper[4885]: I1002 02:10:51.418612 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"228f62a9ac76391aa2abe1ac750cb1186fd2242c2353f7f7b104ff3e535a58a6"} err="failed to get container status \"228f62a9ac76391aa2abe1ac750cb1186fd2242c2353f7f7b104ff3e535a58a6\": rpc error: code = NotFound desc = could not find container \"228f62a9ac76391aa2abe1ac750cb1186fd2242c2353f7f7b104ff3e535a58a6\": container with ID starting with 228f62a9ac76391aa2abe1ac750cb1186fd2242c2353f7f7b104ff3e535a58a6 not found: ID does not exist" Oct 02 02:10:51 crc kubenswrapper[4885]: I1002 02:10:51.418646 4885 scope.go:117] "RemoveContainer" containerID="e0132936e7e53a507a1f13d028c40ada1fe95e61d671f6aedb3a4441f9b838b2" Oct 02 02:10:51 crc kubenswrapper[4885]: E1002 02:10:51.419136 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e0132936e7e53a507a1f13d028c40ada1fe95e61d671f6aedb3a4441f9b838b2\": container with ID starting with e0132936e7e53a507a1f13d028c40ada1fe95e61d671f6aedb3a4441f9b838b2 not found: ID does not exist" containerID="e0132936e7e53a507a1f13d028c40ada1fe95e61d671f6aedb3a4441f9b838b2" Oct 02 02:10:51 crc kubenswrapper[4885]: I1002 02:10:51.419198 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e0132936e7e53a507a1f13d028c40ada1fe95e61d671f6aedb3a4441f9b838b2"} err="failed to get container status \"e0132936e7e53a507a1f13d028c40ada1fe95e61d671f6aedb3a4441f9b838b2\": rpc error: code = NotFound desc = could not find container \"e0132936e7e53a507a1f13d028c40ada1fe95e61d671f6aedb3a4441f9b838b2\": container with ID starting with e0132936e7e53a507a1f13d028c40ada1fe95e61d671f6aedb3a4441f9b838b2 not found: ID does not exist" Oct 02 02:10:51 crc kubenswrapper[4885]: I1002 02:10:51.419243 4885 scope.go:117] "RemoveContainer" containerID="e5484e67e70a405e6c3c2571c20ca7e9cd6ad947543551ea3036da492a8f4009" Oct 02 02:10:51 crc kubenswrapper[4885]: E1002 02:10:51.419751 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e5484e67e70a405e6c3c2571c20ca7e9cd6ad947543551ea3036da492a8f4009\": container with ID starting with e5484e67e70a405e6c3c2571c20ca7e9cd6ad947543551ea3036da492a8f4009 not found: ID does not exist" containerID="e5484e67e70a405e6c3c2571c20ca7e9cd6ad947543551ea3036da492a8f4009" Oct 02 02:10:51 crc kubenswrapper[4885]: I1002 02:10:51.419809 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e5484e67e70a405e6c3c2571c20ca7e9cd6ad947543551ea3036da492a8f4009"} err="failed to get container status \"e5484e67e70a405e6c3c2571c20ca7e9cd6ad947543551ea3036da492a8f4009\": rpc error: code = NotFound desc = could not find container \"e5484e67e70a405e6c3c2571c20ca7e9cd6ad947543551ea3036da492a8f4009\": container with ID starting with e5484e67e70a405e6c3c2571c20ca7e9cd6ad947543551ea3036da492a8f4009 not found: ID does not exist" Oct 02 02:10:52 crc kubenswrapper[4885]: I1002 02:10:52.062640 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a7d3d389-84a8-4987-804d-caa438d0e7f4" path="/var/lib/kubelet/pods/a7d3d389-84a8-4987-804d-caa438d0e7f4/volumes" Oct 02 02:11:43 crc kubenswrapper[4885]: I1002 02:11:43.266014 4885 patch_prober.go:28] interesting pod/machine-config-daemon-rttx8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 02:11:43 crc kubenswrapper[4885]: I1002 02:11:43.266881 4885 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 02:11:51 crc kubenswrapper[4885]: I1002 02:11:51.243189 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-lc8h9"] Oct 02 02:11:51 crc kubenswrapper[4885]: E1002 02:11:51.244220 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7d3d389-84a8-4987-804d-caa438d0e7f4" containerName="registry-server" Oct 02 02:11:51 crc kubenswrapper[4885]: I1002 02:11:51.244237 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7d3d389-84a8-4987-804d-caa438d0e7f4" containerName="registry-server" Oct 02 02:11:51 crc kubenswrapper[4885]: E1002 02:11:51.244276 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7d3d389-84a8-4987-804d-caa438d0e7f4" containerName="extract-content" Oct 02 02:11:51 crc kubenswrapper[4885]: I1002 02:11:51.244285 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7d3d389-84a8-4987-804d-caa438d0e7f4" containerName="extract-content" Oct 02 02:11:51 crc kubenswrapper[4885]: E1002 02:11:51.244301 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7d3d389-84a8-4987-804d-caa438d0e7f4" containerName="extract-utilities" Oct 02 02:11:51 crc kubenswrapper[4885]: I1002 02:11:51.244310 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7d3d389-84a8-4987-804d-caa438d0e7f4" containerName="extract-utilities" Oct 02 02:11:51 crc kubenswrapper[4885]: I1002 02:11:51.244635 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7d3d389-84a8-4987-804d-caa438d0e7f4" containerName="registry-server" Oct 02 02:11:51 crc kubenswrapper[4885]: I1002 02:11:51.246449 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lc8h9" Oct 02 02:11:51 crc kubenswrapper[4885]: I1002 02:11:51.270028 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-lc8h9"] Oct 02 02:11:51 crc kubenswrapper[4885]: I1002 02:11:51.374878 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a78fbf7b-c9e6-42b2-b163-abebcf824f12-catalog-content\") pod \"community-operators-lc8h9\" (UID: \"a78fbf7b-c9e6-42b2-b163-abebcf824f12\") " pod="openshift-marketplace/community-operators-lc8h9" Oct 02 02:11:51 crc kubenswrapper[4885]: I1002 02:11:51.375631 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4dlt8\" (UniqueName: \"kubernetes.io/projected/a78fbf7b-c9e6-42b2-b163-abebcf824f12-kube-api-access-4dlt8\") pod \"community-operators-lc8h9\" (UID: \"a78fbf7b-c9e6-42b2-b163-abebcf824f12\") " pod="openshift-marketplace/community-operators-lc8h9" Oct 02 02:11:51 crc kubenswrapper[4885]: I1002 02:11:51.375952 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a78fbf7b-c9e6-42b2-b163-abebcf824f12-utilities\") pod \"community-operators-lc8h9\" (UID: \"a78fbf7b-c9e6-42b2-b163-abebcf824f12\") " pod="openshift-marketplace/community-operators-lc8h9" Oct 02 02:11:51 crc kubenswrapper[4885]: I1002 02:11:51.477926 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a78fbf7b-c9e6-42b2-b163-abebcf824f12-catalog-content\") pod \"community-operators-lc8h9\" (UID: \"a78fbf7b-c9e6-42b2-b163-abebcf824f12\") " pod="openshift-marketplace/community-operators-lc8h9" Oct 02 02:11:51 crc kubenswrapper[4885]: I1002 02:11:51.478377 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4dlt8\" (UniqueName: \"kubernetes.io/projected/a78fbf7b-c9e6-42b2-b163-abebcf824f12-kube-api-access-4dlt8\") pod \"community-operators-lc8h9\" (UID: \"a78fbf7b-c9e6-42b2-b163-abebcf824f12\") " pod="openshift-marketplace/community-operators-lc8h9" Oct 02 02:11:51 crc kubenswrapper[4885]: I1002 02:11:51.478459 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a78fbf7b-c9e6-42b2-b163-abebcf824f12-catalog-content\") pod \"community-operators-lc8h9\" (UID: \"a78fbf7b-c9e6-42b2-b163-abebcf824f12\") " pod="openshift-marketplace/community-operators-lc8h9" Oct 02 02:11:51 crc kubenswrapper[4885]: I1002 02:11:51.478623 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a78fbf7b-c9e6-42b2-b163-abebcf824f12-utilities\") pod \"community-operators-lc8h9\" (UID: \"a78fbf7b-c9e6-42b2-b163-abebcf824f12\") " pod="openshift-marketplace/community-operators-lc8h9" Oct 02 02:11:51 crc kubenswrapper[4885]: I1002 02:11:51.478854 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a78fbf7b-c9e6-42b2-b163-abebcf824f12-utilities\") pod \"community-operators-lc8h9\" (UID: \"a78fbf7b-c9e6-42b2-b163-abebcf824f12\") " pod="openshift-marketplace/community-operators-lc8h9" Oct 02 02:11:51 crc kubenswrapper[4885]: I1002 02:11:51.504528 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4dlt8\" (UniqueName: \"kubernetes.io/projected/a78fbf7b-c9e6-42b2-b163-abebcf824f12-kube-api-access-4dlt8\") pod \"community-operators-lc8h9\" (UID: \"a78fbf7b-c9e6-42b2-b163-abebcf824f12\") " pod="openshift-marketplace/community-operators-lc8h9" Oct 02 02:11:51 crc kubenswrapper[4885]: I1002 02:11:51.611567 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lc8h9" Oct 02 02:11:52 crc kubenswrapper[4885]: I1002 02:11:52.109451 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-lc8h9"] Oct 02 02:11:53 crc kubenswrapper[4885]: I1002 02:11:53.066008 4885 generic.go:334] "Generic (PLEG): container finished" podID="a78fbf7b-c9e6-42b2-b163-abebcf824f12" containerID="1fbc5f1be01a2bbc4fbebca3ea975a55f3ff350a5d3080699c859d4086913428" exitCode=0 Oct 02 02:11:53 crc kubenswrapper[4885]: I1002 02:11:53.066069 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lc8h9" event={"ID":"a78fbf7b-c9e6-42b2-b163-abebcf824f12","Type":"ContainerDied","Data":"1fbc5f1be01a2bbc4fbebca3ea975a55f3ff350a5d3080699c859d4086913428"} Oct 02 02:11:53 crc kubenswrapper[4885]: I1002 02:11:53.066539 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lc8h9" event={"ID":"a78fbf7b-c9e6-42b2-b163-abebcf824f12","Type":"ContainerStarted","Data":"4c1e387b1c27a2e563b4b1c7eb67b05d6641b5dcd47bad2b85be54750dab9eb6"} Oct 02 02:11:53 crc kubenswrapper[4885]: I1002 02:11:53.070798 4885 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 02:11:55 crc kubenswrapper[4885]: I1002 02:11:55.091829 4885 generic.go:334] "Generic (PLEG): container finished" podID="a78fbf7b-c9e6-42b2-b163-abebcf824f12" containerID="67b8c7d06b744652c0d442a1620d72cc5616cb76178b9a866a4f27092a8c4c45" exitCode=0 Oct 02 02:11:55 crc kubenswrapper[4885]: I1002 02:11:55.091899 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lc8h9" event={"ID":"a78fbf7b-c9e6-42b2-b163-abebcf824f12","Type":"ContainerDied","Data":"67b8c7d06b744652c0d442a1620d72cc5616cb76178b9a866a4f27092a8c4c45"} Oct 02 02:11:56 crc kubenswrapper[4885]: I1002 02:11:56.103291 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lc8h9" event={"ID":"a78fbf7b-c9e6-42b2-b163-abebcf824f12","Type":"ContainerStarted","Data":"511e51e5459f28045367636c2683ab581011e794ab97b678c24ee3cabaf8b68c"} Oct 02 02:11:56 crc kubenswrapper[4885]: I1002 02:11:56.127658 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-lc8h9" podStartSLOduration=2.5399821830000002 podStartE2EDuration="5.127640639s" podCreationTimestamp="2025-10-02 02:11:51 +0000 UTC" firstStartedPulling="2025-10-02 02:11:53.070425596 +0000 UTC m=+1501.882173015" lastFinishedPulling="2025-10-02 02:11:55.658084072 +0000 UTC m=+1504.469831471" observedRunningTime="2025-10-02 02:11:56.124857374 +0000 UTC m=+1504.936604773" watchObservedRunningTime="2025-10-02 02:11:56.127640639 +0000 UTC m=+1504.939388028" Oct 02 02:12:01 crc kubenswrapper[4885]: I1002 02:12:01.611958 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-lc8h9" Oct 02 02:12:01 crc kubenswrapper[4885]: I1002 02:12:01.612463 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-lc8h9" Oct 02 02:12:01 crc kubenswrapper[4885]: I1002 02:12:01.692416 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-lc8h9" Oct 02 02:12:02 crc kubenswrapper[4885]: I1002 02:12:02.233069 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-lc8h9" Oct 02 02:12:02 crc kubenswrapper[4885]: I1002 02:12:02.300650 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-lc8h9"] Oct 02 02:12:04 crc kubenswrapper[4885]: I1002 02:12:04.197563 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-lc8h9" podUID="a78fbf7b-c9e6-42b2-b163-abebcf824f12" containerName="registry-server" containerID="cri-o://511e51e5459f28045367636c2683ab581011e794ab97b678c24ee3cabaf8b68c" gracePeriod=2 Oct 02 02:12:04 crc kubenswrapper[4885]: I1002 02:12:04.671701 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lc8h9" Oct 02 02:12:04 crc kubenswrapper[4885]: I1002 02:12:04.757771 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4dlt8\" (UniqueName: \"kubernetes.io/projected/a78fbf7b-c9e6-42b2-b163-abebcf824f12-kube-api-access-4dlt8\") pod \"a78fbf7b-c9e6-42b2-b163-abebcf824f12\" (UID: \"a78fbf7b-c9e6-42b2-b163-abebcf824f12\") " Oct 02 02:12:04 crc kubenswrapper[4885]: I1002 02:12:04.758193 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a78fbf7b-c9e6-42b2-b163-abebcf824f12-catalog-content\") pod \"a78fbf7b-c9e6-42b2-b163-abebcf824f12\" (UID: \"a78fbf7b-c9e6-42b2-b163-abebcf824f12\") " Oct 02 02:12:04 crc kubenswrapper[4885]: I1002 02:12:04.759423 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a78fbf7b-c9e6-42b2-b163-abebcf824f12-utilities\") pod \"a78fbf7b-c9e6-42b2-b163-abebcf824f12\" (UID: \"a78fbf7b-c9e6-42b2-b163-abebcf824f12\") " Oct 02 02:12:04 crc kubenswrapper[4885]: I1002 02:12:04.765163 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a78fbf7b-c9e6-42b2-b163-abebcf824f12-utilities" (OuterVolumeSpecName: "utilities") pod "a78fbf7b-c9e6-42b2-b163-abebcf824f12" (UID: "a78fbf7b-c9e6-42b2-b163-abebcf824f12"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:12:04 crc kubenswrapper[4885]: I1002 02:12:04.765302 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a78fbf7b-c9e6-42b2-b163-abebcf824f12-kube-api-access-4dlt8" (OuterVolumeSpecName: "kube-api-access-4dlt8") pod "a78fbf7b-c9e6-42b2-b163-abebcf824f12" (UID: "a78fbf7b-c9e6-42b2-b163-abebcf824f12"). InnerVolumeSpecName "kube-api-access-4dlt8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:12:04 crc kubenswrapper[4885]: I1002 02:12:04.866300 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4dlt8\" (UniqueName: \"kubernetes.io/projected/a78fbf7b-c9e6-42b2-b163-abebcf824f12-kube-api-access-4dlt8\") on node \"crc\" DevicePath \"\"" Oct 02 02:12:04 crc kubenswrapper[4885]: I1002 02:12:04.866326 4885 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a78fbf7b-c9e6-42b2-b163-abebcf824f12-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 02:12:05 crc kubenswrapper[4885]: I1002 02:12:05.209783 4885 generic.go:334] "Generic (PLEG): container finished" podID="a78fbf7b-c9e6-42b2-b163-abebcf824f12" containerID="511e51e5459f28045367636c2683ab581011e794ab97b678c24ee3cabaf8b68c" exitCode=0 Oct 02 02:12:05 crc kubenswrapper[4885]: I1002 02:12:05.209843 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lc8h9" event={"ID":"a78fbf7b-c9e6-42b2-b163-abebcf824f12","Type":"ContainerDied","Data":"511e51e5459f28045367636c2683ab581011e794ab97b678c24ee3cabaf8b68c"} Oct 02 02:12:05 crc kubenswrapper[4885]: I1002 02:12:05.209869 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lc8h9" Oct 02 02:12:05 crc kubenswrapper[4885]: I1002 02:12:05.209893 4885 scope.go:117] "RemoveContainer" containerID="511e51e5459f28045367636c2683ab581011e794ab97b678c24ee3cabaf8b68c" Oct 02 02:12:05 crc kubenswrapper[4885]: I1002 02:12:05.209879 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lc8h9" event={"ID":"a78fbf7b-c9e6-42b2-b163-abebcf824f12","Type":"ContainerDied","Data":"4c1e387b1c27a2e563b4b1c7eb67b05d6641b5dcd47bad2b85be54750dab9eb6"} Oct 02 02:12:05 crc kubenswrapper[4885]: I1002 02:12:05.250054 4885 scope.go:117] "RemoveContainer" containerID="67b8c7d06b744652c0d442a1620d72cc5616cb76178b9a866a4f27092a8c4c45" Oct 02 02:12:05 crc kubenswrapper[4885]: I1002 02:12:05.281797 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a78fbf7b-c9e6-42b2-b163-abebcf824f12-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a78fbf7b-c9e6-42b2-b163-abebcf824f12" (UID: "a78fbf7b-c9e6-42b2-b163-abebcf824f12"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:12:05 crc kubenswrapper[4885]: I1002 02:12:05.293778 4885 scope.go:117] "RemoveContainer" containerID="1fbc5f1be01a2bbc4fbebca3ea975a55f3ff350a5d3080699c859d4086913428" Oct 02 02:12:05 crc kubenswrapper[4885]: I1002 02:12:05.356493 4885 scope.go:117] "RemoveContainer" containerID="511e51e5459f28045367636c2683ab581011e794ab97b678c24ee3cabaf8b68c" Oct 02 02:12:05 crc kubenswrapper[4885]: E1002 02:12:05.364741 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"511e51e5459f28045367636c2683ab581011e794ab97b678c24ee3cabaf8b68c\": container with ID starting with 511e51e5459f28045367636c2683ab581011e794ab97b678c24ee3cabaf8b68c not found: ID does not exist" containerID="511e51e5459f28045367636c2683ab581011e794ab97b678c24ee3cabaf8b68c" Oct 02 02:12:05 crc kubenswrapper[4885]: I1002 02:12:05.365042 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"511e51e5459f28045367636c2683ab581011e794ab97b678c24ee3cabaf8b68c"} err="failed to get container status \"511e51e5459f28045367636c2683ab581011e794ab97b678c24ee3cabaf8b68c\": rpc error: code = NotFound desc = could not find container \"511e51e5459f28045367636c2683ab581011e794ab97b678c24ee3cabaf8b68c\": container with ID starting with 511e51e5459f28045367636c2683ab581011e794ab97b678c24ee3cabaf8b68c not found: ID does not exist" Oct 02 02:12:05 crc kubenswrapper[4885]: I1002 02:12:05.365100 4885 scope.go:117] "RemoveContainer" containerID="67b8c7d06b744652c0d442a1620d72cc5616cb76178b9a866a4f27092a8c4c45" Oct 02 02:12:05 crc kubenswrapper[4885]: E1002 02:12:05.366192 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"67b8c7d06b744652c0d442a1620d72cc5616cb76178b9a866a4f27092a8c4c45\": container with ID starting with 67b8c7d06b744652c0d442a1620d72cc5616cb76178b9a866a4f27092a8c4c45 not found: ID does not exist" containerID="67b8c7d06b744652c0d442a1620d72cc5616cb76178b9a866a4f27092a8c4c45" Oct 02 02:12:05 crc kubenswrapper[4885]: I1002 02:12:05.366314 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"67b8c7d06b744652c0d442a1620d72cc5616cb76178b9a866a4f27092a8c4c45"} err="failed to get container status \"67b8c7d06b744652c0d442a1620d72cc5616cb76178b9a866a4f27092a8c4c45\": rpc error: code = NotFound desc = could not find container \"67b8c7d06b744652c0d442a1620d72cc5616cb76178b9a866a4f27092a8c4c45\": container with ID starting with 67b8c7d06b744652c0d442a1620d72cc5616cb76178b9a866a4f27092a8c4c45 not found: ID does not exist" Oct 02 02:12:05 crc kubenswrapper[4885]: I1002 02:12:05.366358 4885 scope.go:117] "RemoveContainer" containerID="1fbc5f1be01a2bbc4fbebca3ea975a55f3ff350a5d3080699c859d4086913428" Oct 02 02:12:05 crc kubenswrapper[4885]: E1002 02:12:05.369616 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1fbc5f1be01a2bbc4fbebca3ea975a55f3ff350a5d3080699c859d4086913428\": container with ID starting with 1fbc5f1be01a2bbc4fbebca3ea975a55f3ff350a5d3080699c859d4086913428 not found: ID does not exist" containerID="1fbc5f1be01a2bbc4fbebca3ea975a55f3ff350a5d3080699c859d4086913428" Oct 02 02:12:05 crc kubenswrapper[4885]: I1002 02:12:05.369844 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1fbc5f1be01a2bbc4fbebca3ea975a55f3ff350a5d3080699c859d4086913428"} err="failed to get container status \"1fbc5f1be01a2bbc4fbebca3ea975a55f3ff350a5d3080699c859d4086913428\": rpc error: code = NotFound desc = could not find container \"1fbc5f1be01a2bbc4fbebca3ea975a55f3ff350a5d3080699c859d4086913428\": container with ID starting with 1fbc5f1be01a2bbc4fbebca3ea975a55f3ff350a5d3080699c859d4086913428 not found: ID does not exist" Oct 02 02:12:05 crc kubenswrapper[4885]: I1002 02:12:05.377823 4885 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a78fbf7b-c9e6-42b2-b163-abebcf824f12-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 02:12:05 crc kubenswrapper[4885]: I1002 02:12:05.569343 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-lc8h9"] Oct 02 02:12:05 crc kubenswrapper[4885]: I1002 02:12:05.583015 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-lc8h9"] Oct 02 02:12:06 crc kubenswrapper[4885]: I1002 02:12:06.067355 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a78fbf7b-c9e6-42b2-b163-abebcf824f12" path="/var/lib/kubelet/pods/a78fbf7b-c9e6-42b2-b163-abebcf824f12/volumes" Oct 02 02:12:06 crc kubenswrapper[4885]: I1002 02:12:06.224945 4885 generic.go:334] "Generic (PLEG): container finished" podID="8cba1665-9cca-44ef-a790-5cd9b174e726" containerID="850533a0036969f7e2fc59ff244f40d27f7b95486b11d451e3b3b9b0a609b7d8" exitCode=0 Oct 02 02:12:06 crc kubenswrapper[4885]: I1002 02:12:06.224989 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rgbbj" event={"ID":"8cba1665-9cca-44ef-a790-5cd9b174e726","Type":"ContainerDied","Data":"850533a0036969f7e2fc59ff244f40d27f7b95486b11d451e3b3b9b0a609b7d8"} Oct 02 02:12:07 crc kubenswrapper[4885]: I1002 02:12:07.738902 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rgbbj" Oct 02 02:12:07 crc kubenswrapper[4885]: I1002 02:12:07.820840 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8cba1665-9cca-44ef-a790-5cd9b174e726-inventory\") pod \"8cba1665-9cca-44ef-a790-5cd9b174e726\" (UID: \"8cba1665-9cca-44ef-a790-5cd9b174e726\") " Oct 02 02:12:07 crc kubenswrapper[4885]: I1002 02:12:07.821655 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8cba1665-9cca-44ef-a790-5cd9b174e726-ssh-key\") pod \"8cba1665-9cca-44ef-a790-5cd9b174e726\" (UID: \"8cba1665-9cca-44ef-a790-5cd9b174e726\") " Oct 02 02:12:07 crc kubenswrapper[4885]: I1002 02:12:07.821777 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jptxh\" (UniqueName: \"kubernetes.io/projected/8cba1665-9cca-44ef-a790-5cd9b174e726-kube-api-access-jptxh\") pod \"8cba1665-9cca-44ef-a790-5cd9b174e726\" (UID: \"8cba1665-9cca-44ef-a790-5cd9b174e726\") " Oct 02 02:12:07 crc kubenswrapper[4885]: I1002 02:12:07.821941 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8cba1665-9cca-44ef-a790-5cd9b174e726-bootstrap-combined-ca-bundle\") pod \"8cba1665-9cca-44ef-a790-5cd9b174e726\" (UID: \"8cba1665-9cca-44ef-a790-5cd9b174e726\") " Oct 02 02:12:07 crc kubenswrapper[4885]: I1002 02:12:07.847515 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cba1665-9cca-44ef-a790-5cd9b174e726-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "8cba1665-9cca-44ef-a790-5cd9b174e726" (UID: "8cba1665-9cca-44ef-a790-5cd9b174e726"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:12:07 crc kubenswrapper[4885]: I1002 02:12:07.848412 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cba1665-9cca-44ef-a790-5cd9b174e726-kube-api-access-jptxh" (OuterVolumeSpecName: "kube-api-access-jptxh") pod "8cba1665-9cca-44ef-a790-5cd9b174e726" (UID: "8cba1665-9cca-44ef-a790-5cd9b174e726"). InnerVolumeSpecName "kube-api-access-jptxh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:12:07 crc kubenswrapper[4885]: I1002 02:12:07.851605 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cba1665-9cca-44ef-a790-5cd9b174e726-inventory" (OuterVolumeSpecName: "inventory") pod "8cba1665-9cca-44ef-a790-5cd9b174e726" (UID: "8cba1665-9cca-44ef-a790-5cd9b174e726"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:12:07 crc kubenswrapper[4885]: I1002 02:12:07.851981 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cba1665-9cca-44ef-a790-5cd9b174e726-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "8cba1665-9cca-44ef-a790-5cd9b174e726" (UID: "8cba1665-9cca-44ef-a790-5cd9b174e726"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:12:07 crc kubenswrapper[4885]: I1002 02:12:07.924064 4885 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8cba1665-9cca-44ef-a790-5cd9b174e726-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 02:12:07 crc kubenswrapper[4885]: I1002 02:12:07.924109 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jptxh\" (UniqueName: \"kubernetes.io/projected/8cba1665-9cca-44ef-a790-5cd9b174e726-kube-api-access-jptxh\") on node \"crc\" DevicePath \"\"" Oct 02 02:12:07 crc kubenswrapper[4885]: I1002 02:12:07.924126 4885 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8cba1665-9cca-44ef-a790-5cd9b174e726-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:12:07 crc kubenswrapper[4885]: I1002 02:12:07.924137 4885 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8cba1665-9cca-44ef-a790-5cd9b174e726-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 02:12:08 crc kubenswrapper[4885]: I1002 02:12:08.306236 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rgbbj" event={"ID":"8cba1665-9cca-44ef-a790-5cd9b174e726","Type":"ContainerDied","Data":"2bcb2f37bef24de52ba79bcbbc8f3afad5d39f45abaea14dcee0350379cf6fcd"} Oct 02 02:12:08 crc kubenswrapper[4885]: I1002 02:12:08.306584 4885 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2bcb2f37bef24de52ba79bcbbc8f3afad5d39f45abaea14dcee0350379cf6fcd" Oct 02 02:12:08 crc kubenswrapper[4885]: I1002 02:12:08.306374 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rgbbj" Oct 02 02:12:08 crc kubenswrapper[4885]: I1002 02:12:08.336423 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-c2g9p"] Oct 02 02:12:08 crc kubenswrapper[4885]: E1002 02:12:08.336961 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a78fbf7b-c9e6-42b2-b163-abebcf824f12" containerName="registry-server" Oct 02 02:12:08 crc kubenswrapper[4885]: I1002 02:12:08.336983 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="a78fbf7b-c9e6-42b2-b163-abebcf824f12" containerName="registry-server" Oct 02 02:12:08 crc kubenswrapper[4885]: E1002 02:12:08.337004 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a78fbf7b-c9e6-42b2-b163-abebcf824f12" containerName="extract-utilities" Oct 02 02:12:08 crc kubenswrapper[4885]: I1002 02:12:08.337013 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="a78fbf7b-c9e6-42b2-b163-abebcf824f12" containerName="extract-utilities" Oct 02 02:12:08 crc kubenswrapper[4885]: E1002 02:12:08.337035 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a78fbf7b-c9e6-42b2-b163-abebcf824f12" containerName="extract-content" Oct 02 02:12:08 crc kubenswrapper[4885]: I1002 02:12:08.337043 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="a78fbf7b-c9e6-42b2-b163-abebcf824f12" containerName="extract-content" Oct 02 02:12:08 crc kubenswrapper[4885]: E1002 02:12:08.337066 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8cba1665-9cca-44ef-a790-5cd9b174e726" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 02 02:12:08 crc kubenswrapper[4885]: I1002 02:12:08.337074 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="8cba1665-9cca-44ef-a790-5cd9b174e726" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 02 02:12:08 crc kubenswrapper[4885]: I1002 02:12:08.337339 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="a78fbf7b-c9e6-42b2-b163-abebcf824f12" containerName="registry-server" Oct 02 02:12:08 crc kubenswrapper[4885]: I1002 02:12:08.337383 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="8cba1665-9cca-44ef-a790-5cd9b174e726" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 02 02:12:08 crc kubenswrapper[4885]: I1002 02:12:08.338394 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-c2g9p" Oct 02 02:12:08 crc kubenswrapper[4885]: I1002 02:12:08.340947 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 02:12:08 crc kubenswrapper[4885]: I1002 02:12:08.341811 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 02:12:08 crc kubenswrapper[4885]: I1002 02:12:08.342230 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-f8tvb" Oct 02 02:12:08 crc kubenswrapper[4885]: I1002 02:12:08.343057 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 02:12:08 crc kubenswrapper[4885]: I1002 02:12:08.346413 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-c2g9p"] Oct 02 02:12:08 crc kubenswrapper[4885]: I1002 02:12:08.535545 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c1ca4902-a7b7-4cfa-a8df-5953754142e0-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-c2g9p\" (UID: \"c1ca4902-a7b7-4cfa-a8df-5953754142e0\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-c2g9p" Oct 02 02:12:08 crc kubenswrapper[4885]: I1002 02:12:08.535648 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c1ca4902-a7b7-4cfa-a8df-5953754142e0-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-c2g9p\" (UID: \"c1ca4902-a7b7-4cfa-a8df-5953754142e0\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-c2g9p" Oct 02 02:12:08 crc kubenswrapper[4885]: I1002 02:12:08.536478 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-klpg2\" (UniqueName: \"kubernetes.io/projected/c1ca4902-a7b7-4cfa-a8df-5953754142e0-kube-api-access-klpg2\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-c2g9p\" (UID: \"c1ca4902-a7b7-4cfa-a8df-5953754142e0\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-c2g9p" Oct 02 02:12:08 crc kubenswrapper[4885]: I1002 02:12:08.639115 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c1ca4902-a7b7-4cfa-a8df-5953754142e0-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-c2g9p\" (UID: \"c1ca4902-a7b7-4cfa-a8df-5953754142e0\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-c2g9p" Oct 02 02:12:08 crc kubenswrapper[4885]: I1002 02:12:08.639228 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c1ca4902-a7b7-4cfa-a8df-5953754142e0-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-c2g9p\" (UID: \"c1ca4902-a7b7-4cfa-a8df-5953754142e0\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-c2g9p" Oct 02 02:12:08 crc kubenswrapper[4885]: I1002 02:12:08.639342 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-klpg2\" (UniqueName: \"kubernetes.io/projected/c1ca4902-a7b7-4cfa-a8df-5953754142e0-kube-api-access-klpg2\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-c2g9p\" (UID: \"c1ca4902-a7b7-4cfa-a8df-5953754142e0\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-c2g9p" Oct 02 02:12:08 crc kubenswrapper[4885]: I1002 02:12:08.643077 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c1ca4902-a7b7-4cfa-a8df-5953754142e0-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-c2g9p\" (UID: \"c1ca4902-a7b7-4cfa-a8df-5953754142e0\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-c2g9p" Oct 02 02:12:08 crc kubenswrapper[4885]: I1002 02:12:08.645797 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c1ca4902-a7b7-4cfa-a8df-5953754142e0-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-c2g9p\" (UID: \"c1ca4902-a7b7-4cfa-a8df-5953754142e0\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-c2g9p" Oct 02 02:12:08 crc kubenswrapper[4885]: I1002 02:12:08.682006 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-klpg2\" (UniqueName: \"kubernetes.io/projected/c1ca4902-a7b7-4cfa-a8df-5953754142e0-kube-api-access-klpg2\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-c2g9p\" (UID: \"c1ca4902-a7b7-4cfa-a8df-5953754142e0\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-c2g9p" Oct 02 02:12:08 crc kubenswrapper[4885]: I1002 02:12:08.695719 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-c2g9p" Oct 02 02:12:09 crc kubenswrapper[4885]: I1002 02:12:09.206546 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-c2g9p"] Oct 02 02:12:09 crc kubenswrapper[4885]: I1002 02:12:09.315823 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-c2g9p" event={"ID":"c1ca4902-a7b7-4cfa-a8df-5953754142e0","Type":"ContainerStarted","Data":"3366dac21c035217b9fd6cb4cf5998cbd044f08f185457f36b30666656c61571"} Oct 02 02:12:10 crc kubenswrapper[4885]: I1002 02:12:10.336815 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-c2g9p" event={"ID":"c1ca4902-a7b7-4cfa-a8df-5953754142e0","Type":"ContainerStarted","Data":"a7955721fbb5465777b93d441594bdea5f72aa22ba463ca30db0d89bbc7fec92"} Oct 02 02:12:10 crc kubenswrapper[4885]: I1002 02:12:10.364683 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-c2g9p" podStartSLOduration=1.855449738 podStartE2EDuration="2.364663983s" podCreationTimestamp="2025-10-02 02:12:08 +0000 UTC" firstStartedPulling="2025-10-02 02:12:09.206638487 +0000 UTC m=+1518.018385886" lastFinishedPulling="2025-10-02 02:12:09.715852702 +0000 UTC m=+1518.527600131" observedRunningTime="2025-10-02 02:12:10.360669972 +0000 UTC m=+1519.172417421" watchObservedRunningTime="2025-10-02 02:12:10.364663983 +0000 UTC m=+1519.176411392" Oct 02 02:12:11 crc kubenswrapper[4885]: I1002 02:12:11.831210 4885 scope.go:117] "RemoveContainer" containerID="fa0d2db4bf10abe0c8e33d54ab6448de1f276f34da0265afacb000713267d8c8" Oct 02 02:12:11 crc kubenswrapper[4885]: I1002 02:12:11.862478 4885 scope.go:117] "RemoveContainer" containerID="3189eb217f435027c5682afe170fc3d3df2f758e8e7f6e08ef786038a88e19f3" Oct 02 02:12:11 crc kubenswrapper[4885]: I1002 02:12:11.892758 4885 scope.go:117] "RemoveContainer" containerID="37415b45895365cfac8d3cda583e36f42405ad139ab5a23a98f34186e2451267" Oct 02 02:12:11 crc kubenswrapper[4885]: I1002 02:12:11.934177 4885 scope.go:117] "RemoveContainer" containerID="196cd6495bbe8e9aef57fcdd9cde8457cfc188166141d4cc6cd72a97e3e3d3a0" Oct 02 02:12:13 crc kubenswrapper[4885]: I1002 02:12:13.265330 4885 patch_prober.go:28] interesting pod/machine-config-daemon-rttx8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 02:12:13 crc kubenswrapper[4885]: I1002 02:12:13.265859 4885 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 02:12:43 crc kubenswrapper[4885]: I1002 02:12:43.265747 4885 patch_prober.go:28] interesting pod/machine-config-daemon-rttx8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 02:12:43 crc kubenswrapper[4885]: I1002 02:12:43.267418 4885 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 02:12:43 crc kubenswrapper[4885]: I1002 02:12:43.267495 4885 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" Oct 02 02:12:43 crc kubenswrapper[4885]: I1002 02:12:43.268381 4885 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"be0fc35a7f63d87317c3e5145a3df69700656f283328299a3b8545263cc22494"} pod="openshift-machine-config-operator/machine-config-daemon-rttx8" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 02:12:43 crc kubenswrapper[4885]: I1002 02:12:43.268440 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" containerName="machine-config-daemon" containerID="cri-o://be0fc35a7f63d87317c3e5145a3df69700656f283328299a3b8545263cc22494" gracePeriod=600 Oct 02 02:12:43 crc kubenswrapper[4885]: E1002 02:12:43.423218 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:12:43 crc kubenswrapper[4885]: I1002 02:12:43.745758 4885 generic.go:334] "Generic (PLEG): container finished" podID="d169c8ba-000a-4218-8e3b-4ae53035b110" containerID="be0fc35a7f63d87317c3e5145a3df69700656f283328299a3b8545263cc22494" exitCode=0 Oct 02 02:12:43 crc kubenswrapper[4885]: I1002 02:12:43.745822 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" event={"ID":"d169c8ba-000a-4218-8e3b-4ae53035b110","Type":"ContainerDied","Data":"be0fc35a7f63d87317c3e5145a3df69700656f283328299a3b8545263cc22494"} Oct 02 02:12:43 crc kubenswrapper[4885]: I1002 02:12:43.745928 4885 scope.go:117] "RemoveContainer" containerID="fd898246d6c734dabec969ae60b5e63b1aba037d46a686b38b0527500e9db0fe" Oct 02 02:12:43 crc kubenswrapper[4885]: I1002 02:12:43.747021 4885 scope.go:117] "RemoveContainer" containerID="be0fc35a7f63d87317c3e5145a3df69700656f283328299a3b8545263cc22494" Oct 02 02:12:43 crc kubenswrapper[4885]: E1002 02:12:43.747681 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:12:46 crc kubenswrapper[4885]: I1002 02:12:46.072170 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-7dl92"] Oct 02 02:12:46 crc kubenswrapper[4885]: I1002 02:12:46.083971 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-7dl92"] Oct 02 02:12:47 crc kubenswrapper[4885]: I1002 02:12:47.041634 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-dnlx6"] Oct 02 02:12:47 crc kubenswrapper[4885]: I1002 02:12:47.057183 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-dnlx6"] Oct 02 02:12:48 crc kubenswrapper[4885]: I1002 02:12:48.063877 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7b8634a8-6199-42bf-b18d-592a8ba3d164" path="/var/lib/kubelet/pods/7b8634a8-6199-42bf-b18d-592a8ba3d164/volumes" Oct 02 02:12:48 crc kubenswrapper[4885]: I1002 02:12:48.068037 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="da80ee88-0786-4a7d-9c62-446412d63021" path="/var/lib/kubelet/pods/da80ee88-0786-4a7d-9c62-446412d63021/volumes" Oct 02 02:12:51 crc kubenswrapper[4885]: I1002 02:12:51.039780 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-jt45g"] Oct 02 02:12:51 crc kubenswrapper[4885]: I1002 02:12:51.059392 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-jt45g"] Oct 02 02:12:52 crc kubenswrapper[4885]: I1002 02:12:52.063862 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="19989c97-8f7e-4141-bbb3-b3ab9aa68d5d" path="/var/lib/kubelet/pods/19989c97-8f7e-4141-bbb3-b3ab9aa68d5d/volumes" Oct 02 02:12:58 crc kubenswrapper[4885]: I1002 02:12:58.041947 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-e8ab-account-create-wrrhx"] Oct 02 02:12:58 crc kubenswrapper[4885]: I1002 02:12:58.066245 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-e8ab-account-create-wrrhx"] Oct 02 02:12:59 crc kubenswrapper[4885]: I1002 02:12:59.048116 4885 scope.go:117] "RemoveContainer" containerID="be0fc35a7f63d87317c3e5145a3df69700656f283328299a3b8545263cc22494" Oct 02 02:12:59 crc kubenswrapper[4885]: E1002 02:12:59.048957 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:13:00 crc kubenswrapper[4885]: I1002 02:13:00.063642 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="352a9474-7bae-4467-9e56-2d320b55494a" path="/var/lib/kubelet/pods/352a9474-7bae-4467-9e56-2d320b55494a/volumes" Oct 02 02:13:01 crc kubenswrapper[4885]: I1002 02:13:01.041729 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-77fb-account-create-jn758"] Oct 02 02:13:01 crc kubenswrapper[4885]: I1002 02:13:01.058144 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-45be-account-create-824gn"] Oct 02 02:13:01 crc kubenswrapper[4885]: I1002 02:13:01.069808 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-77fb-account-create-jn758"] Oct 02 02:13:01 crc kubenswrapper[4885]: I1002 02:13:01.080185 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-45be-account-create-824gn"] Oct 02 02:13:02 crc kubenswrapper[4885]: I1002 02:13:02.056670 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="97dba4b7-d361-4408-9b83-a717653acc0f" path="/var/lib/kubelet/pods/97dba4b7-d361-4408-9b83-a717653acc0f/volumes" Oct 02 02:13:02 crc kubenswrapper[4885]: I1002 02:13:02.057356 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b88a1087-86c9-41ad-a0b9-48a2607bfc46" path="/var/lib/kubelet/pods/b88a1087-86c9-41ad-a0b9-48a2607bfc46/volumes" Oct 02 02:13:09 crc kubenswrapper[4885]: I1002 02:13:09.048255 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-d2d8l"] Oct 02 02:13:09 crc kubenswrapper[4885]: I1002 02:13:09.061272 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-d2d8l"] Oct 02 02:13:10 crc kubenswrapper[4885]: I1002 02:13:10.032544 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-t2x9j"] Oct 02 02:13:10 crc kubenswrapper[4885]: I1002 02:13:10.042471 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-sff5w"] Oct 02 02:13:10 crc kubenswrapper[4885]: I1002 02:13:10.083595 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ab1b5ad-d6a5-4bfa-96bd-9688c397a87b" path="/var/lib/kubelet/pods/5ab1b5ad-d6a5-4bfa-96bd-9688c397a87b/volumes" Oct 02 02:13:10 crc kubenswrapper[4885]: I1002 02:13:10.084527 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-t2x9j"] Oct 02 02:13:10 crc kubenswrapper[4885]: I1002 02:13:10.084576 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-sff5w"] Oct 02 02:13:11 crc kubenswrapper[4885]: I1002 02:13:11.047159 4885 scope.go:117] "RemoveContainer" containerID="be0fc35a7f63d87317c3e5145a3df69700656f283328299a3b8545263cc22494" Oct 02 02:13:11 crc kubenswrapper[4885]: E1002 02:13:11.047683 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:13:12 crc kubenswrapper[4885]: I1002 02:13:12.014478 4885 scope.go:117] "RemoveContainer" containerID="e8887f0cf1208a83481402ef4a9f149d8af949103cfd35b53383a946e2763dd8" Oct 02 02:13:12 crc kubenswrapper[4885]: I1002 02:13:12.046279 4885 scope.go:117] "RemoveContainer" containerID="af98d4d991ccc2571344d27684d911494c041b09417917524d78ad7f42ba6f84" Oct 02 02:13:12 crc kubenswrapper[4885]: I1002 02:13:12.061146 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ab748dfb-c9b3-412c-8c8d-fe40d3c1d9e7" path="/var/lib/kubelet/pods/ab748dfb-c9b3-412c-8c8d-fe40d3c1d9e7/volumes" Oct 02 02:13:12 crc kubenswrapper[4885]: I1002 02:13:12.061702 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7eb74c5-7bff-43b5-9bd5-89b05eb43ff6" path="/var/lib/kubelet/pods/e7eb74c5-7bff-43b5-9bd5-89b05eb43ff6/volumes" Oct 02 02:13:12 crc kubenswrapper[4885]: I1002 02:13:12.089983 4885 scope.go:117] "RemoveContainer" containerID="af22f9c7d8e8ba413cc823f2e85402994be30148633b974c8c2c302351d03e4e" Oct 02 02:13:12 crc kubenswrapper[4885]: I1002 02:13:12.142789 4885 scope.go:117] "RemoveContainer" containerID="c9e734a87f217a4029999cbcf9a65483cb0267cdff71063783527577082855bf" Oct 02 02:13:12 crc kubenswrapper[4885]: I1002 02:13:12.180989 4885 scope.go:117] "RemoveContainer" containerID="05489a5bf3fb1c22237f40e797f5260aa4732bffe7ad2335ce7c1139122fa939" Oct 02 02:13:12 crc kubenswrapper[4885]: I1002 02:13:12.250647 4885 scope.go:117] "RemoveContainer" containerID="49363e2fbd3e9324689ef02f8d32d917ce193a0b4ab530b68afaac7b0433606a" Oct 02 02:13:12 crc kubenswrapper[4885]: I1002 02:13:12.278292 4885 scope.go:117] "RemoveContainer" containerID="417844a7916dd6611a997b92463595cb69243430f38bd4bc93e7d8ec63b1a43b" Oct 02 02:13:12 crc kubenswrapper[4885]: I1002 02:13:12.310204 4885 scope.go:117] "RemoveContainer" containerID="b1d356a6538df1c7aaf0a73e390a4014364807d6d4f4a03d7384039af90d4972" Oct 02 02:13:26 crc kubenswrapper[4885]: I1002 02:13:26.046753 4885 scope.go:117] "RemoveContainer" containerID="be0fc35a7f63d87317c3e5145a3df69700656f283328299a3b8545263cc22494" Oct 02 02:13:26 crc kubenswrapper[4885]: E1002 02:13:26.047912 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:13:33 crc kubenswrapper[4885]: I1002 02:13:33.078072 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-f842-account-create-fsxd8"] Oct 02 02:13:33 crc kubenswrapper[4885]: I1002 02:13:33.085903 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-fe7d-account-create-qtl4l"] Oct 02 02:13:33 crc kubenswrapper[4885]: I1002 02:13:33.094911 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-fe7d-account-create-qtl4l"] Oct 02 02:13:33 crc kubenswrapper[4885]: I1002 02:13:33.101961 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-1c48-account-create-6cz92"] Oct 02 02:13:33 crc kubenswrapper[4885]: I1002 02:13:33.108299 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-f842-account-create-fsxd8"] Oct 02 02:13:33 crc kubenswrapper[4885]: I1002 02:13:33.114329 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-1c48-account-create-6cz92"] Oct 02 02:13:34 crc kubenswrapper[4885]: I1002 02:13:34.067094 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2147f705-82c1-4cce-98c8-85c7aee85542" path="/var/lib/kubelet/pods/2147f705-82c1-4cce-98c8-85c7aee85542/volumes" Oct 02 02:13:34 crc kubenswrapper[4885]: I1002 02:13:34.067907 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="32627c09-5c38-4a0b-90a6-d6c665654402" path="/var/lib/kubelet/pods/32627c09-5c38-4a0b-90a6-d6c665654402/volumes" Oct 02 02:13:34 crc kubenswrapper[4885]: I1002 02:13:34.068647 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ef4d5777-35ad-4551-a2df-28af567ebeb5" path="/var/lib/kubelet/pods/ef4d5777-35ad-4551-a2df-28af567ebeb5/volumes" Oct 02 02:13:37 crc kubenswrapper[4885]: I1002 02:13:37.041155 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-tvjc2"] Oct 02 02:13:37 crc kubenswrapper[4885]: I1002 02:13:37.047579 4885 scope.go:117] "RemoveContainer" containerID="be0fc35a7f63d87317c3e5145a3df69700656f283328299a3b8545263cc22494" Oct 02 02:13:37 crc kubenswrapper[4885]: E1002 02:13:37.048128 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:13:37 crc kubenswrapper[4885]: I1002 02:13:37.054062 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-tvjc2"] Oct 02 02:13:38 crc kubenswrapper[4885]: I1002 02:13:38.069200 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d0092499-91ac-4e53-8c91-04d30bd49e68" path="/var/lib/kubelet/pods/d0092499-91ac-4e53-8c91-04d30bd49e68/volumes" Oct 02 02:13:47 crc kubenswrapper[4885]: I1002 02:13:47.565627 4885 generic.go:334] "Generic (PLEG): container finished" podID="c1ca4902-a7b7-4cfa-a8df-5953754142e0" containerID="a7955721fbb5465777b93d441594bdea5f72aa22ba463ca30db0d89bbc7fec92" exitCode=0 Oct 02 02:13:47 crc kubenswrapper[4885]: I1002 02:13:47.565805 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-c2g9p" event={"ID":"c1ca4902-a7b7-4cfa-a8df-5953754142e0","Type":"ContainerDied","Data":"a7955721fbb5465777b93d441594bdea5f72aa22ba463ca30db0d89bbc7fec92"} Oct 02 02:13:49 crc kubenswrapper[4885]: I1002 02:13:49.096761 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-c2g9p" Oct 02 02:13:49 crc kubenswrapper[4885]: I1002 02:13:49.217921 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c1ca4902-a7b7-4cfa-a8df-5953754142e0-ssh-key\") pod \"c1ca4902-a7b7-4cfa-a8df-5953754142e0\" (UID: \"c1ca4902-a7b7-4cfa-a8df-5953754142e0\") " Oct 02 02:13:49 crc kubenswrapper[4885]: I1002 02:13:49.218153 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c1ca4902-a7b7-4cfa-a8df-5953754142e0-inventory\") pod \"c1ca4902-a7b7-4cfa-a8df-5953754142e0\" (UID: \"c1ca4902-a7b7-4cfa-a8df-5953754142e0\") " Oct 02 02:13:49 crc kubenswrapper[4885]: I1002 02:13:49.218276 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-klpg2\" (UniqueName: \"kubernetes.io/projected/c1ca4902-a7b7-4cfa-a8df-5953754142e0-kube-api-access-klpg2\") pod \"c1ca4902-a7b7-4cfa-a8df-5953754142e0\" (UID: \"c1ca4902-a7b7-4cfa-a8df-5953754142e0\") " Oct 02 02:13:49 crc kubenswrapper[4885]: I1002 02:13:49.225767 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c1ca4902-a7b7-4cfa-a8df-5953754142e0-kube-api-access-klpg2" (OuterVolumeSpecName: "kube-api-access-klpg2") pod "c1ca4902-a7b7-4cfa-a8df-5953754142e0" (UID: "c1ca4902-a7b7-4cfa-a8df-5953754142e0"). InnerVolumeSpecName "kube-api-access-klpg2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:13:49 crc kubenswrapper[4885]: I1002 02:13:49.252396 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1ca4902-a7b7-4cfa-a8df-5953754142e0-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c1ca4902-a7b7-4cfa-a8df-5953754142e0" (UID: "c1ca4902-a7b7-4cfa-a8df-5953754142e0"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:13:49 crc kubenswrapper[4885]: I1002 02:13:49.265117 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1ca4902-a7b7-4cfa-a8df-5953754142e0-inventory" (OuterVolumeSpecName: "inventory") pod "c1ca4902-a7b7-4cfa-a8df-5953754142e0" (UID: "c1ca4902-a7b7-4cfa-a8df-5953754142e0"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:13:49 crc kubenswrapper[4885]: I1002 02:13:49.320245 4885 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c1ca4902-a7b7-4cfa-a8df-5953754142e0-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 02:13:49 crc kubenswrapper[4885]: I1002 02:13:49.320298 4885 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c1ca4902-a7b7-4cfa-a8df-5953754142e0-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 02:13:49 crc kubenswrapper[4885]: I1002 02:13:49.320313 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-klpg2\" (UniqueName: \"kubernetes.io/projected/c1ca4902-a7b7-4cfa-a8df-5953754142e0-kube-api-access-klpg2\") on node \"crc\" DevicePath \"\"" Oct 02 02:13:49 crc kubenswrapper[4885]: I1002 02:13:49.593443 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-c2g9p" event={"ID":"c1ca4902-a7b7-4cfa-a8df-5953754142e0","Type":"ContainerDied","Data":"3366dac21c035217b9fd6cb4cf5998cbd044f08f185457f36b30666656c61571"} Oct 02 02:13:49 crc kubenswrapper[4885]: I1002 02:13:49.593510 4885 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3366dac21c035217b9fd6cb4cf5998cbd044f08f185457f36b30666656c61571" Oct 02 02:13:49 crc kubenswrapper[4885]: I1002 02:13:49.593533 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-c2g9p" Oct 02 02:13:49 crc kubenswrapper[4885]: I1002 02:13:49.703389 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zlzq7"] Oct 02 02:13:49 crc kubenswrapper[4885]: E1002 02:13:49.704166 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1ca4902-a7b7-4cfa-a8df-5953754142e0" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Oct 02 02:13:49 crc kubenswrapper[4885]: I1002 02:13:49.704203 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1ca4902-a7b7-4cfa-a8df-5953754142e0" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Oct 02 02:13:49 crc kubenswrapper[4885]: I1002 02:13:49.704599 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1ca4902-a7b7-4cfa-a8df-5953754142e0" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Oct 02 02:13:49 crc kubenswrapper[4885]: I1002 02:13:49.707144 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zlzq7" Oct 02 02:13:49 crc kubenswrapper[4885]: I1002 02:13:49.709528 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 02:13:49 crc kubenswrapper[4885]: I1002 02:13:49.711329 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 02:13:49 crc kubenswrapper[4885]: I1002 02:13:49.711376 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 02:13:49 crc kubenswrapper[4885]: I1002 02:13:49.711389 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-f8tvb" Oct 02 02:13:49 crc kubenswrapper[4885]: I1002 02:13:49.716059 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zlzq7"] Oct 02 02:13:49 crc kubenswrapper[4885]: I1002 02:13:49.834253 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2nspt\" (UniqueName: \"kubernetes.io/projected/fd55e079-473a-43d3-aca2-ee0d91c06aca-kube-api-access-2nspt\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-zlzq7\" (UID: \"fd55e079-473a-43d3-aca2-ee0d91c06aca\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zlzq7" Oct 02 02:13:49 crc kubenswrapper[4885]: I1002 02:13:49.834663 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fd55e079-473a-43d3-aca2-ee0d91c06aca-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-zlzq7\" (UID: \"fd55e079-473a-43d3-aca2-ee0d91c06aca\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zlzq7" Oct 02 02:13:49 crc kubenswrapper[4885]: I1002 02:13:49.834797 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fd55e079-473a-43d3-aca2-ee0d91c06aca-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-zlzq7\" (UID: \"fd55e079-473a-43d3-aca2-ee0d91c06aca\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zlzq7" Oct 02 02:13:49 crc kubenswrapper[4885]: I1002 02:13:49.937339 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fd55e079-473a-43d3-aca2-ee0d91c06aca-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-zlzq7\" (UID: \"fd55e079-473a-43d3-aca2-ee0d91c06aca\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zlzq7" Oct 02 02:13:49 crc kubenswrapper[4885]: I1002 02:13:49.937651 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2nspt\" (UniqueName: \"kubernetes.io/projected/fd55e079-473a-43d3-aca2-ee0d91c06aca-kube-api-access-2nspt\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-zlzq7\" (UID: \"fd55e079-473a-43d3-aca2-ee0d91c06aca\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zlzq7" Oct 02 02:13:49 crc kubenswrapper[4885]: I1002 02:13:49.937714 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fd55e079-473a-43d3-aca2-ee0d91c06aca-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-zlzq7\" (UID: \"fd55e079-473a-43d3-aca2-ee0d91c06aca\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zlzq7" Oct 02 02:13:49 crc kubenswrapper[4885]: I1002 02:13:49.961293 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fd55e079-473a-43d3-aca2-ee0d91c06aca-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-zlzq7\" (UID: \"fd55e079-473a-43d3-aca2-ee0d91c06aca\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zlzq7" Oct 02 02:13:49 crc kubenswrapper[4885]: I1002 02:13:49.961902 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fd55e079-473a-43d3-aca2-ee0d91c06aca-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-zlzq7\" (UID: \"fd55e079-473a-43d3-aca2-ee0d91c06aca\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zlzq7" Oct 02 02:13:49 crc kubenswrapper[4885]: I1002 02:13:49.966437 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2nspt\" (UniqueName: \"kubernetes.io/projected/fd55e079-473a-43d3-aca2-ee0d91c06aca-kube-api-access-2nspt\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-zlzq7\" (UID: \"fd55e079-473a-43d3-aca2-ee0d91c06aca\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zlzq7" Oct 02 02:13:50 crc kubenswrapper[4885]: I1002 02:13:50.031369 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zlzq7" Oct 02 02:13:50 crc kubenswrapper[4885]: I1002 02:13:50.404069 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zlzq7"] Oct 02 02:13:50 crc kubenswrapper[4885]: I1002 02:13:50.606152 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zlzq7" event={"ID":"fd55e079-473a-43d3-aca2-ee0d91c06aca","Type":"ContainerStarted","Data":"c6e790fdf5cb359cb6ebcf0181e14d078f88971f27beb0604e561baacce6397b"} Oct 02 02:13:51 crc kubenswrapper[4885]: I1002 02:13:51.046669 4885 scope.go:117] "RemoveContainer" containerID="be0fc35a7f63d87317c3e5145a3df69700656f283328299a3b8545263cc22494" Oct 02 02:13:51 crc kubenswrapper[4885]: E1002 02:13:51.047439 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:13:51 crc kubenswrapper[4885]: I1002 02:13:51.658018 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zlzq7" event={"ID":"fd55e079-473a-43d3-aca2-ee0d91c06aca","Type":"ContainerStarted","Data":"2136983b5d5ebb483fb50460a6385ad0e4adb10625ffb0f53409e8d0177f7c76"} Oct 02 02:13:51 crc kubenswrapper[4885]: I1002 02:13:51.692965 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zlzq7" podStartSLOduration=2.268040891 podStartE2EDuration="2.69294412s" podCreationTimestamp="2025-10-02 02:13:49 +0000 UTC" firstStartedPulling="2025-10-02 02:13:50.412526659 +0000 UTC m=+1619.224274058" lastFinishedPulling="2025-10-02 02:13:50.837429848 +0000 UTC m=+1619.649177287" observedRunningTime="2025-10-02 02:13:51.686497974 +0000 UTC m=+1620.498245373" watchObservedRunningTime="2025-10-02 02:13:51.69294412 +0000 UTC m=+1620.504691519" Oct 02 02:14:06 crc kubenswrapper[4885]: I1002 02:14:06.046969 4885 scope.go:117] "RemoveContainer" containerID="be0fc35a7f63d87317c3e5145a3df69700656f283328299a3b8545263cc22494" Oct 02 02:14:06 crc kubenswrapper[4885]: E1002 02:14:06.047756 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:14:09 crc kubenswrapper[4885]: I1002 02:14:09.049333 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-spxxt"] Oct 02 02:14:09 crc kubenswrapper[4885]: I1002 02:14:09.062069 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-spxxt"] Oct 02 02:14:10 crc kubenswrapper[4885]: I1002 02:14:10.067953 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a9ed972-b8c7-4ff7-af7b-899842ecc19c" path="/var/lib/kubelet/pods/6a9ed972-b8c7-4ff7-af7b-899842ecc19c/volumes" Oct 02 02:14:12 crc kubenswrapper[4885]: I1002 02:14:12.499000 4885 scope.go:117] "RemoveContainer" containerID="f25a01d3dd742ffada042f77bc5438c703eadf29a06cfa21a978326393093670" Oct 02 02:14:12 crc kubenswrapper[4885]: I1002 02:14:12.535890 4885 scope.go:117] "RemoveContainer" containerID="8d716973aace109fb672bfa98c5dffccc3dd5fa5f26180d6b4abf24b760c199c" Oct 02 02:14:12 crc kubenswrapper[4885]: I1002 02:14:12.611979 4885 scope.go:117] "RemoveContainer" containerID="4f1562bee17ffaa4f7203145677f0398a269b28de57c734ed5c2f625b10afe1c" Oct 02 02:14:12 crc kubenswrapper[4885]: I1002 02:14:12.648100 4885 scope.go:117] "RemoveContainer" containerID="29d8b0d4bf1525e7f40bd8accd471c2f106dfae1d0fb0eb9c3ccac5d65c6ee07" Oct 02 02:14:12 crc kubenswrapper[4885]: I1002 02:14:12.681631 4885 scope.go:117] "RemoveContainer" containerID="2673effa29be2d50feab73fa410b48c8a801fdbb1e2991b20d9fd88761ccdee5" Oct 02 02:14:12 crc kubenswrapper[4885]: I1002 02:14:12.723455 4885 scope.go:117] "RemoveContainer" containerID="d1c29d0f0646e6b585de70535fb0505faeed21ff4aac88795e86f26e0758e872" Oct 02 02:14:19 crc kubenswrapper[4885]: I1002 02:14:19.046946 4885 scope.go:117] "RemoveContainer" containerID="be0fc35a7f63d87317c3e5145a3df69700656f283328299a3b8545263cc22494" Oct 02 02:14:19 crc kubenswrapper[4885]: E1002 02:14:19.047918 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:14:22 crc kubenswrapper[4885]: I1002 02:14:22.066282 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-dr95w"] Oct 02 02:14:22 crc kubenswrapper[4885]: I1002 02:14:22.073943 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-dr95w"] Oct 02 02:14:24 crc kubenswrapper[4885]: I1002 02:14:24.059504 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d4a98a2-1f96-47a2-bacc-a2dd42bc4119" path="/var/lib/kubelet/pods/1d4a98a2-1f96-47a2-bacc-a2dd42bc4119/volumes" Oct 02 02:14:33 crc kubenswrapper[4885]: I1002 02:14:33.039210 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-xf24l"] Oct 02 02:14:33 crc kubenswrapper[4885]: I1002 02:14:33.046572 4885 scope.go:117] "RemoveContainer" containerID="be0fc35a7f63d87317c3e5145a3df69700656f283328299a3b8545263cc22494" Oct 02 02:14:33 crc kubenswrapper[4885]: E1002 02:14:33.046963 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:14:33 crc kubenswrapper[4885]: I1002 02:14:33.049676 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-xf24l"] Oct 02 02:14:34 crc kubenswrapper[4885]: I1002 02:14:34.069236 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c036a628-a241-45ea-bd2a-5d6d50804f61" path="/var/lib/kubelet/pods/c036a628-a241-45ea-bd2a-5d6d50804f61/volumes" Oct 02 02:14:47 crc kubenswrapper[4885]: I1002 02:14:47.046531 4885 scope.go:117] "RemoveContainer" containerID="be0fc35a7f63d87317c3e5145a3df69700656f283328299a3b8545263cc22494" Oct 02 02:14:47 crc kubenswrapper[4885]: E1002 02:14:47.048890 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:14:59 crc kubenswrapper[4885]: I1002 02:14:59.050914 4885 scope.go:117] "RemoveContainer" containerID="be0fc35a7f63d87317c3e5145a3df69700656f283328299a3b8545263cc22494" Oct 02 02:14:59 crc kubenswrapper[4885]: E1002 02:14:59.051816 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:15:00 crc kubenswrapper[4885]: I1002 02:15:00.161895 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322855-22v6b"] Oct 02 02:15:00 crc kubenswrapper[4885]: I1002 02:15:00.162986 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322855-22v6b" Oct 02 02:15:00 crc kubenswrapper[4885]: I1002 02:15:00.170067 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 02 02:15:00 crc kubenswrapper[4885]: I1002 02:15:00.170326 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 02 02:15:00 crc kubenswrapper[4885]: I1002 02:15:00.177549 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322855-22v6b"] Oct 02 02:15:00 crc kubenswrapper[4885]: I1002 02:15:00.210105 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mmglv\" (UniqueName: \"kubernetes.io/projected/6954232a-a827-4d3b-abf7-e552100fd000-kube-api-access-mmglv\") pod \"collect-profiles-29322855-22v6b\" (UID: \"6954232a-a827-4d3b-abf7-e552100fd000\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322855-22v6b" Oct 02 02:15:00 crc kubenswrapper[4885]: I1002 02:15:00.210250 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6954232a-a827-4d3b-abf7-e552100fd000-config-volume\") pod \"collect-profiles-29322855-22v6b\" (UID: \"6954232a-a827-4d3b-abf7-e552100fd000\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322855-22v6b" Oct 02 02:15:00 crc kubenswrapper[4885]: I1002 02:15:00.210478 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6954232a-a827-4d3b-abf7-e552100fd000-secret-volume\") pod \"collect-profiles-29322855-22v6b\" (UID: \"6954232a-a827-4d3b-abf7-e552100fd000\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322855-22v6b" Oct 02 02:15:00 crc kubenswrapper[4885]: I1002 02:15:00.312615 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6954232a-a827-4d3b-abf7-e552100fd000-secret-volume\") pod \"collect-profiles-29322855-22v6b\" (UID: \"6954232a-a827-4d3b-abf7-e552100fd000\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322855-22v6b" Oct 02 02:15:00 crc kubenswrapper[4885]: I1002 02:15:00.312785 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mmglv\" (UniqueName: \"kubernetes.io/projected/6954232a-a827-4d3b-abf7-e552100fd000-kube-api-access-mmglv\") pod \"collect-profiles-29322855-22v6b\" (UID: \"6954232a-a827-4d3b-abf7-e552100fd000\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322855-22v6b" Oct 02 02:15:00 crc kubenswrapper[4885]: I1002 02:15:00.312842 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6954232a-a827-4d3b-abf7-e552100fd000-config-volume\") pod \"collect-profiles-29322855-22v6b\" (UID: \"6954232a-a827-4d3b-abf7-e552100fd000\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322855-22v6b" Oct 02 02:15:00 crc kubenswrapper[4885]: I1002 02:15:00.313939 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6954232a-a827-4d3b-abf7-e552100fd000-config-volume\") pod \"collect-profiles-29322855-22v6b\" (UID: \"6954232a-a827-4d3b-abf7-e552100fd000\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322855-22v6b" Oct 02 02:15:00 crc kubenswrapper[4885]: I1002 02:15:00.319807 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6954232a-a827-4d3b-abf7-e552100fd000-secret-volume\") pod \"collect-profiles-29322855-22v6b\" (UID: \"6954232a-a827-4d3b-abf7-e552100fd000\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322855-22v6b" Oct 02 02:15:00 crc kubenswrapper[4885]: I1002 02:15:00.344406 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mmglv\" (UniqueName: \"kubernetes.io/projected/6954232a-a827-4d3b-abf7-e552100fd000-kube-api-access-mmglv\") pod \"collect-profiles-29322855-22v6b\" (UID: \"6954232a-a827-4d3b-abf7-e552100fd000\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322855-22v6b" Oct 02 02:15:00 crc kubenswrapper[4885]: I1002 02:15:00.479982 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322855-22v6b" Oct 02 02:15:00 crc kubenswrapper[4885]: I1002 02:15:00.991002 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322855-22v6b"] Oct 02 02:15:01 crc kubenswrapper[4885]: I1002 02:15:01.506624 4885 generic.go:334] "Generic (PLEG): container finished" podID="6954232a-a827-4d3b-abf7-e552100fd000" containerID="69dff0f53d137fa8c49e2b7a285c1e7cf30a27dfaf3772c78aa36403cd87322d" exitCode=0 Oct 02 02:15:01 crc kubenswrapper[4885]: I1002 02:15:01.506691 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29322855-22v6b" event={"ID":"6954232a-a827-4d3b-abf7-e552100fd000","Type":"ContainerDied","Data":"69dff0f53d137fa8c49e2b7a285c1e7cf30a27dfaf3772c78aa36403cd87322d"} Oct 02 02:15:01 crc kubenswrapper[4885]: I1002 02:15:01.506902 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29322855-22v6b" event={"ID":"6954232a-a827-4d3b-abf7-e552100fd000","Type":"ContainerStarted","Data":"23fd190737787ea64b999d88727e94af2bfa4e565748eea77f9d89bf82f05dae"} Oct 02 02:15:02 crc kubenswrapper[4885]: I1002 02:15:02.946126 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322855-22v6b" Oct 02 02:15:03 crc kubenswrapper[4885]: I1002 02:15:03.048843 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-vj2tf"] Oct 02 02:15:03 crc kubenswrapper[4885]: I1002 02:15:03.063389 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-vj2tf"] Oct 02 02:15:03 crc kubenswrapper[4885]: I1002 02:15:03.074712 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6954232a-a827-4d3b-abf7-e552100fd000-secret-volume\") pod \"6954232a-a827-4d3b-abf7-e552100fd000\" (UID: \"6954232a-a827-4d3b-abf7-e552100fd000\") " Oct 02 02:15:03 crc kubenswrapper[4885]: I1002 02:15:03.074922 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mmglv\" (UniqueName: \"kubernetes.io/projected/6954232a-a827-4d3b-abf7-e552100fd000-kube-api-access-mmglv\") pod \"6954232a-a827-4d3b-abf7-e552100fd000\" (UID: \"6954232a-a827-4d3b-abf7-e552100fd000\") " Oct 02 02:15:03 crc kubenswrapper[4885]: I1002 02:15:03.075140 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6954232a-a827-4d3b-abf7-e552100fd000-config-volume\") pod \"6954232a-a827-4d3b-abf7-e552100fd000\" (UID: \"6954232a-a827-4d3b-abf7-e552100fd000\") " Oct 02 02:15:03 crc kubenswrapper[4885]: I1002 02:15:03.075814 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6954232a-a827-4d3b-abf7-e552100fd000-config-volume" (OuterVolumeSpecName: "config-volume") pod "6954232a-a827-4d3b-abf7-e552100fd000" (UID: "6954232a-a827-4d3b-abf7-e552100fd000"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:15:03 crc kubenswrapper[4885]: I1002 02:15:03.081039 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6954232a-a827-4d3b-abf7-e552100fd000-kube-api-access-mmglv" (OuterVolumeSpecName: "kube-api-access-mmglv") pod "6954232a-a827-4d3b-abf7-e552100fd000" (UID: "6954232a-a827-4d3b-abf7-e552100fd000"). InnerVolumeSpecName "kube-api-access-mmglv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:15:03 crc kubenswrapper[4885]: I1002 02:15:03.082870 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6954232a-a827-4d3b-abf7-e552100fd000-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "6954232a-a827-4d3b-abf7-e552100fd000" (UID: "6954232a-a827-4d3b-abf7-e552100fd000"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:15:03 crc kubenswrapper[4885]: I1002 02:15:03.177290 4885 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6954232a-a827-4d3b-abf7-e552100fd000-config-volume\") on node \"crc\" DevicePath \"\"" Oct 02 02:15:03 crc kubenswrapper[4885]: I1002 02:15:03.177325 4885 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6954232a-a827-4d3b-abf7-e552100fd000-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 02 02:15:03 crc kubenswrapper[4885]: I1002 02:15:03.177338 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mmglv\" (UniqueName: \"kubernetes.io/projected/6954232a-a827-4d3b-abf7-e552100fd000-kube-api-access-mmglv\") on node \"crc\" DevicePath \"\"" Oct 02 02:15:03 crc kubenswrapper[4885]: I1002 02:15:03.540065 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29322855-22v6b" event={"ID":"6954232a-a827-4d3b-abf7-e552100fd000","Type":"ContainerDied","Data":"23fd190737787ea64b999d88727e94af2bfa4e565748eea77f9d89bf82f05dae"} Oct 02 02:15:03 crc kubenswrapper[4885]: I1002 02:15:03.540115 4885 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="23fd190737787ea64b999d88727e94af2bfa4e565748eea77f9d89bf82f05dae" Oct 02 02:15:03 crc kubenswrapper[4885]: I1002 02:15:03.540176 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322855-22v6b" Oct 02 02:15:04 crc kubenswrapper[4885]: I1002 02:15:04.070544 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8fa66af9-7b4b-480c-8f96-bda850e2eef6" path="/var/lib/kubelet/pods/8fa66af9-7b4b-480c-8f96-bda850e2eef6/volumes" Oct 02 02:15:05 crc kubenswrapper[4885]: I1002 02:15:05.567803 4885 generic.go:334] "Generic (PLEG): container finished" podID="fd55e079-473a-43d3-aca2-ee0d91c06aca" containerID="2136983b5d5ebb483fb50460a6385ad0e4adb10625ffb0f53409e8d0177f7c76" exitCode=0 Oct 02 02:15:05 crc kubenswrapper[4885]: I1002 02:15:05.567918 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zlzq7" event={"ID":"fd55e079-473a-43d3-aca2-ee0d91c06aca","Type":"ContainerDied","Data":"2136983b5d5ebb483fb50460a6385ad0e4adb10625ffb0f53409e8d0177f7c76"} Oct 02 02:15:07 crc kubenswrapper[4885]: I1002 02:15:07.222702 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zlzq7" Oct 02 02:15:07 crc kubenswrapper[4885]: I1002 02:15:07.282828 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fd55e079-473a-43d3-aca2-ee0d91c06aca-inventory\") pod \"fd55e079-473a-43d3-aca2-ee0d91c06aca\" (UID: \"fd55e079-473a-43d3-aca2-ee0d91c06aca\") " Oct 02 02:15:07 crc kubenswrapper[4885]: I1002 02:15:07.282972 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2nspt\" (UniqueName: \"kubernetes.io/projected/fd55e079-473a-43d3-aca2-ee0d91c06aca-kube-api-access-2nspt\") pod \"fd55e079-473a-43d3-aca2-ee0d91c06aca\" (UID: \"fd55e079-473a-43d3-aca2-ee0d91c06aca\") " Oct 02 02:15:07 crc kubenswrapper[4885]: I1002 02:15:07.283005 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fd55e079-473a-43d3-aca2-ee0d91c06aca-ssh-key\") pod \"fd55e079-473a-43d3-aca2-ee0d91c06aca\" (UID: \"fd55e079-473a-43d3-aca2-ee0d91c06aca\") " Oct 02 02:15:07 crc kubenswrapper[4885]: I1002 02:15:07.305627 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fd55e079-473a-43d3-aca2-ee0d91c06aca-kube-api-access-2nspt" (OuterVolumeSpecName: "kube-api-access-2nspt") pod "fd55e079-473a-43d3-aca2-ee0d91c06aca" (UID: "fd55e079-473a-43d3-aca2-ee0d91c06aca"). InnerVolumeSpecName "kube-api-access-2nspt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:15:07 crc kubenswrapper[4885]: I1002 02:15:07.325440 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd55e079-473a-43d3-aca2-ee0d91c06aca-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "fd55e079-473a-43d3-aca2-ee0d91c06aca" (UID: "fd55e079-473a-43d3-aca2-ee0d91c06aca"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:15:07 crc kubenswrapper[4885]: I1002 02:15:07.336840 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd55e079-473a-43d3-aca2-ee0d91c06aca-inventory" (OuterVolumeSpecName: "inventory") pod "fd55e079-473a-43d3-aca2-ee0d91c06aca" (UID: "fd55e079-473a-43d3-aca2-ee0d91c06aca"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:15:07 crc kubenswrapper[4885]: I1002 02:15:07.385271 4885 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fd55e079-473a-43d3-aca2-ee0d91c06aca-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 02:15:07 crc kubenswrapper[4885]: I1002 02:15:07.385301 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2nspt\" (UniqueName: \"kubernetes.io/projected/fd55e079-473a-43d3-aca2-ee0d91c06aca-kube-api-access-2nspt\") on node \"crc\" DevicePath \"\"" Oct 02 02:15:07 crc kubenswrapper[4885]: I1002 02:15:07.385311 4885 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fd55e079-473a-43d3-aca2-ee0d91c06aca-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 02:15:07 crc kubenswrapper[4885]: I1002 02:15:07.597678 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zlzq7" event={"ID":"fd55e079-473a-43d3-aca2-ee0d91c06aca","Type":"ContainerDied","Data":"c6e790fdf5cb359cb6ebcf0181e14d078f88971f27beb0604e561baacce6397b"} Oct 02 02:15:07 crc kubenswrapper[4885]: I1002 02:15:07.597737 4885 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c6e790fdf5cb359cb6ebcf0181e14d078f88971f27beb0604e561baacce6397b" Oct 02 02:15:07 crc kubenswrapper[4885]: I1002 02:15:07.597765 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zlzq7" Oct 02 02:15:07 crc kubenswrapper[4885]: I1002 02:15:07.746952 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5l99v"] Oct 02 02:15:07 crc kubenswrapper[4885]: E1002 02:15:07.747734 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6954232a-a827-4d3b-abf7-e552100fd000" containerName="collect-profiles" Oct 02 02:15:07 crc kubenswrapper[4885]: I1002 02:15:07.747766 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="6954232a-a827-4d3b-abf7-e552100fd000" containerName="collect-profiles" Oct 02 02:15:07 crc kubenswrapper[4885]: E1002 02:15:07.747810 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd55e079-473a-43d3-aca2-ee0d91c06aca" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 02 02:15:07 crc kubenswrapper[4885]: I1002 02:15:07.747818 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd55e079-473a-43d3-aca2-ee0d91c06aca" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 02 02:15:07 crc kubenswrapper[4885]: I1002 02:15:07.748155 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd55e079-473a-43d3-aca2-ee0d91c06aca" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 02 02:15:07 crc kubenswrapper[4885]: I1002 02:15:07.748173 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="6954232a-a827-4d3b-abf7-e552100fd000" containerName="collect-profiles" Oct 02 02:15:07 crc kubenswrapper[4885]: I1002 02:15:07.749122 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5l99v" Oct 02 02:15:07 crc kubenswrapper[4885]: I1002 02:15:07.753413 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-f8tvb" Oct 02 02:15:07 crc kubenswrapper[4885]: I1002 02:15:07.753442 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 02:15:07 crc kubenswrapper[4885]: I1002 02:15:07.753923 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 02:15:07 crc kubenswrapper[4885]: I1002 02:15:07.754166 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 02:15:07 crc kubenswrapper[4885]: I1002 02:15:07.757691 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5l99v"] Oct 02 02:15:07 crc kubenswrapper[4885]: I1002 02:15:07.790883 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/829f7f0e-4288-42a2-b585-99beef8cf451-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-5l99v\" (UID: \"829f7f0e-4288-42a2-b585-99beef8cf451\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5l99v" Oct 02 02:15:07 crc kubenswrapper[4885]: I1002 02:15:07.790927 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/829f7f0e-4288-42a2-b585-99beef8cf451-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-5l99v\" (UID: \"829f7f0e-4288-42a2-b585-99beef8cf451\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5l99v" Oct 02 02:15:07 crc kubenswrapper[4885]: I1002 02:15:07.790955 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wkwpc\" (UniqueName: \"kubernetes.io/projected/829f7f0e-4288-42a2-b585-99beef8cf451-kube-api-access-wkwpc\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-5l99v\" (UID: \"829f7f0e-4288-42a2-b585-99beef8cf451\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5l99v" Oct 02 02:15:07 crc kubenswrapper[4885]: I1002 02:15:07.892083 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wkwpc\" (UniqueName: \"kubernetes.io/projected/829f7f0e-4288-42a2-b585-99beef8cf451-kube-api-access-wkwpc\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-5l99v\" (UID: \"829f7f0e-4288-42a2-b585-99beef8cf451\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5l99v" Oct 02 02:15:07 crc kubenswrapper[4885]: I1002 02:15:07.892238 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/829f7f0e-4288-42a2-b585-99beef8cf451-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-5l99v\" (UID: \"829f7f0e-4288-42a2-b585-99beef8cf451\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5l99v" Oct 02 02:15:07 crc kubenswrapper[4885]: I1002 02:15:07.892279 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/829f7f0e-4288-42a2-b585-99beef8cf451-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-5l99v\" (UID: \"829f7f0e-4288-42a2-b585-99beef8cf451\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5l99v" Oct 02 02:15:07 crc kubenswrapper[4885]: I1002 02:15:07.896160 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/829f7f0e-4288-42a2-b585-99beef8cf451-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-5l99v\" (UID: \"829f7f0e-4288-42a2-b585-99beef8cf451\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5l99v" Oct 02 02:15:07 crc kubenswrapper[4885]: I1002 02:15:07.896708 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/829f7f0e-4288-42a2-b585-99beef8cf451-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-5l99v\" (UID: \"829f7f0e-4288-42a2-b585-99beef8cf451\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5l99v" Oct 02 02:15:07 crc kubenswrapper[4885]: I1002 02:15:07.909069 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wkwpc\" (UniqueName: \"kubernetes.io/projected/829f7f0e-4288-42a2-b585-99beef8cf451-kube-api-access-wkwpc\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-5l99v\" (UID: \"829f7f0e-4288-42a2-b585-99beef8cf451\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5l99v" Oct 02 02:15:08 crc kubenswrapper[4885]: I1002 02:15:08.068613 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5l99v" Oct 02 02:15:08 crc kubenswrapper[4885]: I1002 02:15:08.679883 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5l99v"] Oct 02 02:15:08 crc kubenswrapper[4885]: W1002 02:15:08.682204 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod829f7f0e_4288_42a2_b585_99beef8cf451.slice/crio-c489e665f0c51b4e6e5ba0173ff6a64acc6f5952cdc48c42a5b3259ed16264bc WatchSource:0}: Error finding container c489e665f0c51b4e6e5ba0173ff6a64acc6f5952cdc48c42a5b3259ed16264bc: Status 404 returned error can't find the container with id c489e665f0c51b4e6e5ba0173ff6a64acc6f5952cdc48c42a5b3259ed16264bc Oct 02 02:15:09 crc kubenswrapper[4885]: I1002 02:15:09.622384 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5l99v" event={"ID":"829f7f0e-4288-42a2-b585-99beef8cf451","Type":"ContainerStarted","Data":"cb5861744593d945ba55072950e0cedc4ed03ae0bc15218bfec2d9d7da9be920"} Oct 02 02:15:09 crc kubenswrapper[4885]: I1002 02:15:09.622785 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5l99v" event={"ID":"829f7f0e-4288-42a2-b585-99beef8cf451","Type":"ContainerStarted","Data":"c489e665f0c51b4e6e5ba0173ff6a64acc6f5952cdc48c42a5b3259ed16264bc"} Oct 02 02:15:09 crc kubenswrapper[4885]: I1002 02:15:09.649614 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5l99v" podStartSLOduration=2.110969715 podStartE2EDuration="2.649590653s" podCreationTimestamp="2025-10-02 02:15:07 +0000 UTC" firstStartedPulling="2025-10-02 02:15:08.687960109 +0000 UTC m=+1697.499707518" lastFinishedPulling="2025-10-02 02:15:09.226581027 +0000 UTC m=+1698.038328456" observedRunningTime="2025-10-02 02:15:09.639642198 +0000 UTC m=+1698.451389637" watchObservedRunningTime="2025-10-02 02:15:09.649590653 +0000 UTC m=+1698.461338092" Oct 02 02:15:10 crc kubenswrapper[4885]: I1002 02:15:10.046807 4885 scope.go:117] "RemoveContainer" containerID="be0fc35a7f63d87317c3e5145a3df69700656f283328299a3b8545263cc22494" Oct 02 02:15:10 crc kubenswrapper[4885]: E1002 02:15:10.046999 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:15:11 crc kubenswrapper[4885]: I1002 02:15:11.049297 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-gp9st"] Oct 02 02:15:11 crc kubenswrapper[4885]: I1002 02:15:11.065179 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-gp9st"] Oct 02 02:15:12 crc kubenswrapper[4885]: I1002 02:15:12.061026 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0feba5bf-20c5-4578-9cfc-f37d6511c2d6" path="/var/lib/kubelet/pods/0feba5bf-20c5-4578-9cfc-f37d6511c2d6/volumes" Oct 02 02:15:12 crc kubenswrapper[4885]: I1002 02:15:12.882651 4885 scope.go:117] "RemoveContainer" containerID="8d9aa7b7c3aacfe216138f05551b49ff0afe4a9f79e28ef88f3baf68e620b0fb" Oct 02 02:15:12 crc kubenswrapper[4885]: I1002 02:15:12.960923 4885 scope.go:117] "RemoveContainer" containerID="1a1f0197c0399a742a3103b0db978338c9e66ae73dd5516c960bfa34cfe0a60a" Oct 02 02:15:13 crc kubenswrapper[4885]: I1002 02:15:13.014664 4885 scope.go:117] "RemoveContainer" containerID="c6c205ead834e44b7376e00c80be9ca0db099be3ec3ef0e4d6aa2a17921241ce" Oct 02 02:15:13 crc kubenswrapper[4885]: I1002 02:15:13.048022 4885 scope.go:117] "RemoveContainer" containerID="0e6ab596da7147218797e106083b8cfd31507cec30d5f1645bcc1a5f69320aac" Oct 02 02:15:15 crc kubenswrapper[4885]: I1002 02:15:15.691313 4885 generic.go:334] "Generic (PLEG): container finished" podID="829f7f0e-4288-42a2-b585-99beef8cf451" containerID="cb5861744593d945ba55072950e0cedc4ed03ae0bc15218bfec2d9d7da9be920" exitCode=0 Oct 02 02:15:15 crc kubenswrapper[4885]: I1002 02:15:15.691414 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5l99v" event={"ID":"829f7f0e-4288-42a2-b585-99beef8cf451","Type":"ContainerDied","Data":"cb5861744593d945ba55072950e0cedc4ed03ae0bc15218bfec2d9d7da9be920"} Oct 02 02:15:17 crc kubenswrapper[4885]: I1002 02:15:17.273368 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5l99v" Oct 02 02:15:17 crc kubenswrapper[4885]: I1002 02:15:17.318353 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/829f7f0e-4288-42a2-b585-99beef8cf451-ssh-key\") pod \"829f7f0e-4288-42a2-b585-99beef8cf451\" (UID: \"829f7f0e-4288-42a2-b585-99beef8cf451\") " Oct 02 02:15:17 crc kubenswrapper[4885]: I1002 02:15:17.318455 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wkwpc\" (UniqueName: \"kubernetes.io/projected/829f7f0e-4288-42a2-b585-99beef8cf451-kube-api-access-wkwpc\") pod \"829f7f0e-4288-42a2-b585-99beef8cf451\" (UID: \"829f7f0e-4288-42a2-b585-99beef8cf451\") " Oct 02 02:15:17 crc kubenswrapper[4885]: I1002 02:15:17.318567 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/829f7f0e-4288-42a2-b585-99beef8cf451-inventory\") pod \"829f7f0e-4288-42a2-b585-99beef8cf451\" (UID: \"829f7f0e-4288-42a2-b585-99beef8cf451\") " Oct 02 02:15:17 crc kubenswrapper[4885]: I1002 02:15:17.338169 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/829f7f0e-4288-42a2-b585-99beef8cf451-kube-api-access-wkwpc" (OuterVolumeSpecName: "kube-api-access-wkwpc") pod "829f7f0e-4288-42a2-b585-99beef8cf451" (UID: "829f7f0e-4288-42a2-b585-99beef8cf451"). InnerVolumeSpecName "kube-api-access-wkwpc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:15:17 crc kubenswrapper[4885]: I1002 02:15:17.355284 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/829f7f0e-4288-42a2-b585-99beef8cf451-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "829f7f0e-4288-42a2-b585-99beef8cf451" (UID: "829f7f0e-4288-42a2-b585-99beef8cf451"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:15:17 crc kubenswrapper[4885]: I1002 02:15:17.376230 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/829f7f0e-4288-42a2-b585-99beef8cf451-inventory" (OuterVolumeSpecName: "inventory") pod "829f7f0e-4288-42a2-b585-99beef8cf451" (UID: "829f7f0e-4288-42a2-b585-99beef8cf451"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:15:17 crc kubenswrapper[4885]: I1002 02:15:17.420471 4885 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/829f7f0e-4288-42a2-b585-99beef8cf451-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 02:15:17 crc kubenswrapper[4885]: I1002 02:15:17.420697 4885 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/829f7f0e-4288-42a2-b585-99beef8cf451-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 02:15:17 crc kubenswrapper[4885]: I1002 02:15:17.420785 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wkwpc\" (UniqueName: \"kubernetes.io/projected/829f7f0e-4288-42a2-b585-99beef8cf451-kube-api-access-wkwpc\") on node \"crc\" DevicePath \"\"" Oct 02 02:15:17 crc kubenswrapper[4885]: I1002 02:15:17.724396 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5l99v" event={"ID":"829f7f0e-4288-42a2-b585-99beef8cf451","Type":"ContainerDied","Data":"c489e665f0c51b4e6e5ba0173ff6a64acc6f5952cdc48c42a5b3259ed16264bc"} Oct 02 02:15:17 crc kubenswrapper[4885]: I1002 02:15:17.724462 4885 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c489e665f0c51b4e6e5ba0173ff6a64acc6f5952cdc48c42a5b3259ed16264bc" Oct 02 02:15:17 crc kubenswrapper[4885]: I1002 02:15:17.724473 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5l99v" Oct 02 02:15:17 crc kubenswrapper[4885]: I1002 02:15:17.818439 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-p9p9g"] Oct 02 02:15:17 crc kubenswrapper[4885]: E1002 02:15:17.818887 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="829f7f0e-4288-42a2-b585-99beef8cf451" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 02 02:15:17 crc kubenswrapper[4885]: I1002 02:15:17.818907 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="829f7f0e-4288-42a2-b585-99beef8cf451" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 02 02:15:17 crc kubenswrapper[4885]: I1002 02:15:17.819071 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="829f7f0e-4288-42a2-b585-99beef8cf451" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 02 02:15:17 crc kubenswrapper[4885]: I1002 02:15:17.819711 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-p9p9g" Oct 02 02:15:17 crc kubenswrapper[4885]: I1002 02:15:17.822720 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 02:15:17 crc kubenswrapper[4885]: I1002 02:15:17.822858 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-f8tvb" Oct 02 02:15:17 crc kubenswrapper[4885]: I1002 02:15:17.822920 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 02:15:17 crc kubenswrapper[4885]: I1002 02:15:17.823852 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 02:15:17 crc kubenswrapper[4885]: I1002 02:15:17.826973 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-p9p9g"] Oct 02 02:15:17 crc kubenswrapper[4885]: I1002 02:15:17.932594 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/48ef0e0a-ae52-4d11-9023-d53f4ab716ac-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-p9p9g\" (UID: \"48ef0e0a-ae52-4d11-9023-d53f4ab716ac\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-p9p9g" Oct 02 02:15:17 crc kubenswrapper[4885]: I1002 02:15:17.932699 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wm8s7\" (UniqueName: \"kubernetes.io/projected/48ef0e0a-ae52-4d11-9023-d53f4ab716ac-kube-api-access-wm8s7\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-p9p9g\" (UID: \"48ef0e0a-ae52-4d11-9023-d53f4ab716ac\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-p9p9g" Oct 02 02:15:17 crc kubenswrapper[4885]: I1002 02:15:17.932784 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/48ef0e0a-ae52-4d11-9023-d53f4ab716ac-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-p9p9g\" (UID: \"48ef0e0a-ae52-4d11-9023-d53f4ab716ac\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-p9p9g" Oct 02 02:15:18 crc kubenswrapper[4885]: I1002 02:15:18.028645 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-tjcz4"] Oct 02 02:15:18 crc kubenswrapper[4885]: I1002 02:15:18.034590 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-tjcz4"] Oct 02 02:15:18 crc kubenswrapper[4885]: I1002 02:15:18.035017 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/48ef0e0a-ae52-4d11-9023-d53f4ab716ac-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-p9p9g\" (UID: \"48ef0e0a-ae52-4d11-9023-d53f4ab716ac\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-p9p9g" Oct 02 02:15:18 crc kubenswrapper[4885]: I1002 02:15:18.035235 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/48ef0e0a-ae52-4d11-9023-d53f4ab716ac-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-p9p9g\" (UID: \"48ef0e0a-ae52-4d11-9023-d53f4ab716ac\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-p9p9g" Oct 02 02:15:18 crc kubenswrapper[4885]: I1002 02:15:18.035446 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wm8s7\" (UniqueName: \"kubernetes.io/projected/48ef0e0a-ae52-4d11-9023-d53f4ab716ac-kube-api-access-wm8s7\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-p9p9g\" (UID: \"48ef0e0a-ae52-4d11-9023-d53f4ab716ac\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-p9p9g" Oct 02 02:15:18 crc kubenswrapper[4885]: I1002 02:15:18.040786 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/48ef0e0a-ae52-4d11-9023-d53f4ab716ac-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-p9p9g\" (UID: \"48ef0e0a-ae52-4d11-9023-d53f4ab716ac\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-p9p9g" Oct 02 02:15:18 crc kubenswrapper[4885]: I1002 02:15:18.044555 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/48ef0e0a-ae52-4d11-9023-d53f4ab716ac-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-p9p9g\" (UID: \"48ef0e0a-ae52-4d11-9023-d53f4ab716ac\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-p9p9g" Oct 02 02:15:18 crc kubenswrapper[4885]: I1002 02:15:18.056605 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wm8s7\" (UniqueName: \"kubernetes.io/projected/48ef0e0a-ae52-4d11-9023-d53f4ab716ac-kube-api-access-wm8s7\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-p9p9g\" (UID: \"48ef0e0a-ae52-4d11-9023-d53f4ab716ac\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-p9p9g" Oct 02 02:15:18 crc kubenswrapper[4885]: I1002 02:15:18.061195 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c66bb6a-5a3c-4223-aba6-c5d9d0ea83ba" path="/var/lib/kubelet/pods/9c66bb6a-5a3c-4223-aba6-c5d9d0ea83ba/volumes" Oct 02 02:15:18 crc kubenswrapper[4885]: I1002 02:15:18.144008 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-p9p9g" Oct 02 02:15:18 crc kubenswrapper[4885]: I1002 02:15:18.692451 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-p9p9g"] Oct 02 02:15:18 crc kubenswrapper[4885]: I1002 02:15:18.736923 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-p9p9g" event={"ID":"48ef0e0a-ae52-4d11-9023-d53f4ab716ac","Type":"ContainerStarted","Data":"e539a6ee346edde38e7e4ad8ad9d21a85c3002ec92c06dfed294b7f36fc9ae08"} Oct 02 02:15:19 crc kubenswrapper[4885]: I1002 02:15:19.748702 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-p9p9g" event={"ID":"48ef0e0a-ae52-4d11-9023-d53f4ab716ac","Type":"ContainerStarted","Data":"8710cc62cd5f897ef3c5772309b65e4c6ebb45c15e66d996b8b5bafb53d8df90"} Oct 02 02:15:19 crc kubenswrapper[4885]: I1002 02:15:19.776226 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-p9p9g" podStartSLOduration=2.30250333 podStartE2EDuration="2.776195884s" podCreationTimestamp="2025-10-02 02:15:17 +0000 UTC" firstStartedPulling="2025-10-02 02:15:18.711075907 +0000 UTC m=+1707.522823346" lastFinishedPulling="2025-10-02 02:15:19.184768491 +0000 UTC m=+1707.996515900" observedRunningTime="2025-10-02 02:15:19.769575231 +0000 UTC m=+1708.581322670" watchObservedRunningTime="2025-10-02 02:15:19.776195884 +0000 UTC m=+1708.587943323" Oct 02 02:15:21 crc kubenswrapper[4885]: I1002 02:15:21.047511 4885 scope.go:117] "RemoveContainer" containerID="be0fc35a7f63d87317c3e5145a3df69700656f283328299a3b8545263cc22494" Oct 02 02:15:21 crc kubenswrapper[4885]: E1002 02:15:21.048040 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:15:33 crc kubenswrapper[4885]: I1002 02:15:33.046771 4885 scope.go:117] "RemoveContainer" containerID="be0fc35a7f63d87317c3e5145a3df69700656f283328299a3b8545263cc22494" Oct 02 02:15:33 crc kubenswrapper[4885]: E1002 02:15:33.047904 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:15:41 crc kubenswrapper[4885]: I1002 02:15:41.086679 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-2tfgv"] Oct 02 02:15:41 crc kubenswrapper[4885]: I1002 02:15:41.106400 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-fjgfv"] Oct 02 02:15:41 crc kubenswrapper[4885]: I1002 02:15:41.122872 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-2tfgv"] Oct 02 02:15:41 crc kubenswrapper[4885]: I1002 02:15:41.131602 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-vl726"] Oct 02 02:15:41 crc kubenswrapper[4885]: I1002 02:15:41.139725 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-fjgfv"] Oct 02 02:15:41 crc kubenswrapper[4885]: I1002 02:15:41.148147 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-vl726"] Oct 02 02:15:42 crc kubenswrapper[4885]: I1002 02:15:42.066380 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925401e1-aa66-44fc-a81b-80d48c105c9e" path="/var/lib/kubelet/pods/925401e1-aa66-44fc-a81b-80d48c105c9e/volumes" Oct 02 02:15:42 crc kubenswrapper[4885]: I1002 02:15:42.072054 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="93487043-76c8-4e9b-a290-8a8411ccda2c" path="/var/lib/kubelet/pods/93487043-76c8-4e9b-a290-8a8411ccda2c/volumes" Oct 02 02:15:42 crc kubenswrapper[4885]: I1002 02:15:42.073360 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9ffa5a6b-478d-4f4d-967d-61b6b79a8097" path="/var/lib/kubelet/pods/9ffa5a6b-478d-4f4d-967d-61b6b79a8097/volumes" Oct 02 02:15:44 crc kubenswrapper[4885]: I1002 02:15:44.047743 4885 scope.go:117] "RemoveContainer" containerID="be0fc35a7f63d87317c3e5145a3df69700656f283328299a3b8545263cc22494" Oct 02 02:15:44 crc kubenswrapper[4885]: E1002 02:15:44.049069 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:15:51 crc kubenswrapper[4885]: I1002 02:15:51.048867 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-3f2e-account-create-lp5n9"] Oct 02 02:15:51 crc kubenswrapper[4885]: I1002 02:15:51.065889 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-5b7d-account-create-r6fp9"] Oct 02 02:15:51 crc kubenswrapper[4885]: I1002 02:15:51.075967 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-b3c0-account-create-tcx7f"] Oct 02 02:15:51 crc kubenswrapper[4885]: I1002 02:15:51.086484 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-5b7d-account-create-r6fp9"] Oct 02 02:15:51 crc kubenswrapper[4885]: I1002 02:15:51.096517 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-3f2e-account-create-lp5n9"] Oct 02 02:15:51 crc kubenswrapper[4885]: I1002 02:15:51.104178 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-b3c0-account-create-tcx7f"] Oct 02 02:15:52 crc kubenswrapper[4885]: I1002 02:15:52.078179 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5e1ee780-2c31-401f-9822-cc58ca573908" path="/var/lib/kubelet/pods/5e1ee780-2c31-401f-9822-cc58ca573908/volumes" Oct 02 02:15:52 crc kubenswrapper[4885]: I1002 02:15:52.078966 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6d1c46db-7827-474e-8dec-7b3fb2070449" path="/var/lib/kubelet/pods/6d1c46db-7827-474e-8dec-7b3fb2070449/volumes" Oct 02 02:15:52 crc kubenswrapper[4885]: I1002 02:15:52.079646 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af2b6dad-dd76-4c03-9902-d190e8a5949e" path="/var/lib/kubelet/pods/af2b6dad-dd76-4c03-9902-d190e8a5949e/volumes" Oct 02 02:15:59 crc kubenswrapper[4885]: I1002 02:15:59.046644 4885 scope.go:117] "RemoveContainer" containerID="be0fc35a7f63d87317c3e5145a3df69700656f283328299a3b8545263cc22494" Oct 02 02:15:59 crc kubenswrapper[4885]: E1002 02:15:59.047445 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:16:00 crc kubenswrapper[4885]: I1002 02:16:00.259522 4885 generic.go:334] "Generic (PLEG): container finished" podID="48ef0e0a-ae52-4d11-9023-d53f4ab716ac" containerID="8710cc62cd5f897ef3c5772309b65e4c6ebb45c15e66d996b8b5bafb53d8df90" exitCode=0 Oct 02 02:16:00 crc kubenswrapper[4885]: I1002 02:16:00.259725 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-p9p9g" event={"ID":"48ef0e0a-ae52-4d11-9023-d53f4ab716ac","Type":"ContainerDied","Data":"8710cc62cd5f897ef3c5772309b65e4c6ebb45c15e66d996b8b5bafb53d8df90"} Oct 02 02:16:01 crc kubenswrapper[4885]: I1002 02:16:01.825320 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-p9p9g" Oct 02 02:16:01 crc kubenswrapper[4885]: I1002 02:16:01.956802 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wm8s7\" (UniqueName: \"kubernetes.io/projected/48ef0e0a-ae52-4d11-9023-d53f4ab716ac-kube-api-access-wm8s7\") pod \"48ef0e0a-ae52-4d11-9023-d53f4ab716ac\" (UID: \"48ef0e0a-ae52-4d11-9023-d53f4ab716ac\") " Oct 02 02:16:01 crc kubenswrapper[4885]: I1002 02:16:01.957003 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/48ef0e0a-ae52-4d11-9023-d53f4ab716ac-inventory\") pod \"48ef0e0a-ae52-4d11-9023-d53f4ab716ac\" (UID: \"48ef0e0a-ae52-4d11-9023-d53f4ab716ac\") " Oct 02 02:16:01 crc kubenswrapper[4885]: I1002 02:16:01.957089 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/48ef0e0a-ae52-4d11-9023-d53f4ab716ac-ssh-key\") pod \"48ef0e0a-ae52-4d11-9023-d53f4ab716ac\" (UID: \"48ef0e0a-ae52-4d11-9023-d53f4ab716ac\") " Oct 02 02:16:01 crc kubenswrapper[4885]: I1002 02:16:01.967740 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48ef0e0a-ae52-4d11-9023-d53f4ab716ac-kube-api-access-wm8s7" (OuterVolumeSpecName: "kube-api-access-wm8s7") pod "48ef0e0a-ae52-4d11-9023-d53f4ab716ac" (UID: "48ef0e0a-ae52-4d11-9023-d53f4ab716ac"). InnerVolumeSpecName "kube-api-access-wm8s7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:16:01 crc kubenswrapper[4885]: I1002 02:16:01.996061 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48ef0e0a-ae52-4d11-9023-d53f4ab716ac-inventory" (OuterVolumeSpecName: "inventory") pod "48ef0e0a-ae52-4d11-9023-d53f4ab716ac" (UID: "48ef0e0a-ae52-4d11-9023-d53f4ab716ac"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:16:02 crc kubenswrapper[4885]: I1002 02:16:02.023489 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48ef0e0a-ae52-4d11-9023-d53f4ab716ac-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "48ef0e0a-ae52-4d11-9023-d53f4ab716ac" (UID: "48ef0e0a-ae52-4d11-9023-d53f4ab716ac"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:16:02 crc kubenswrapper[4885]: I1002 02:16:02.060102 4885 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/48ef0e0a-ae52-4d11-9023-d53f4ab716ac-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 02:16:02 crc kubenswrapper[4885]: I1002 02:16:02.060154 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wm8s7\" (UniqueName: \"kubernetes.io/projected/48ef0e0a-ae52-4d11-9023-d53f4ab716ac-kube-api-access-wm8s7\") on node \"crc\" DevicePath \"\"" Oct 02 02:16:02 crc kubenswrapper[4885]: I1002 02:16:02.060184 4885 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/48ef0e0a-ae52-4d11-9023-d53f4ab716ac-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 02:16:02 crc kubenswrapper[4885]: I1002 02:16:02.320050 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-p9p9g" event={"ID":"48ef0e0a-ae52-4d11-9023-d53f4ab716ac","Type":"ContainerDied","Data":"e539a6ee346edde38e7e4ad8ad9d21a85c3002ec92c06dfed294b7f36fc9ae08"} Oct 02 02:16:02 crc kubenswrapper[4885]: I1002 02:16:02.320095 4885 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e539a6ee346edde38e7e4ad8ad9d21a85c3002ec92c06dfed294b7f36fc9ae08" Oct 02 02:16:02 crc kubenswrapper[4885]: I1002 02:16:02.320161 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-p9p9g" Oct 02 02:16:02 crc kubenswrapper[4885]: I1002 02:16:02.411022 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-mv2tw"] Oct 02 02:16:02 crc kubenswrapper[4885]: E1002 02:16:02.411461 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48ef0e0a-ae52-4d11-9023-d53f4ab716ac" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 02 02:16:02 crc kubenswrapper[4885]: I1002 02:16:02.411483 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="48ef0e0a-ae52-4d11-9023-d53f4ab716ac" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 02 02:16:02 crc kubenswrapper[4885]: I1002 02:16:02.413107 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="48ef0e0a-ae52-4d11-9023-d53f4ab716ac" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 02 02:16:02 crc kubenswrapper[4885]: I1002 02:16:02.414058 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-mv2tw" Oct 02 02:16:02 crc kubenswrapper[4885]: I1002 02:16:02.420732 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 02:16:02 crc kubenswrapper[4885]: I1002 02:16:02.420770 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 02:16:02 crc kubenswrapper[4885]: I1002 02:16:02.420803 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-f8tvb" Oct 02 02:16:02 crc kubenswrapper[4885]: I1002 02:16:02.425179 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 02:16:02 crc kubenswrapper[4885]: I1002 02:16:02.428349 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-mv2tw"] Oct 02 02:16:02 crc kubenswrapper[4885]: I1002 02:16:02.569228 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7tpqt\" (UniqueName: \"kubernetes.io/projected/c6157c99-0cbf-4edb-9a16-04c99b7b2385-kube-api-access-7tpqt\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-mv2tw\" (UID: \"c6157c99-0cbf-4edb-9a16-04c99b7b2385\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-mv2tw" Oct 02 02:16:02 crc kubenswrapper[4885]: I1002 02:16:02.569852 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c6157c99-0cbf-4edb-9a16-04c99b7b2385-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-mv2tw\" (UID: \"c6157c99-0cbf-4edb-9a16-04c99b7b2385\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-mv2tw" Oct 02 02:16:02 crc kubenswrapper[4885]: I1002 02:16:02.570006 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c6157c99-0cbf-4edb-9a16-04c99b7b2385-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-mv2tw\" (UID: \"c6157c99-0cbf-4edb-9a16-04c99b7b2385\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-mv2tw" Oct 02 02:16:02 crc kubenswrapper[4885]: I1002 02:16:02.673983 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c6157c99-0cbf-4edb-9a16-04c99b7b2385-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-mv2tw\" (UID: \"c6157c99-0cbf-4edb-9a16-04c99b7b2385\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-mv2tw" Oct 02 02:16:02 crc kubenswrapper[4885]: I1002 02:16:02.674410 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7tpqt\" (UniqueName: \"kubernetes.io/projected/c6157c99-0cbf-4edb-9a16-04c99b7b2385-kube-api-access-7tpqt\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-mv2tw\" (UID: \"c6157c99-0cbf-4edb-9a16-04c99b7b2385\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-mv2tw" Oct 02 02:16:02 crc kubenswrapper[4885]: I1002 02:16:02.674813 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c6157c99-0cbf-4edb-9a16-04c99b7b2385-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-mv2tw\" (UID: \"c6157c99-0cbf-4edb-9a16-04c99b7b2385\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-mv2tw" Oct 02 02:16:02 crc kubenswrapper[4885]: I1002 02:16:02.681577 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c6157c99-0cbf-4edb-9a16-04c99b7b2385-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-mv2tw\" (UID: \"c6157c99-0cbf-4edb-9a16-04c99b7b2385\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-mv2tw" Oct 02 02:16:02 crc kubenswrapper[4885]: I1002 02:16:02.681610 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c6157c99-0cbf-4edb-9a16-04c99b7b2385-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-mv2tw\" (UID: \"c6157c99-0cbf-4edb-9a16-04c99b7b2385\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-mv2tw" Oct 02 02:16:02 crc kubenswrapper[4885]: I1002 02:16:02.703501 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7tpqt\" (UniqueName: \"kubernetes.io/projected/c6157c99-0cbf-4edb-9a16-04c99b7b2385-kube-api-access-7tpqt\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-mv2tw\" (UID: \"c6157c99-0cbf-4edb-9a16-04c99b7b2385\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-mv2tw" Oct 02 02:16:02 crc kubenswrapper[4885]: I1002 02:16:02.740450 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-mv2tw" Oct 02 02:16:03 crc kubenswrapper[4885]: I1002 02:16:03.347989 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-mv2tw"] Oct 02 02:16:04 crc kubenswrapper[4885]: I1002 02:16:04.348479 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-mv2tw" event={"ID":"c6157c99-0cbf-4edb-9a16-04c99b7b2385","Type":"ContainerStarted","Data":"099fe83fdf2d64cca3a7bf644e79fe010743c7b11fe561160ac6462e8faf6e2b"} Oct 02 02:16:05 crc kubenswrapper[4885]: I1002 02:16:05.359875 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-mv2tw" event={"ID":"c6157c99-0cbf-4edb-9a16-04c99b7b2385","Type":"ContainerStarted","Data":"4590915c3207de5e15360a0936a7eee63b8bbba49bf43485dadc329a1c9086cf"} Oct 02 02:16:05 crc kubenswrapper[4885]: I1002 02:16:05.381238 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-mv2tw" podStartSLOduration=2.328593882 podStartE2EDuration="3.381217447s" podCreationTimestamp="2025-10-02 02:16:02 +0000 UTC" firstStartedPulling="2025-10-02 02:16:03.361193243 +0000 UTC m=+1752.172940682" lastFinishedPulling="2025-10-02 02:16:04.413816808 +0000 UTC m=+1753.225564247" observedRunningTime="2025-10-02 02:16:05.375249415 +0000 UTC m=+1754.186996814" watchObservedRunningTime="2025-10-02 02:16:05.381217447 +0000 UTC m=+1754.192964866" Oct 02 02:16:12 crc kubenswrapper[4885]: I1002 02:16:12.059607 4885 scope.go:117] "RemoveContainer" containerID="be0fc35a7f63d87317c3e5145a3df69700656f283328299a3b8545263cc22494" Oct 02 02:16:12 crc kubenswrapper[4885]: E1002 02:16:12.060648 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:16:13 crc kubenswrapper[4885]: I1002 02:16:13.218934 4885 scope.go:117] "RemoveContainer" containerID="d1c077f945323e970ddc5489d47a203f6ed334052d474e1176c845e8518d0596" Oct 02 02:16:13 crc kubenswrapper[4885]: I1002 02:16:13.252098 4885 scope.go:117] "RemoveContainer" containerID="521d56adf2bb767b794788e9e4a1227efa64a91af1a488857a83713bc3a46de9" Oct 02 02:16:13 crc kubenswrapper[4885]: I1002 02:16:13.344011 4885 scope.go:117] "RemoveContainer" containerID="f2a80ac179f355c797c3ac3052f0e7b4127c9bc773ee633cb5fd43ac644ecf3a" Oct 02 02:16:13 crc kubenswrapper[4885]: I1002 02:16:13.404864 4885 scope.go:117] "RemoveContainer" containerID="f0f6b0f74803390b757a6e1c56567398bbc625990fe01c57c6c902f5bf5310c8" Oct 02 02:16:13 crc kubenswrapper[4885]: I1002 02:16:13.452751 4885 scope.go:117] "RemoveContainer" containerID="d6c5dcd3f24c7e7a3ba090cbc513323434befe12c944ae98c3d55b394329fd49" Oct 02 02:16:13 crc kubenswrapper[4885]: I1002 02:16:13.496962 4885 scope.go:117] "RemoveContainer" containerID="38e9ae7aa4568465f545b227259addc8c26bbdb84e7f74a00eacfc79ca7d54b2" Oct 02 02:16:13 crc kubenswrapper[4885]: I1002 02:16:13.534753 4885 scope.go:117] "RemoveContainer" containerID="1492fc89ff995999425a0af3d182ae6feac4ff3c58a85684452fa3c1b7ea5fd4" Oct 02 02:16:14 crc kubenswrapper[4885]: I1002 02:16:14.077015 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-zcpc9"] Oct 02 02:16:14 crc kubenswrapper[4885]: I1002 02:16:14.077384 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-zcpc9"] Oct 02 02:16:16 crc kubenswrapper[4885]: I1002 02:16:16.072188 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cb797f9e-26a8-42fb-bac6-f5c145103aca" path="/var/lib/kubelet/pods/cb797f9e-26a8-42fb-bac6-f5c145103aca/volumes" Oct 02 02:16:16 crc kubenswrapper[4885]: I1002 02:16:16.280928 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-7sx7v"] Oct 02 02:16:16 crc kubenswrapper[4885]: I1002 02:16:16.283085 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7sx7v" Oct 02 02:16:16 crc kubenswrapper[4885]: I1002 02:16:16.294909 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7sx7v"] Oct 02 02:16:16 crc kubenswrapper[4885]: I1002 02:16:16.408378 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5dkf4\" (UniqueName: \"kubernetes.io/projected/28bb6cba-e588-43e1-84da-78eae93d9cca-kube-api-access-5dkf4\") pod \"redhat-marketplace-7sx7v\" (UID: \"28bb6cba-e588-43e1-84da-78eae93d9cca\") " pod="openshift-marketplace/redhat-marketplace-7sx7v" Oct 02 02:16:16 crc kubenswrapper[4885]: I1002 02:16:16.408459 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/28bb6cba-e588-43e1-84da-78eae93d9cca-utilities\") pod \"redhat-marketplace-7sx7v\" (UID: \"28bb6cba-e588-43e1-84da-78eae93d9cca\") " pod="openshift-marketplace/redhat-marketplace-7sx7v" Oct 02 02:16:16 crc kubenswrapper[4885]: I1002 02:16:16.408850 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/28bb6cba-e588-43e1-84da-78eae93d9cca-catalog-content\") pod \"redhat-marketplace-7sx7v\" (UID: \"28bb6cba-e588-43e1-84da-78eae93d9cca\") " pod="openshift-marketplace/redhat-marketplace-7sx7v" Oct 02 02:16:16 crc kubenswrapper[4885]: I1002 02:16:16.510584 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/28bb6cba-e588-43e1-84da-78eae93d9cca-catalog-content\") pod \"redhat-marketplace-7sx7v\" (UID: \"28bb6cba-e588-43e1-84da-78eae93d9cca\") " pod="openshift-marketplace/redhat-marketplace-7sx7v" Oct 02 02:16:16 crc kubenswrapper[4885]: I1002 02:16:16.510723 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5dkf4\" (UniqueName: \"kubernetes.io/projected/28bb6cba-e588-43e1-84da-78eae93d9cca-kube-api-access-5dkf4\") pod \"redhat-marketplace-7sx7v\" (UID: \"28bb6cba-e588-43e1-84da-78eae93d9cca\") " pod="openshift-marketplace/redhat-marketplace-7sx7v" Oct 02 02:16:16 crc kubenswrapper[4885]: I1002 02:16:16.510753 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/28bb6cba-e588-43e1-84da-78eae93d9cca-utilities\") pod \"redhat-marketplace-7sx7v\" (UID: \"28bb6cba-e588-43e1-84da-78eae93d9cca\") " pod="openshift-marketplace/redhat-marketplace-7sx7v" Oct 02 02:16:16 crc kubenswrapper[4885]: I1002 02:16:16.511230 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/28bb6cba-e588-43e1-84da-78eae93d9cca-catalog-content\") pod \"redhat-marketplace-7sx7v\" (UID: \"28bb6cba-e588-43e1-84da-78eae93d9cca\") " pod="openshift-marketplace/redhat-marketplace-7sx7v" Oct 02 02:16:16 crc kubenswrapper[4885]: I1002 02:16:16.511310 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/28bb6cba-e588-43e1-84da-78eae93d9cca-utilities\") pod \"redhat-marketplace-7sx7v\" (UID: \"28bb6cba-e588-43e1-84da-78eae93d9cca\") " pod="openshift-marketplace/redhat-marketplace-7sx7v" Oct 02 02:16:16 crc kubenswrapper[4885]: I1002 02:16:16.531814 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5dkf4\" (UniqueName: \"kubernetes.io/projected/28bb6cba-e588-43e1-84da-78eae93d9cca-kube-api-access-5dkf4\") pod \"redhat-marketplace-7sx7v\" (UID: \"28bb6cba-e588-43e1-84da-78eae93d9cca\") " pod="openshift-marketplace/redhat-marketplace-7sx7v" Oct 02 02:16:16 crc kubenswrapper[4885]: I1002 02:16:16.603529 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7sx7v" Oct 02 02:16:17 crc kubenswrapper[4885]: I1002 02:16:17.132884 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7sx7v"] Oct 02 02:16:17 crc kubenswrapper[4885]: W1002 02:16:17.147035 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod28bb6cba_e588_43e1_84da_78eae93d9cca.slice/crio-aa88a1e8f368e28eb684f60af1569cfc62c8b70feecbb81117da556f8ea9c7c0 WatchSource:0}: Error finding container aa88a1e8f368e28eb684f60af1569cfc62c8b70feecbb81117da556f8ea9c7c0: Status 404 returned error can't find the container with id aa88a1e8f368e28eb684f60af1569cfc62c8b70feecbb81117da556f8ea9c7c0 Oct 02 02:16:17 crc kubenswrapper[4885]: I1002 02:16:17.494838 4885 generic.go:334] "Generic (PLEG): container finished" podID="28bb6cba-e588-43e1-84da-78eae93d9cca" containerID="a5cb1fe7827631294c837efcc5ab6e17d0acd87fb30869dd770883f4d00f5c8f" exitCode=0 Oct 02 02:16:17 crc kubenswrapper[4885]: I1002 02:16:17.494927 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7sx7v" event={"ID":"28bb6cba-e588-43e1-84da-78eae93d9cca","Type":"ContainerDied","Data":"a5cb1fe7827631294c837efcc5ab6e17d0acd87fb30869dd770883f4d00f5c8f"} Oct 02 02:16:17 crc kubenswrapper[4885]: I1002 02:16:17.495118 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7sx7v" event={"ID":"28bb6cba-e588-43e1-84da-78eae93d9cca","Type":"ContainerStarted","Data":"aa88a1e8f368e28eb684f60af1569cfc62c8b70feecbb81117da556f8ea9c7c0"} Oct 02 02:16:18 crc kubenswrapper[4885]: I1002 02:16:18.508844 4885 generic.go:334] "Generic (PLEG): container finished" podID="28bb6cba-e588-43e1-84da-78eae93d9cca" containerID="8dc9b7c87de21dc0f1acc99b577f169d70e756d2a7a410dc8006f64e6460e1ac" exitCode=0 Oct 02 02:16:18 crc kubenswrapper[4885]: I1002 02:16:18.508946 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7sx7v" event={"ID":"28bb6cba-e588-43e1-84da-78eae93d9cca","Type":"ContainerDied","Data":"8dc9b7c87de21dc0f1acc99b577f169d70e756d2a7a410dc8006f64e6460e1ac"} Oct 02 02:16:19 crc kubenswrapper[4885]: I1002 02:16:19.479191 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-ttp6v"] Oct 02 02:16:19 crc kubenswrapper[4885]: I1002 02:16:19.482992 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ttp6v" Oct 02 02:16:19 crc kubenswrapper[4885]: I1002 02:16:19.500661 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ttp6v"] Oct 02 02:16:19 crc kubenswrapper[4885]: I1002 02:16:19.524429 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7sx7v" event={"ID":"28bb6cba-e588-43e1-84da-78eae93d9cca","Type":"ContainerStarted","Data":"3e93a8141e05fbed30469b632365b694bd1c096d933b58e51d299ee6158f737d"} Oct 02 02:16:19 crc kubenswrapper[4885]: I1002 02:16:19.556316 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-7sx7v" podStartSLOduration=2.120467961 podStartE2EDuration="3.556273804s" podCreationTimestamp="2025-10-02 02:16:16 +0000 UTC" firstStartedPulling="2025-10-02 02:16:17.497160545 +0000 UTC m=+1766.308907954" lastFinishedPulling="2025-10-02 02:16:18.932966368 +0000 UTC m=+1767.744713797" observedRunningTime="2025-10-02 02:16:19.548747474 +0000 UTC m=+1768.360494873" watchObservedRunningTime="2025-10-02 02:16:19.556273804 +0000 UTC m=+1768.368021243" Oct 02 02:16:19 crc kubenswrapper[4885]: I1002 02:16:19.594546 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7f36a0d-b258-47d7-a09a-13581e0e7d1e-catalog-content\") pod \"certified-operators-ttp6v\" (UID: \"e7f36a0d-b258-47d7-a09a-13581e0e7d1e\") " pod="openshift-marketplace/certified-operators-ttp6v" Oct 02 02:16:19 crc kubenswrapper[4885]: I1002 02:16:19.594652 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7f36a0d-b258-47d7-a09a-13581e0e7d1e-utilities\") pod \"certified-operators-ttp6v\" (UID: \"e7f36a0d-b258-47d7-a09a-13581e0e7d1e\") " pod="openshift-marketplace/certified-operators-ttp6v" Oct 02 02:16:19 crc kubenswrapper[4885]: I1002 02:16:19.594997 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v9xbx\" (UniqueName: \"kubernetes.io/projected/e7f36a0d-b258-47d7-a09a-13581e0e7d1e-kube-api-access-v9xbx\") pod \"certified-operators-ttp6v\" (UID: \"e7f36a0d-b258-47d7-a09a-13581e0e7d1e\") " pod="openshift-marketplace/certified-operators-ttp6v" Oct 02 02:16:19 crc kubenswrapper[4885]: I1002 02:16:19.697066 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v9xbx\" (UniqueName: \"kubernetes.io/projected/e7f36a0d-b258-47d7-a09a-13581e0e7d1e-kube-api-access-v9xbx\") pod \"certified-operators-ttp6v\" (UID: \"e7f36a0d-b258-47d7-a09a-13581e0e7d1e\") " pod="openshift-marketplace/certified-operators-ttp6v" Oct 02 02:16:19 crc kubenswrapper[4885]: I1002 02:16:19.697546 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7f36a0d-b258-47d7-a09a-13581e0e7d1e-catalog-content\") pod \"certified-operators-ttp6v\" (UID: \"e7f36a0d-b258-47d7-a09a-13581e0e7d1e\") " pod="openshift-marketplace/certified-operators-ttp6v" Oct 02 02:16:19 crc kubenswrapper[4885]: I1002 02:16:19.697749 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7f36a0d-b258-47d7-a09a-13581e0e7d1e-utilities\") pod \"certified-operators-ttp6v\" (UID: \"e7f36a0d-b258-47d7-a09a-13581e0e7d1e\") " pod="openshift-marketplace/certified-operators-ttp6v" Oct 02 02:16:19 crc kubenswrapper[4885]: I1002 02:16:19.697967 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7f36a0d-b258-47d7-a09a-13581e0e7d1e-catalog-content\") pod \"certified-operators-ttp6v\" (UID: \"e7f36a0d-b258-47d7-a09a-13581e0e7d1e\") " pod="openshift-marketplace/certified-operators-ttp6v" Oct 02 02:16:19 crc kubenswrapper[4885]: I1002 02:16:19.698351 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7f36a0d-b258-47d7-a09a-13581e0e7d1e-utilities\") pod \"certified-operators-ttp6v\" (UID: \"e7f36a0d-b258-47d7-a09a-13581e0e7d1e\") " pod="openshift-marketplace/certified-operators-ttp6v" Oct 02 02:16:19 crc kubenswrapper[4885]: I1002 02:16:19.733918 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v9xbx\" (UniqueName: \"kubernetes.io/projected/e7f36a0d-b258-47d7-a09a-13581e0e7d1e-kube-api-access-v9xbx\") pod \"certified-operators-ttp6v\" (UID: \"e7f36a0d-b258-47d7-a09a-13581e0e7d1e\") " pod="openshift-marketplace/certified-operators-ttp6v" Oct 02 02:16:19 crc kubenswrapper[4885]: I1002 02:16:19.809006 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ttp6v" Oct 02 02:16:20 crc kubenswrapper[4885]: I1002 02:16:20.313828 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ttp6v"] Oct 02 02:16:20 crc kubenswrapper[4885]: I1002 02:16:20.534213 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ttp6v" event={"ID":"e7f36a0d-b258-47d7-a09a-13581e0e7d1e","Type":"ContainerStarted","Data":"c0d693b748377df3eebd036ea2374876420490318358e337da0446bf35793101"} Oct 02 02:16:21 crc kubenswrapper[4885]: I1002 02:16:21.548312 4885 generic.go:334] "Generic (PLEG): container finished" podID="e7f36a0d-b258-47d7-a09a-13581e0e7d1e" containerID="b0d2ec6d584bb86a842ee57086b30f28aae4a35dd2089d74997d6a3f4a3e056c" exitCode=0 Oct 02 02:16:21 crc kubenswrapper[4885]: I1002 02:16:21.548423 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ttp6v" event={"ID":"e7f36a0d-b258-47d7-a09a-13581e0e7d1e","Type":"ContainerDied","Data":"b0d2ec6d584bb86a842ee57086b30f28aae4a35dd2089d74997d6a3f4a3e056c"} Oct 02 02:16:22 crc kubenswrapper[4885]: I1002 02:16:22.563702 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ttp6v" event={"ID":"e7f36a0d-b258-47d7-a09a-13581e0e7d1e","Type":"ContainerStarted","Data":"a33313dac05267d45d43183f371f7f43d02557c15a345f7d20ba4398c27fdb93"} Oct 02 02:16:23 crc kubenswrapper[4885]: I1002 02:16:23.576128 4885 generic.go:334] "Generic (PLEG): container finished" podID="e7f36a0d-b258-47d7-a09a-13581e0e7d1e" containerID="a33313dac05267d45d43183f371f7f43d02557c15a345f7d20ba4398c27fdb93" exitCode=0 Oct 02 02:16:23 crc kubenswrapper[4885]: I1002 02:16:23.576385 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ttp6v" event={"ID":"e7f36a0d-b258-47d7-a09a-13581e0e7d1e","Type":"ContainerDied","Data":"a33313dac05267d45d43183f371f7f43d02557c15a345f7d20ba4398c27fdb93"} Oct 02 02:16:24 crc kubenswrapper[4885]: I1002 02:16:24.047001 4885 scope.go:117] "RemoveContainer" containerID="be0fc35a7f63d87317c3e5145a3df69700656f283328299a3b8545263cc22494" Oct 02 02:16:24 crc kubenswrapper[4885]: E1002 02:16:24.047621 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:16:24 crc kubenswrapper[4885]: I1002 02:16:24.591937 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ttp6v" event={"ID":"e7f36a0d-b258-47d7-a09a-13581e0e7d1e","Type":"ContainerStarted","Data":"dd0d1141b3f67d5e81de8ff9121076893b7a40b48d6b2fa68b990c29f33f39b2"} Oct 02 02:16:24 crc kubenswrapper[4885]: I1002 02:16:24.632832 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-ttp6v" podStartSLOduration=3.205864824 podStartE2EDuration="5.632809363s" podCreationTimestamp="2025-10-02 02:16:19 +0000 UTC" firstStartedPulling="2025-10-02 02:16:21.550714886 +0000 UTC m=+1770.362462325" lastFinishedPulling="2025-10-02 02:16:23.977659455 +0000 UTC m=+1772.789406864" observedRunningTime="2025-10-02 02:16:24.6156881 +0000 UTC m=+1773.427435539" watchObservedRunningTime="2025-10-02 02:16:24.632809363 +0000 UTC m=+1773.444556792" Oct 02 02:16:26 crc kubenswrapper[4885]: I1002 02:16:26.603791 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-7sx7v" Oct 02 02:16:26 crc kubenswrapper[4885]: I1002 02:16:26.604233 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-7sx7v" Oct 02 02:16:26 crc kubenswrapper[4885]: I1002 02:16:26.676580 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-7sx7v" Oct 02 02:16:27 crc kubenswrapper[4885]: I1002 02:16:27.679253 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-7sx7v" Oct 02 02:16:28 crc kubenswrapper[4885]: I1002 02:16:28.474612 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7sx7v"] Oct 02 02:16:29 crc kubenswrapper[4885]: I1002 02:16:29.646046 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-7sx7v" podUID="28bb6cba-e588-43e1-84da-78eae93d9cca" containerName="registry-server" containerID="cri-o://3e93a8141e05fbed30469b632365b694bd1c096d933b58e51d299ee6158f737d" gracePeriod=2 Oct 02 02:16:29 crc kubenswrapper[4885]: I1002 02:16:29.809800 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-ttp6v" Oct 02 02:16:29 crc kubenswrapper[4885]: I1002 02:16:29.809893 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-ttp6v" Oct 02 02:16:29 crc kubenswrapper[4885]: I1002 02:16:29.907721 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-ttp6v" Oct 02 02:16:30 crc kubenswrapper[4885]: I1002 02:16:30.210366 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7sx7v" Oct 02 02:16:30 crc kubenswrapper[4885]: I1002 02:16:30.325867 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/28bb6cba-e588-43e1-84da-78eae93d9cca-utilities\") pod \"28bb6cba-e588-43e1-84da-78eae93d9cca\" (UID: \"28bb6cba-e588-43e1-84da-78eae93d9cca\") " Oct 02 02:16:30 crc kubenswrapper[4885]: I1002 02:16:30.328778 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/28bb6cba-e588-43e1-84da-78eae93d9cca-catalog-content\") pod \"28bb6cba-e588-43e1-84da-78eae93d9cca\" (UID: \"28bb6cba-e588-43e1-84da-78eae93d9cca\") " Oct 02 02:16:30 crc kubenswrapper[4885]: I1002 02:16:30.328500 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/28bb6cba-e588-43e1-84da-78eae93d9cca-utilities" (OuterVolumeSpecName: "utilities") pod "28bb6cba-e588-43e1-84da-78eae93d9cca" (UID: "28bb6cba-e588-43e1-84da-78eae93d9cca"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:16:30 crc kubenswrapper[4885]: I1002 02:16:30.339463 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5dkf4\" (UniqueName: \"kubernetes.io/projected/28bb6cba-e588-43e1-84da-78eae93d9cca-kube-api-access-5dkf4\") pod \"28bb6cba-e588-43e1-84da-78eae93d9cca\" (UID: \"28bb6cba-e588-43e1-84da-78eae93d9cca\") " Oct 02 02:16:30 crc kubenswrapper[4885]: I1002 02:16:30.340341 4885 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/28bb6cba-e588-43e1-84da-78eae93d9cca-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 02:16:30 crc kubenswrapper[4885]: I1002 02:16:30.354090 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/28bb6cba-e588-43e1-84da-78eae93d9cca-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "28bb6cba-e588-43e1-84da-78eae93d9cca" (UID: "28bb6cba-e588-43e1-84da-78eae93d9cca"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:16:30 crc kubenswrapper[4885]: I1002 02:16:30.355573 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/28bb6cba-e588-43e1-84da-78eae93d9cca-kube-api-access-5dkf4" (OuterVolumeSpecName: "kube-api-access-5dkf4") pod "28bb6cba-e588-43e1-84da-78eae93d9cca" (UID: "28bb6cba-e588-43e1-84da-78eae93d9cca"). InnerVolumeSpecName "kube-api-access-5dkf4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:16:30 crc kubenswrapper[4885]: I1002 02:16:30.441531 4885 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/28bb6cba-e588-43e1-84da-78eae93d9cca-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 02:16:30 crc kubenswrapper[4885]: I1002 02:16:30.441558 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5dkf4\" (UniqueName: \"kubernetes.io/projected/28bb6cba-e588-43e1-84da-78eae93d9cca-kube-api-access-5dkf4\") on node \"crc\" DevicePath \"\"" Oct 02 02:16:30 crc kubenswrapper[4885]: I1002 02:16:30.661168 4885 generic.go:334] "Generic (PLEG): container finished" podID="28bb6cba-e588-43e1-84da-78eae93d9cca" containerID="3e93a8141e05fbed30469b632365b694bd1c096d933b58e51d299ee6158f737d" exitCode=0 Oct 02 02:16:30 crc kubenswrapper[4885]: I1002 02:16:30.661236 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7sx7v" Oct 02 02:16:30 crc kubenswrapper[4885]: I1002 02:16:30.661288 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7sx7v" event={"ID":"28bb6cba-e588-43e1-84da-78eae93d9cca","Type":"ContainerDied","Data":"3e93a8141e05fbed30469b632365b694bd1c096d933b58e51d299ee6158f737d"} Oct 02 02:16:30 crc kubenswrapper[4885]: I1002 02:16:30.662582 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7sx7v" event={"ID":"28bb6cba-e588-43e1-84da-78eae93d9cca","Type":"ContainerDied","Data":"aa88a1e8f368e28eb684f60af1569cfc62c8b70feecbb81117da556f8ea9c7c0"} Oct 02 02:16:30 crc kubenswrapper[4885]: I1002 02:16:30.662642 4885 scope.go:117] "RemoveContainer" containerID="3e93a8141e05fbed30469b632365b694bd1c096d933b58e51d299ee6158f737d" Oct 02 02:16:30 crc kubenswrapper[4885]: I1002 02:16:30.710341 4885 scope.go:117] "RemoveContainer" containerID="8dc9b7c87de21dc0f1acc99b577f169d70e756d2a7a410dc8006f64e6460e1ac" Oct 02 02:16:30 crc kubenswrapper[4885]: I1002 02:16:30.714656 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7sx7v"] Oct 02 02:16:30 crc kubenswrapper[4885]: I1002 02:16:30.727794 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-7sx7v"] Oct 02 02:16:30 crc kubenswrapper[4885]: I1002 02:16:30.749682 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-ttp6v" Oct 02 02:16:30 crc kubenswrapper[4885]: I1002 02:16:30.757490 4885 scope.go:117] "RemoveContainer" containerID="a5cb1fe7827631294c837efcc5ab6e17d0acd87fb30869dd770883f4d00f5c8f" Oct 02 02:16:30 crc kubenswrapper[4885]: I1002 02:16:30.823751 4885 scope.go:117] "RemoveContainer" containerID="3e93a8141e05fbed30469b632365b694bd1c096d933b58e51d299ee6158f737d" Oct 02 02:16:30 crc kubenswrapper[4885]: E1002 02:16:30.824327 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3e93a8141e05fbed30469b632365b694bd1c096d933b58e51d299ee6158f737d\": container with ID starting with 3e93a8141e05fbed30469b632365b694bd1c096d933b58e51d299ee6158f737d not found: ID does not exist" containerID="3e93a8141e05fbed30469b632365b694bd1c096d933b58e51d299ee6158f737d" Oct 02 02:16:30 crc kubenswrapper[4885]: I1002 02:16:30.824369 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3e93a8141e05fbed30469b632365b694bd1c096d933b58e51d299ee6158f737d"} err="failed to get container status \"3e93a8141e05fbed30469b632365b694bd1c096d933b58e51d299ee6158f737d\": rpc error: code = NotFound desc = could not find container \"3e93a8141e05fbed30469b632365b694bd1c096d933b58e51d299ee6158f737d\": container with ID starting with 3e93a8141e05fbed30469b632365b694bd1c096d933b58e51d299ee6158f737d not found: ID does not exist" Oct 02 02:16:30 crc kubenswrapper[4885]: I1002 02:16:30.824395 4885 scope.go:117] "RemoveContainer" containerID="8dc9b7c87de21dc0f1acc99b577f169d70e756d2a7a410dc8006f64e6460e1ac" Oct 02 02:16:30 crc kubenswrapper[4885]: E1002 02:16:30.825038 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8dc9b7c87de21dc0f1acc99b577f169d70e756d2a7a410dc8006f64e6460e1ac\": container with ID starting with 8dc9b7c87de21dc0f1acc99b577f169d70e756d2a7a410dc8006f64e6460e1ac not found: ID does not exist" containerID="8dc9b7c87de21dc0f1acc99b577f169d70e756d2a7a410dc8006f64e6460e1ac" Oct 02 02:16:30 crc kubenswrapper[4885]: I1002 02:16:30.825111 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8dc9b7c87de21dc0f1acc99b577f169d70e756d2a7a410dc8006f64e6460e1ac"} err="failed to get container status \"8dc9b7c87de21dc0f1acc99b577f169d70e756d2a7a410dc8006f64e6460e1ac\": rpc error: code = NotFound desc = could not find container \"8dc9b7c87de21dc0f1acc99b577f169d70e756d2a7a410dc8006f64e6460e1ac\": container with ID starting with 8dc9b7c87de21dc0f1acc99b577f169d70e756d2a7a410dc8006f64e6460e1ac not found: ID does not exist" Oct 02 02:16:30 crc kubenswrapper[4885]: I1002 02:16:30.825136 4885 scope.go:117] "RemoveContainer" containerID="a5cb1fe7827631294c837efcc5ab6e17d0acd87fb30869dd770883f4d00f5c8f" Oct 02 02:16:30 crc kubenswrapper[4885]: E1002 02:16:30.825627 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a5cb1fe7827631294c837efcc5ab6e17d0acd87fb30869dd770883f4d00f5c8f\": container with ID starting with a5cb1fe7827631294c837efcc5ab6e17d0acd87fb30869dd770883f4d00f5c8f not found: ID does not exist" containerID="a5cb1fe7827631294c837efcc5ab6e17d0acd87fb30869dd770883f4d00f5c8f" Oct 02 02:16:30 crc kubenswrapper[4885]: I1002 02:16:30.825660 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a5cb1fe7827631294c837efcc5ab6e17d0acd87fb30869dd770883f4d00f5c8f"} err="failed to get container status \"a5cb1fe7827631294c837efcc5ab6e17d0acd87fb30869dd770883f4d00f5c8f\": rpc error: code = NotFound desc = could not find container \"a5cb1fe7827631294c837efcc5ab6e17d0acd87fb30869dd770883f4d00f5c8f\": container with ID starting with a5cb1fe7827631294c837efcc5ab6e17d0acd87fb30869dd770883f4d00f5c8f not found: ID does not exist" Oct 02 02:16:31 crc kubenswrapper[4885]: I1002 02:16:31.465094 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-ttp6v"] Oct 02 02:16:32 crc kubenswrapper[4885]: I1002 02:16:32.069661 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="28bb6cba-e588-43e1-84da-78eae93d9cca" path="/var/lib/kubelet/pods/28bb6cba-e588-43e1-84da-78eae93d9cca/volumes" Oct 02 02:16:32 crc kubenswrapper[4885]: I1002 02:16:32.685431 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-ttp6v" podUID="e7f36a0d-b258-47d7-a09a-13581e0e7d1e" containerName="registry-server" containerID="cri-o://dd0d1141b3f67d5e81de8ff9121076893b7a40b48d6b2fa68b990c29f33f39b2" gracePeriod=2 Oct 02 02:16:33 crc kubenswrapper[4885]: I1002 02:16:33.229078 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ttp6v" Oct 02 02:16:33 crc kubenswrapper[4885]: I1002 02:16:33.326138 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7f36a0d-b258-47d7-a09a-13581e0e7d1e-catalog-content\") pod \"e7f36a0d-b258-47d7-a09a-13581e0e7d1e\" (UID: \"e7f36a0d-b258-47d7-a09a-13581e0e7d1e\") " Oct 02 02:16:33 crc kubenswrapper[4885]: I1002 02:16:33.326309 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v9xbx\" (UniqueName: \"kubernetes.io/projected/e7f36a0d-b258-47d7-a09a-13581e0e7d1e-kube-api-access-v9xbx\") pod \"e7f36a0d-b258-47d7-a09a-13581e0e7d1e\" (UID: \"e7f36a0d-b258-47d7-a09a-13581e0e7d1e\") " Oct 02 02:16:33 crc kubenswrapper[4885]: I1002 02:16:33.326432 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7f36a0d-b258-47d7-a09a-13581e0e7d1e-utilities\") pod \"e7f36a0d-b258-47d7-a09a-13581e0e7d1e\" (UID: \"e7f36a0d-b258-47d7-a09a-13581e0e7d1e\") " Oct 02 02:16:33 crc kubenswrapper[4885]: I1002 02:16:33.327640 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e7f36a0d-b258-47d7-a09a-13581e0e7d1e-utilities" (OuterVolumeSpecName: "utilities") pod "e7f36a0d-b258-47d7-a09a-13581e0e7d1e" (UID: "e7f36a0d-b258-47d7-a09a-13581e0e7d1e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:16:33 crc kubenswrapper[4885]: I1002 02:16:33.335367 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7f36a0d-b258-47d7-a09a-13581e0e7d1e-kube-api-access-v9xbx" (OuterVolumeSpecName: "kube-api-access-v9xbx") pod "e7f36a0d-b258-47d7-a09a-13581e0e7d1e" (UID: "e7f36a0d-b258-47d7-a09a-13581e0e7d1e"). InnerVolumeSpecName "kube-api-access-v9xbx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:16:33 crc kubenswrapper[4885]: I1002 02:16:33.379320 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e7f36a0d-b258-47d7-a09a-13581e0e7d1e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e7f36a0d-b258-47d7-a09a-13581e0e7d1e" (UID: "e7f36a0d-b258-47d7-a09a-13581e0e7d1e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:16:33 crc kubenswrapper[4885]: I1002 02:16:33.429203 4885 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7f36a0d-b258-47d7-a09a-13581e0e7d1e-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 02:16:33 crc kubenswrapper[4885]: I1002 02:16:33.429252 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v9xbx\" (UniqueName: \"kubernetes.io/projected/e7f36a0d-b258-47d7-a09a-13581e0e7d1e-kube-api-access-v9xbx\") on node \"crc\" DevicePath \"\"" Oct 02 02:16:33 crc kubenswrapper[4885]: I1002 02:16:33.429304 4885 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7f36a0d-b258-47d7-a09a-13581e0e7d1e-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 02:16:33 crc kubenswrapper[4885]: I1002 02:16:33.702319 4885 generic.go:334] "Generic (PLEG): container finished" podID="e7f36a0d-b258-47d7-a09a-13581e0e7d1e" containerID="dd0d1141b3f67d5e81de8ff9121076893b7a40b48d6b2fa68b990c29f33f39b2" exitCode=0 Oct 02 02:16:33 crc kubenswrapper[4885]: I1002 02:16:33.702402 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ttp6v" event={"ID":"e7f36a0d-b258-47d7-a09a-13581e0e7d1e","Type":"ContainerDied","Data":"dd0d1141b3f67d5e81de8ff9121076893b7a40b48d6b2fa68b990c29f33f39b2"} Oct 02 02:16:33 crc kubenswrapper[4885]: I1002 02:16:33.702475 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ttp6v" event={"ID":"e7f36a0d-b258-47d7-a09a-13581e0e7d1e","Type":"ContainerDied","Data":"c0d693b748377df3eebd036ea2374876420490318358e337da0446bf35793101"} Oct 02 02:16:33 crc kubenswrapper[4885]: I1002 02:16:33.702509 4885 scope.go:117] "RemoveContainer" containerID="dd0d1141b3f67d5e81de8ff9121076893b7a40b48d6b2fa68b990c29f33f39b2" Oct 02 02:16:33 crc kubenswrapper[4885]: I1002 02:16:33.702421 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ttp6v" Oct 02 02:16:33 crc kubenswrapper[4885]: I1002 02:16:33.747626 4885 scope.go:117] "RemoveContainer" containerID="a33313dac05267d45d43183f371f7f43d02557c15a345f7d20ba4398c27fdb93" Oct 02 02:16:33 crc kubenswrapper[4885]: I1002 02:16:33.772220 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-ttp6v"] Oct 02 02:16:33 crc kubenswrapper[4885]: I1002 02:16:33.781102 4885 scope.go:117] "RemoveContainer" containerID="b0d2ec6d584bb86a842ee57086b30f28aae4a35dd2089d74997d6a3f4a3e056c" Oct 02 02:16:33 crc kubenswrapper[4885]: I1002 02:16:33.790410 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-ttp6v"] Oct 02 02:16:33 crc kubenswrapper[4885]: I1002 02:16:33.864427 4885 scope.go:117] "RemoveContainer" containerID="dd0d1141b3f67d5e81de8ff9121076893b7a40b48d6b2fa68b990c29f33f39b2" Oct 02 02:16:33 crc kubenswrapper[4885]: E1002 02:16:33.865294 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dd0d1141b3f67d5e81de8ff9121076893b7a40b48d6b2fa68b990c29f33f39b2\": container with ID starting with dd0d1141b3f67d5e81de8ff9121076893b7a40b48d6b2fa68b990c29f33f39b2 not found: ID does not exist" containerID="dd0d1141b3f67d5e81de8ff9121076893b7a40b48d6b2fa68b990c29f33f39b2" Oct 02 02:16:33 crc kubenswrapper[4885]: I1002 02:16:33.865348 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd0d1141b3f67d5e81de8ff9121076893b7a40b48d6b2fa68b990c29f33f39b2"} err="failed to get container status \"dd0d1141b3f67d5e81de8ff9121076893b7a40b48d6b2fa68b990c29f33f39b2\": rpc error: code = NotFound desc = could not find container \"dd0d1141b3f67d5e81de8ff9121076893b7a40b48d6b2fa68b990c29f33f39b2\": container with ID starting with dd0d1141b3f67d5e81de8ff9121076893b7a40b48d6b2fa68b990c29f33f39b2 not found: ID does not exist" Oct 02 02:16:33 crc kubenswrapper[4885]: I1002 02:16:33.865381 4885 scope.go:117] "RemoveContainer" containerID="a33313dac05267d45d43183f371f7f43d02557c15a345f7d20ba4398c27fdb93" Oct 02 02:16:33 crc kubenswrapper[4885]: E1002 02:16:33.865864 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a33313dac05267d45d43183f371f7f43d02557c15a345f7d20ba4398c27fdb93\": container with ID starting with a33313dac05267d45d43183f371f7f43d02557c15a345f7d20ba4398c27fdb93 not found: ID does not exist" containerID="a33313dac05267d45d43183f371f7f43d02557c15a345f7d20ba4398c27fdb93" Oct 02 02:16:33 crc kubenswrapper[4885]: I1002 02:16:33.865904 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a33313dac05267d45d43183f371f7f43d02557c15a345f7d20ba4398c27fdb93"} err="failed to get container status \"a33313dac05267d45d43183f371f7f43d02557c15a345f7d20ba4398c27fdb93\": rpc error: code = NotFound desc = could not find container \"a33313dac05267d45d43183f371f7f43d02557c15a345f7d20ba4398c27fdb93\": container with ID starting with a33313dac05267d45d43183f371f7f43d02557c15a345f7d20ba4398c27fdb93 not found: ID does not exist" Oct 02 02:16:33 crc kubenswrapper[4885]: I1002 02:16:33.865933 4885 scope.go:117] "RemoveContainer" containerID="b0d2ec6d584bb86a842ee57086b30f28aae4a35dd2089d74997d6a3f4a3e056c" Oct 02 02:16:33 crc kubenswrapper[4885]: E1002 02:16:33.866197 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b0d2ec6d584bb86a842ee57086b30f28aae4a35dd2089d74997d6a3f4a3e056c\": container with ID starting with b0d2ec6d584bb86a842ee57086b30f28aae4a35dd2089d74997d6a3f4a3e056c not found: ID does not exist" containerID="b0d2ec6d584bb86a842ee57086b30f28aae4a35dd2089d74997d6a3f4a3e056c" Oct 02 02:16:33 crc kubenswrapper[4885]: I1002 02:16:33.866228 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b0d2ec6d584bb86a842ee57086b30f28aae4a35dd2089d74997d6a3f4a3e056c"} err="failed to get container status \"b0d2ec6d584bb86a842ee57086b30f28aae4a35dd2089d74997d6a3f4a3e056c\": rpc error: code = NotFound desc = could not find container \"b0d2ec6d584bb86a842ee57086b30f28aae4a35dd2089d74997d6a3f4a3e056c\": container with ID starting with b0d2ec6d584bb86a842ee57086b30f28aae4a35dd2089d74997d6a3f4a3e056c not found: ID does not exist" Oct 02 02:16:34 crc kubenswrapper[4885]: I1002 02:16:34.067512 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7f36a0d-b258-47d7-a09a-13581e0e7d1e" path="/var/lib/kubelet/pods/e7f36a0d-b258-47d7-a09a-13581e0e7d1e/volumes" Oct 02 02:16:35 crc kubenswrapper[4885]: I1002 02:16:35.041563 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-t94gt"] Oct 02 02:16:35 crc kubenswrapper[4885]: I1002 02:16:35.047395 4885 scope.go:117] "RemoveContainer" containerID="be0fc35a7f63d87317c3e5145a3df69700656f283328299a3b8545263cc22494" Oct 02 02:16:35 crc kubenswrapper[4885]: E1002 02:16:35.047973 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:16:35 crc kubenswrapper[4885]: I1002 02:16:35.050697 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-t94gt"] Oct 02 02:16:36 crc kubenswrapper[4885]: I1002 02:16:36.067559 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="50bc98ab-1629-4591-8e85-bfa8f17962dc" path="/var/lib/kubelet/pods/50bc98ab-1629-4591-8e85-bfa8f17962dc/volumes" Oct 02 02:16:40 crc kubenswrapper[4885]: I1002 02:16:40.038642 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-cm2bt"] Oct 02 02:16:40 crc kubenswrapper[4885]: I1002 02:16:40.068179 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-cm2bt"] Oct 02 02:16:42 crc kubenswrapper[4885]: I1002 02:16:42.072966 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ffc2d43b-ac70-44d3-9fa5-3dc0f6159f46" path="/var/lib/kubelet/pods/ffc2d43b-ac70-44d3-9fa5-3dc0f6159f46/volumes" Oct 02 02:16:50 crc kubenswrapper[4885]: I1002 02:16:50.047093 4885 scope.go:117] "RemoveContainer" containerID="be0fc35a7f63d87317c3e5145a3df69700656f283328299a3b8545263cc22494" Oct 02 02:16:50 crc kubenswrapper[4885]: E1002 02:16:50.047913 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:16:59 crc kubenswrapper[4885]: I1002 02:16:59.008374 4885 generic.go:334] "Generic (PLEG): container finished" podID="c6157c99-0cbf-4edb-9a16-04c99b7b2385" containerID="4590915c3207de5e15360a0936a7eee63b8bbba49bf43485dadc329a1c9086cf" exitCode=2 Oct 02 02:16:59 crc kubenswrapper[4885]: I1002 02:16:59.008889 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-mv2tw" event={"ID":"c6157c99-0cbf-4edb-9a16-04c99b7b2385","Type":"ContainerDied","Data":"4590915c3207de5e15360a0936a7eee63b8bbba49bf43485dadc329a1c9086cf"} Oct 02 02:17:00 crc kubenswrapper[4885]: I1002 02:17:00.533213 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-mv2tw" Oct 02 02:17:00 crc kubenswrapper[4885]: I1002 02:17:00.596174 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c6157c99-0cbf-4edb-9a16-04c99b7b2385-ssh-key\") pod \"c6157c99-0cbf-4edb-9a16-04c99b7b2385\" (UID: \"c6157c99-0cbf-4edb-9a16-04c99b7b2385\") " Oct 02 02:17:00 crc kubenswrapper[4885]: I1002 02:17:00.596299 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7tpqt\" (UniqueName: \"kubernetes.io/projected/c6157c99-0cbf-4edb-9a16-04c99b7b2385-kube-api-access-7tpqt\") pod \"c6157c99-0cbf-4edb-9a16-04c99b7b2385\" (UID: \"c6157c99-0cbf-4edb-9a16-04c99b7b2385\") " Oct 02 02:17:00 crc kubenswrapper[4885]: I1002 02:17:00.596360 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c6157c99-0cbf-4edb-9a16-04c99b7b2385-inventory\") pod \"c6157c99-0cbf-4edb-9a16-04c99b7b2385\" (UID: \"c6157c99-0cbf-4edb-9a16-04c99b7b2385\") " Oct 02 02:17:00 crc kubenswrapper[4885]: I1002 02:17:00.602066 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c6157c99-0cbf-4edb-9a16-04c99b7b2385-kube-api-access-7tpqt" (OuterVolumeSpecName: "kube-api-access-7tpqt") pod "c6157c99-0cbf-4edb-9a16-04c99b7b2385" (UID: "c6157c99-0cbf-4edb-9a16-04c99b7b2385"). InnerVolumeSpecName "kube-api-access-7tpqt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:17:00 crc kubenswrapper[4885]: I1002 02:17:00.625125 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6157c99-0cbf-4edb-9a16-04c99b7b2385-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c6157c99-0cbf-4edb-9a16-04c99b7b2385" (UID: "c6157c99-0cbf-4edb-9a16-04c99b7b2385"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:17:00 crc kubenswrapper[4885]: I1002 02:17:00.640745 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6157c99-0cbf-4edb-9a16-04c99b7b2385-inventory" (OuterVolumeSpecName: "inventory") pod "c6157c99-0cbf-4edb-9a16-04c99b7b2385" (UID: "c6157c99-0cbf-4edb-9a16-04c99b7b2385"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:17:00 crc kubenswrapper[4885]: I1002 02:17:00.698427 4885 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c6157c99-0cbf-4edb-9a16-04c99b7b2385-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 02:17:00 crc kubenswrapper[4885]: I1002 02:17:00.698461 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7tpqt\" (UniqueName: \"kubernetes.io/projected/c6157c99-0cbf-4edb-9a16-04c99b7b2385-kube-api-access-7tpqt\") on node \"crc\" DevicePath \"\"" Oct 02 02:17:00 crc kubenswrapper[4885]: I1002 02:17:00.698474 4885 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c6157c99-0cbf-4edb-9a16-04c99b7b2385-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 02:17:01 crc kubenswrapper[4885]: I1002 02:17:01.034628 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-mv2tw" event={"ID":"c6157c99-0cbf-4edb-9a16-04c99b7b2385","Type":"ContainerDied","Data":"099fe83fdf2d64cca3a7bf644e79fe010743c7b11fe561160ac6462e8faf6e2b"} Oct 02 02:17:01 crc kubenswrapper[4885]: I1002 02:17:01.035071 4885 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="099fe83fdf2d64cca3a7bf644e79fe010743c7b11fe561160ac6462e8faf6e2b" Oct 02 02:17:01 crc kubenswrapper[4885]: I1002 02:17:01.034740 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-mv2tw" Oct 02 02:17:01 crc kubenswrapper[4885]: E1002 02:17:01.337572 4885 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc6157c99_0cbf_4edb_9a16_04c99b7b2385.slice/crio-099fe83fdf2d64cca3a7bf644e79fe010743c7b11fe561160ac6462e8faf6e2b\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc6157c99_0cbf_4edb_9a16_04c99b7b2385.slice\": RecentStats: unable to find data in memory cache]" Oct 02 02:17:04 crc kubenswrapper[4885]: I1002 02:17:04.047046 4885 scope.go:117] "RemoveContainer" containerID="be0fc35a7f63d87317c3e5145a3df69700656f283328299a3b8545263cc22494" Oct 02 02:17:04 crc kubenswrapper[4885]: E1002 02:17:04.047947 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:17:08 crc kubenswrapper[4885]: I1002 02:17:08.032608 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wvh9j"] Oct 02 02:17:08 crc kubenswrapper[4885]: E1002 02:17:08.033393 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7f36a0d-b258-47d7-a09a-13581e0e7d1e" containerName="extract-content" Oct 02 02:17:08 crc kubenswrapper[4885]: I1002 02:17:08.033407 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7f36a0d-b258-47d7-a09a-13581e0e7d1e" containerName="extract-content" Oct 02 02:17:08 crc kubenswrapper[4885]: E1002 02:17:08.033421 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6157c99-0cbf-4edb-9a16-04c99b7b2385" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 02 02:17:08 crc kubenswrapper[4885]: I1002 02:17:08.033429 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6157c99-0cbf-4edb-9a16-04c99b7b2385" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 02 02:17:08 crc kubenswrapper[4885]: E1002 02:17:08.033448 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7f36a0d-b258-47d7-a09a-13581e0e7d1e" containerName="registry-server" Oct 02 02:17:08 crc kubenswrapper[4885]: I1002 02:17:08.033458 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7f36a0d-b258-47d7-a09a-13581e0e7d1e" containerName="registry-server" Oct 02 02:17:08 crc kubenswrapper[4885]: E1002 02:17:08.033478 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28bb6cba-e588-43e1-84da-78eae93d9cca" containerName="registry-server" Oct 02 02:17:08 crc kubenswrapper[4885]: I1002 02:17:08.033487 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="28bb6cba-e588-43e1-84da-78eae93d9cca" containerName="registry-server" Oct 02 02:17:08 crc kubenswrapper[4885]: E1002 02:17:08.033507 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7f36a0d-b258-47d7-a09a-13581e0e7d1e" containerName="extract-utilities" Oct 02 02:17:08 crc kubenswrapper[4885]: I1002 02:17:08.033514 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7f36a0d-b258-47d7-a09a-13581e0e7d1e" containerName="extract-utilities" Oct 02 02:17:08 crc kubenswrapper[4885]: E1002 02:17:08.033537 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28bb6cba-e588-43e1-84da-78eae93d9cca" containerName="extract-utilities" Oct 02 02:17:08 crc kubenswrapper[4885]: I1002 02:17:08.033544 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="28bb6cba-e588-43e1-84da-78eae93d9cca" containerName="extract-utilities" Oct 02 02:17:08 crc kubenswrapper[4885]: E1002 02:17:08.033553 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28bb6cba-e588-43e1-84da-78eae93d9cca" containerName="extract-content" Oct 02 02:17:08 crc kubenswrapper[4885]: I1002 02:17:08.033560 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="28bb6cba-e588-43e1-84da-78eae93d9cca" containerName="extract-content" Oct 02 02:17:08 crc kubenswrapper[4885]: I1002 02:17:08.033780 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="28bb6cba-e588-43e1-84da-78eae93d9cca" containerName="registry-server" Oct 02 02:17:08 crc kubenswrapper[4885]: I1002 02:17:08.033796 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="e7f36a0d-b258-47d7-a09a-13581e0e7d1e" containerName="registry-server" Oct 02 02:17:08 crc kubenswrapper[4885]: I1002 02:17:08.033824 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6157c99-0cbf-4edb-9a16-04c99b7b2385" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 02 02:17:08 crc kubenswrapper[4885]: I1002 02:17:08.034630 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wvh9j" Oct 02 02:17:08 crc kubenswrapper[4885]: I1002 02:17:08.037688 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 02:17:08 crc kubenswrapper[4885]: I1002 02:17:08.038387 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 02:17:08 crc kubenswrapper[4885]: I1002 02:17:08.038680 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-f8tvb" Oct 02 02:17:08 crc kubenswrapper[4885]: I1002 02:17:08.041638 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 02:17:08 crc kubenswrapper[4885]: I1002 02:17:08.073840 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/93d5499a-81ec-4dbc-b769-94d3e8690fbe-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-wvh9j\" (UID: \"93d5499a-81ec-4dbc-b769-94d3e8690fbe\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wvh9j" Oct 02 02:17:08 crc kubenswrapper[4885]: I1002 02:17:08.073901 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/93d5499a-81ec-4dbc-b769-94d3e8690fbe-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-wvh9j\" (UID: \"93d5499a-81ec-4dbc-b769-94d3e8690fbe\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wvh9j" Oct 02 02:17:08 crc kubenswrapper[4885]: I1002 02:17:08.074113 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5knrx\" (UniqueName: \"kubernetes.io/projected/93d5499a-81ec-4dbc-b769-94d3e8690fbe-kube-api-access-5knrx\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-wvh9j\" (UID: \"93d5499a-81ec-4dbc-b769-94d3e8690fbe\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wvh9j" Oct 02 02:17:08 crc kubenswrapper[4885]: I1002 02:17:08.079350 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wvh9j"] Oct 02 02:17:08 crc kubenswrapper[4885]: I1002 02:17:08.176553 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5knrx\" (UniqueName: \"kubernetes.io/projected/93d5499a-81ec-4dbc-b769-94d3e8690fbe-kube-api-access-5knrx\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-wvh9j\" (UID: \"93d5499a-81ec-4dbc-b769-94d3e8690fbe\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wvh9j" Oct 02 02:17:08 crc kubenswrapper[4885]: I1002 02:17:08.176874 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/93d5499a-81ec-4dbc-b769-94d3e8690fbe-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-wvh9j\" (UID: \"93d5499a-81ec-4dbc-b769-94d3e8690fbe\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wvh9j" Oct 02 02:17:08 crc kubenswrapper[4885]: I1002 02:17:08.176956 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/93d5499a-81ec-4dbc-b769-94d3e8690fbe-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-wvh9j\" (UID: \"93d5499a-81ec-4dbc-b769-94d3e8690fbe\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wvh9j" Oct 02 02:17:08 crc kubenswrapper[4885]: I1002 02:17:08.186955 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/93d5499a-81ec-4dbc-b769-94d3e8690fbe-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-wvh9j\" (UID: \"93d5499a-81ec-4dbc-b769-94d3e8690fbe\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wvh9j" Oct 02 02:17:08 crc kubenswrapper[4885]: I1002 02:17:08.187974 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/93d5499a-81ec-4dbc-b769-94d3e8690fbe-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-wvh9j\" (UID: \"93d5499a-81ec-4dbc-b769-94d3e8690fbe\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wvh9j" Oct 02 02:17:08 crc kubenswrapper[4885]: I1002 02:17:08.206185 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5knrx\" (UniqueName: \"kubernetes.io/projected/93d5499a-81ec-4dbc-b769-94d3e8690fbe-kube-api-access-5knrx\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-wvh9j\" (UID: \"93d5499a-81ec-4dbc-b769-94d3e8690fbe\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wvh9j" Oct 02 02:17:08 crc kubenswrapper[4885]: I1002 02:17:08.353880 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wvh9j" Oct 02 02:17:08 crc kubenswrapper[4885]: I1002 02:17:08.976628 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wvh9j"] Oct 02 02:17:08 crc kubenswrapper[4885]: I1002 02:17:08.984595 4885 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 02:17:09 crc kubenswrapper[4885]: I1002 02:17:09.146894 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wvh9j" event={"ID":"93d5499a-81ec-4dbc-b769-94d3e8690fbe","Type":"ContainerStarted","Data":"c06603b6948b1b6e7711870973ac027406ee2e9c253a41c35980c96205168a12"} Oct 02 02:17:10 crc kubenswrapper[4885]: I1002 02:17:10.161368 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wvh9j" event={"ID":"93d5499a-81ec-4dbc-b769-94d3e8690fbe","Type":"ContainerStarted","Data":"5405c9c0f4574a393e5bfcb954ecf8d7262f26f5a0d7fb9e58d6543921e64420"} Oct 02 02:17:10 crc kubenswrapper[4885]: I1002 02:17:10.194695 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wvh9j" podStartSLOduration=1.526149281 podStartE2EDuration="2.194673498s" podCreationTimestamp="2025-10-02 02:17:08 +0000 UTC" firstStartedPulling="2025-10-02 02:17:08.984278043 +0000 UTC m=+1817.796025452" lastFinishedPulling="2025-10-02 02:17:09.65280223 +0000 UTC m=+1818.464549669" observedRunningTime="2025-10-02 02:17:10.181822325 +0000 UTC m=+1818.993569754" watchObservedRunningTime="2025-10-02 02:17:10.194673498 +0000 UTC m=+1819.006420937" Oct 02 02:17:13 crc kubenswrapper[4885]: I1002 02:17:13.767963 4885 scope.go:117] "RemoveContainer" containerID="4c8b11178b78970976d78c639d18fb811b383b7789114f921f64a5148460950c" Oct 02 02:17:13 crc kubenswrapper[4885]: I1002 02:17:13.865491 4885 scope.go:117] "RemoveContainer" containerID="afb02741a030d7ca36e1fee4413fde2d4f695ccf7f80c7624b2fd6e3270d7a1d" Oct 02 02:17:13 crc kubenswrapper[4885]: I1002 02:17:13.937234 4885 scope.go:117] "RemoveContainer" containerID="af23b39faeaa1d212bba186d27a3acdf681f31ba75d794163d93887c5df66788" Oct 02 02:17:18 crc kubenswrapper[4885]: I1002 02:17:18.047925 4885 scope.go:117] "RemoveContainer" containerID="be0fc35a7f63d87317c3e5145a3df69700656f283328299a3b8545263cc22494" Oct 02 02:17:18 crc kubenswrapper[4885]: E1002 02:17:18.049243 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:17:23 crc kubenswrapper[4885]: I1002 02:17:23.044245 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-fsn9x"] Oct 02 02:17:23 crc kubenswrapper[4885]: I1002 02:17:23.056046 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-fsn9x"] Oct 02 02:17:24 crc kubenswrapper[4885]: I1002 02:17:24.077056 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a91d000d-543e-42b4-b815-12243df15ea9" path="/var/lib/kubelet/pods/a91d000d-543e-42b4-b815-12243df15ea9/volumes" Oct 02 02:17:33 crc kubenswrapper[4885]: I1002 02:17:33.046720 4885 scope.go:117] "RemoveContainer" containerID="be0fc35a7f63d87317c3e5145a3df69700656f283328299a3b8545263cc22494" Oct 02 02:17:33 crc kubenswrapper[4885]: E1002 02:17:33.047752 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:17:46 crc kubenswrapper[4885]: I1002 02:17:46.047117 4885 scope.go:117] "RemoveContainer" containerID="be0fc35a7f63d87317c3e5145a3df69700656f283328299a3b8545263cc22494" Oct 02 02:17:46 crc kubenswrapper[4885]: I1002 02:17:46.593765 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" event={"ID":"d169c8ba-000a-4218-8e3b-4ae53035b110","Type":"ContainerStarted","Data":"562009e16fbeefe72a46fc9d563e2da31ec7f209f7fd3da12e264d10773a1c88"} Oct 02 02:17:57 crc kubenswrapper[4885]: I1002 02:17:57.748241 4885 generic.go:334] "Generic (PLEG): container finished" podID="93d5499a-81ec-4dbc-b769-94d3e8690fbe" containerID="5405c9c0f4574a393e5bfcb954ecf8d7262f26f5a0d7fb9e58d6543921e64420" exitCode=0 Oct 02 02:17:57 crc kubenswrapper[4885]: I1002 02:17:57.748321 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wvh9j" event={"ID":"93d5499a-81ec-4dbc-b769-94d3e8690fbe","Type":"ContainerDied","Data":"5405c9c0f4574a393e5bfcb954ecf8d7262f26f5a0d7fb9e58d6543921e64420"} Oct 02 02:17:59 crc kubenswrapper[4885]: I1002 02:17:59.251670 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wvh9j" Oct 02 02:17:59 crc kubenswrapper[4885]: I1002 02:17:59.334525 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5knrx\" (UniqueName: \"kubernetes.io/projected/93d5499a-81ec-4dbc-b769-94d3e8690fbe-kube-api-access-5knrx\") pod \"93d5499a-81ec-4dbc-b769-94d3e8690fbe\" (UID: \"93d5499a-81ec-4dbc-b769-94d3e8690fbe\") " Oct 02 02:17:59 crc kubenswrapper[4885]: I1002 02:17:59.335407 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/93d5499a-81ec-4dbc-b769-94d3e8690fbe-inventory\") pod \"93d5499a-81ec-4dbc-b769-94d3e8690fbe\" (UID: \"93d5499a-81ec-4dbc-b769-94d3e8690fbe\") " Oct 02 02:17:59 crc kubenswrapper[4885]: I1002 02:17:59.335542 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/93d5499a-81ec-4dbc-b769-94d3e8690fbe-ssh-key\") pod \"93d5499a-81ec-4dbc-b769-94d3e8690fbe\" (UID: \"93d5499a-81ec-4dbc-b769-94d3e8690fbe\") " Oct 02 02:17:59 crc kubenswrapper[4885]: I1002 02:17:59.340362 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93d5499a-81ec-4dbc-b769-94d3e8690fbe-kube-api-access-5knrx" (OuterVolumeSpecName: "kube-api-access-5knrx") pod "93d5499a-81ec-4dbc-b769-94d3e8690fbe" (UID: "93d5499a-81ec-4dbc-b769-94d3e8690fbe"). InnerVolumeSpecName "kube-api-access-5knrx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:17:59 crc kubenswrapper[4885]: I1002 02:17:59.367639 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93d5499a-81ec-4dbc-b769-94d3e8690fbe-inventory" (OuterVolumeSpecName: "inventory") pod "93d5499a-81ec-4dbc-b769-94d3e8690fbe" (UID: "93d5499a-81ec-4dbc-b769-94d3e8690fbe"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:17:59 crc kubenswrapper[4885]: I1002 02:17:59.373220 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93d5499a-81ec-4dbc-b769-94d3e8690fbe-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "93d5499a-81ec-4dbc-b769-94d3e8690fbe" (UID: "93d5499a-81ec-4dbc-b769-94d3e8690fbe"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:17:59 crc kubenswrapper[4885]: I1002 02:17:59.438452 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5knrx\" (UniqueName: \"kubernetes.io/projected/93d5499a-81ec-4dbc-b769-94d3e8690fbe-kube-api-access-5knrx\") on node \"crc\" DevicePath \"\"" Oct 02 02:17:59 crc kubenswrapper[4885]: I1002 02:17:59.438507 4885 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/93d5499a-81ec-4dbc-b769-94d3e8690fbe-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 02:17:59 crc kubenswrapper[4885]: I1002 02:17:59.438526 4885 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/93d5499a-81ec-4dbc-b769-94d3e8690fbe-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 02:17:59 crc kubenswrapper[4885]: I1002 02:17:59.771970 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wvh9j" event={"ID":"93d5499a-81ec-4dbc-b769-94d3e8690fbe","Type":"ContainerDied","Data":"c06603b6948b1b6e7711870973ac027406ee2e9c253a41c35980c96205168a12"} Oct 02 02:17:59 crc kubenswrapper[4885]: I1002 02:17:59.772033 4885 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c06603b6948b1b6e7711870973ac027406ee2e9c253a41c35980c96205168a12" Oct 02 02:17:59 crc kubenswrapper[4885]: I1002 02:17:59.772583 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wvh9j" Oct 02 02:17:59 crc kubenswrapper[4885]: I1002 02:17:59.869480 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-njjfc"] Oct 02 02:17:59 crc kubenswrapper[4885]: E1002 02:17:59.869946 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93d5499a-81ec-4dbc-b769-94d3e8690fbe" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 02 02:17:59 crc kubenswrapper[4885]: I1002 02:17:59.869968 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="93d5499a-81ec-4dbc-b769-94d3e8690fbe" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 02 02:17:59 crc kubenswrapper[4885]: I1002 02:17:59.870205 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="93d5499a-81ec-4dbc-b769-94d3e8690fbe" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 02 02:17:59 crc kubenswrapper[4885]: I1002 02:17:59.870964 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-njjfc" Oct 02 02:17:59 crc kubenswrapper[4885]: I1002 02:17:59.875070 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 02:17:59 crc kubenswrapper[4885]: I1002 02:17:59.875143 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 02:17:59 crc kubenswrapper[4885]: I1002 02:17:59.875160 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-f8tvb" Oct 02 02:17:59 crc kubenswrapper[4885]: I1002 02:17:59.875305 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 02:17:59 crc kubenswrapper[4885]: I1002 02:17:59.884990 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-njjfc"] Oct 02 02:17:59 crc kubenswrapper[4885]: I1002 02:17:59.947625 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/e6dc3bce-d7cf-4ce8-9577-dd1f6813bff1-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-njjfc\" (UID: \"e6dc3bce-d7cf-4ce8-9577-dd1f6813bff1\") " pod="openstack/ssh-known-hosts-edpm-deployment-njjfc" Oct 02 02:17:59 crc kubenswrapper[4885]: I1002 02:17:59.947671 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e6dc3bce-d7cf-4ce8-9577-dd1f6813bff1-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-njjfc\" (UID: \"e6dc3bce-d7cf-4ce8-9577-dd1f6813bff1\") " pod="openstack/ssh-known-hosts-edpm-deployment-njjfc" Oct 02 02:17:59 crc kubenswrapper[4885]: I1002 02:17:59.947790 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-28fbc\" (UniqueName: \"kubernetes.io/projected/e6dc3bce-d7cf-4ce8-9577-dd1f6813bff1-kube-api-access-28fbc\") pod \"ssh-known-hosts-edpm-deployment-njjfc\" (UID: \"e6dc3bce-d7cf-4ce8-9577-dd1f6813bff1\") " pod="openstack/ssh-known-hosts-edpm-deployment-njjfc" Oct 02 02:18:00 crc kubenswrapper[4885]: I1002 02:18:00.049964 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-28fbc\" (UniqueName: \"kubernetes.io/projected/e6dc3bce-d7cf-4ce8-9577-dd1f6813bff1-kube-api-access-28fbc\") pod \"ssh-known-hosts-edpm-deployment-njjfc\" (UID: \"e6dc3bce-d7cf-4ce8-9577-dd1f6813bff1\") " pod="openstack/ssh-known-hosts-edpm-deployment-njjfc" Oct 02 02:18:00 crc kubenswrapper[4885]: I1002 02:18:00.050142 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/e6dc3bce-d7cf-4ce8-9577-dd1f6813bff1-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-njjfc\" (UID: \"e6dc3bce-d7cf-4ce8-9577-dd1f6813bff1\") " pod="openstack/ssh-known-hosts-edpm-deployment-njjfc" Oct 02 02:18:00 crc kubenswrapper[4885]: I1002 02:18:00.050186 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e6dc3bce-d7cf-4ce8-9577-dd1f6813bff1-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-njjfc\" (UID: \"e6dc3bce-d7cf-4ce8-9577-dd1f6813bff1\") " pod="openstack/ssh-known-hosts-edpm-deployment-njjfc" Oct 02 02:18:00 crc kubenswrapper[4885]: I1002 02:18:00.056356 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/e6dc3bce-d7cf-4ce8-9577-dd1f6813bff1-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-njjfc\" (UID: \"e6dc3bce-d7cf-4ce8-9577-dd1f6813bff1\") " pod="openstack/ssh-known-hosts-edpm-deployment-njjfc" Oct 02 02:18:00 crc kubenswrapper[4885]: I1002 02:18:00.058079 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e6dc3bce-d7cf-4ce8-9577-dd1f6813bff1-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-njjfc\" (UID: \"e6dc3bce-d7cf-4ce8-9577-dd1f6813bff1\") " pod="openstack/ssh-known-hosts-edpm-deployment-njjfc" Oct 02 02:18:00 crc kubenswrapper[4885]: I1002 02:18:00.080444 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-28fbc\" (UniqueName: \"kubernetes.io/projected/e6dc3bce-d7cf-4ce8-9577-dd1f6813bff1-kube-api-access-28fbc\") pod \"ssh-known-hosts-edpm-deployment-njjfc\" (UID: \"e6dc3bce-d7cf-4ce8-9577-dd1f6813bff1\") " pod="openstack/ssh-known-hosts-edpm-deployment-njjfc" Oct 02 02:18:00 crc kubenswrapper[4885]: I1002 02:18:00.195934 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-njjfc" Oct 02 02:18:00 crc kubenswrapper[4885]: I1002 02:18:00.802122 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-njjfc"] Oct 02 02:18:01 crc kubenswrapper[4885]: I1002 02:18:01.816341 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-njjfc" event={"ID":"e6dc3bce-d7cf-4ce8-9577-dd1f6813bff1","Type":"ContainerStarted","Data":"ba88c79207dc752a4c616c56fdeee570943f125daee2b326eb8c02a48fffcfb6"} Oct 02 02:18:01 crc kubenswrapper[4885]: I1002 02:18:01.816699 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-njjfc" event={"ID":"e6dc3bce-d7cf-4ce8-9577-dd1f6813bff1","Type":"ContainerStarted","Data":"61506a94c6693aae5eb32c8863da68e991ff4038a988ad6f4a8772c2a630a81f"} Oct 02 02:18:01 crc kubenswrapper[4885]: I1002 02:18:01.848087 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-njjfc" podStartSLOduration=2.3598545619999998 podStartE2EDuration="2.84806459s" podCreationTimestamp="2025-10-02 02:17:59 +0000 UTC" firstStartedPulling="2025-10-02 02:18:00.800101018 +0000 UTC m=+1869.611848427" lastFinishedPulling="2025-10-02 02:18:01.288311016 +0000 UTC m=+1870.100058455" observedRunningTime="2025-10-02 02:18:01.835461375 +0000 UTC m=+1870.647208804" watchObservedRunningTime="2025-10-02 02:18:01.84806459 +0000 UTC m=+1870.659812009" Oct 02 02:18:08 crc kubenswrapper[4885]: I1002 02:18:08.903033 4885 generic.go:334] "Generic (PLEG): container finished" podID="e6dc3bce-d7cf-4ce8-9577-dd1f6813bff1" containerID="ba88c79207dc752a4c616c56fdeee570943f125daee2b326eb8c02a48fffcfb6" exitCode=0 Oct 02 02:18:08 crc kubenswrapper[4885]: I1002 02:18:08.903178 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-njjfc" event={"ID":"e6dc3bce-d7cf-4ce8-9577-dd1f6813bff1","Type":"ContainerDied","Data":"ba88c79207dc752a4c616c56fdeee570943f125daee2b326eb8c02a48fffcfb6"} Oct 02 02:18:10 crc kubenswrapper[4885]: I1002 02:18:10.473090 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-njjfc" Oct 02 02:18:10 crc kubenswrapper[4885]: I1002 02:18:10.584207 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-28fbc\" (UniqueName: \"kubernetes.io/projected/e6dc3bce-d7cf-4ce8-9577-dd1f6813bff1-kube-api-access-28fbc\") pod \"e6dc3bce-d7cf-4ce8-9577-dd1f6813bff1\" (UID: \"e6dc3bce-d7cf-4ce8-9577-dd1f6813bff1\") " Oct 02 02:18:10 crc kubenswrapper[4885]: I1002 02:18:10.584484 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/e6dc3bce-d7cf-4ce8-9577-dd1f6813bff1-inventory-0\") pod \"e6dc3bce-d7cf-4ce8-9577-dd1f6813bff1\" (UID: \"e6dc3bce-d7cf-4ce8-9577-dd1f6813bff1\") " Oct 02 02:18:10 crc kubenswrapper[4885]: I1002 02:18:10.584619 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e6dc3bce-d7cf-4ce8-9577-dd1f6813bff1-ssh-key-openstack-edpm-ipam\") pod \"e6dc3bce-d7cf-4ce8-9577-dd1f6813bff1\" (UID: \"e6dc3bce-d7cf-4ce8-9577-dd1f6813bff1\") " Oct 02 02:18:10 crc kubenswrapper[4885]: I1002 02:18:10.589991 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e6dc3bce-d7cf-4ce8-9577-dd1f6813bff1-kube-api-access-28fbc" (OuterVolumeSpecName: "kube-api-access-28fbc") pod "e6dc3bce-d7cf-4ce8-9577-dd1f6813bff1" (UID: "e6dc3bce-d7cf-4ce8-9577-dd1f6813bff1"). InnerVolumeSpecName "kube-api-access-28fbc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:18:10 crc kubenswrapper[4885]: I1002 02:18:10.622044 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e6dc3bce-d7cf-4ce8-9577-dd1f6813bff1-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "e6dc3bce-d7cf-4ce8-9577-dd1f6813bff1" (UID: "e6dc3bce-d7cf-4ce8-9577-dd1f6813bff1"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:18:10 crc kubenswrapper[4885]: I1002 02:18:10.624241 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e6dc3bce-d7cf-4ce8-9577-dd1f6813bff1-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "e6dc3bce-d7cf-4ce8-9577-dd1f6813bff1" (UID: "e6dc3bce-d7cf-4ce8-9577-dd1f6813bff1"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:18:10 crc kubenswrapper[4885]: I1002 02:18:10.688823 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-28fbc\" (UniqueName: \"kubernetes.io/projected/e6dc3bce-d7cf-4ce8-9577-dd1f6813bff1-kube-api-access-28fbc\") on node \"crc\" DevicePath \"\"" Oct 02 02:18:10 crc kubenswrapper[4885]: I1002 02:18:10.688913 4885 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/e6dc3bce-d7cf-4ce8-9577-dd1f6813bff1-inventory-0\") on node \"crc\" DevicePath \"\"" Oct 02 02:18:10 crc kubenswrapper[4885]: I1002 02:18:10.688931 4885 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e6dc3bce-d7cf-4ce8-9577-dd1f6813bff1-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Oct 02 02:18:10 crc kubenswrapper[4885]: I1002 02:18:10.977574 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-njjfc" event={"ID":"e6dc3bce-d7cf-4ce8-9577-dd1f6813bff1","Type":"ContainerDied","Data":"61506a94c6693aae5eb32c8863da68e991ff4038a988ad6f4a8772c2a630a81f"} Oct 02 02:18:10 crc kubenswrapper[4885]: I1002 02:18:10.977891 4885 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="61506a94c6693aae5eb32c8863da68e991ff4038a988ad6f4a8772c2a630a81f" Oct 02 02:18:10 crc kubenswrapper[4885]: I1002 02:18:10.977699 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-njjfc" Oct 02 02:18:11 crc kubenswrapper[4885]: I1002 02:18:11.040200 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-jnpmf"] Oct 02 02:18:11 crc kubenswrapper[4885]: E1002 02:18:11.040654 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6dc3bce-d7cf-4ce8-9577-dd1f6813bff1" containerName="ssh-known-hosts-edpm-deployment" Oct 02 02:18:11 crc kubenswrapper[4885]: I1002 02:18:11.040670 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6dc3bce-d7cf-4ce8-9577-dd1f6813bff1" containerName="ssh-known-hosts-edpm-deployment" Oct 02 02:18:11 crc kubenswrapper[4885]: I1002 02:18:11.040927 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="e6dc3bce-d7cf-4ce8-9577-dd1f6813bff1" containerName="ssh-known-hosts-edpm-deployment" Oct 02 02:18:11 crc kubenswrapper[4885]: I1002 02:18:11.041772 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jnpmf" Oct 02 02:18:11 crc kubenswrapper[4885]: I1002 02:18:11.046586 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 02:18:11 crc kubenswrapper[4885]: I1002 02:18:11.046794 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 02:18:11 crc kubenswrapper[4885]: I1002 02:18:11.047049 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 02:18:11 crc kubenswrapper[4885]: I1002 02:18:11.047310 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-f8tvb" Oct 02 02:18:11 crc kubenswrapper[4885]: I1002 02:18:11.061377 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-jnpmf"] Oct 02 02:18:11 crc kubenswrapper[4885]: I1002 02:18:11.096768 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8mkx6\" (UniqueName: \"kubernetes.io/projected/3093cd67-4ea2-4256-bcf4-7e2533870ba6-kube-api-access-8mkx6\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-jnpmf\" (UID: \"3093cd67-4ea2-4256-bcf4-7e2533870ba6\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jnpmf" Oct 02 02:18:11 crc kubenswrapper[4885]: I1002 02:18:11.096957 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3093cd67-4ea2-4256-bcf4-7e2533870ba6-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-jnpmf\" (UID: \"3093cd67-4ea2-4256-bcf4-7e2533870ba6\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jnpmf" Oct 02 02:18:11 crc kubenswrapper[4885]: I1002 02:18:11.097160 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3093cd67-4ea2-4256-bcf4-7e2533870ba6-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-jnpmf\" (UID: \"3093cd67-4ea2-4256-bcf4-7e2533870ba6\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jnpmf" Oct 02 02:18:11 crc kubenswrapper[4885]: I1002 02:18:11.198908 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8mkx6\" (UniqueName: \"kubernetes.io/projected/3093cd67-4ea2-4256-bcf4-7e2533870ba6-kube-api-access-8mkx6\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-jnpmf\" (UID: \"3093cd67-4ea2-4256-bcf4-7e2533870ba6\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jnpmf" Oct 02 02:18:11 crc kubenswrapper[4885]: I1002 02:18:11.199025 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3093cd67-4ea2-4256-bcf4-7e2533870ba6-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-jnpmf\" (UID: \"3093cd67-4ea2-4256-bcf4-7e2533870ba6\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jnpmf" Oct 02 02:18:11 crc kubenswrapper[4885]: I1002 02:18:11.199161 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3093cd67-4ea2-4256-bcf4-7e2533870ba6-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-jnpmf\" (UID: \"3093cd67-4ea2-4256-bcf4-7e2533870ba6\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jnpmf" Oct 02 02:18:11 crc kubenswrapper[4885]: I1002 02:18:11.203089 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3093cd67-4ea2-4256-bcf4-7e2533870ba6-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-jnpmf\" (UID: \"3093cd67-4ea2-4256-bcf4-7e2533870ba6\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jnpmf" Oct 02 02:18:11 crc kubenswrapper[4885]: I1002 02:18:11.203302 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3093cd67-4ea2-4256-bcf4-7e2533870ba6-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-jnpmf\" (UID: \"3093cd67-4ea2-4256-bcf4-7e2533870ba6\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jnpmf" Oct 02 02:18:11 crc kubenswrapper[4885]: I1002 02:18:11.215073 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8mkx6\" (UniqueName: \"kubernetes.io/projected/3093cd67-4ea2-4256-bcf4-7e2533870ba6-kube-api-access-8mkx6\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-jnpmf\" (UID: \"3093cd67-4ea2-4256-bcf4-7e2533870ba6\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jnpmf" Oct 02 02:18:11 crc kubenswrapper[4885]: I1002 02:18:11.367863 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jnpmf" Oct 02 02:18:11 crc kubenswrapper[4885]: I1002 02:18:11.718943 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-jnpmf"] Oct 02 02:18:11 crc kubenswrapper[4885]: I1002 02:18:11.993159 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jnpmf" event={"ID":"3093cd67-4ea2-4256-bcf4-7e2533870ba6","Type":"ContainerStarted","Data":"e5c0ce0915f170a120f902750b87b52f11f6a38595fc9b9eb0c5e49ea21e95c3"} Oct 02 02:18:13 crc kubenswrapper[4885]: I1002 02:18:13.004838 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jnpmf" event={"ID":"3093cd67-4ea2-4256-bcf4-7e2533870ba6","Type":"ContainerStarted","Data":"8c856645b4c1b35cda2353b2458917621ade03436b25a2cd26b0e1cfc82122ff"} Oct 02 02:18:13 crc kubenswrapper[4885]: I1002 02:18:13.033123 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jnpmf" podStartSLOduration=1.529601692 podStartE2EDuration="2.033095288s" podCreationTimestamp="2025-10-02 02:18:11 +0000 UTC" firstStartedPulling="2025-10-02 02:18:11.728429343 +0000 UTC m=+1880.540176742" lastFinishedPulling="2025-10-02 02:18:12.231922939 +0000 UTC m=+1881.043670338" observedRunningTime="2025-10-02 02:18:13.028073615 +0000 UTC m=+1881.839821014" watchObservedRunningTime="2025-10-02 02:18:13.033095288 +0000 UTC m=+1881.844842717" Oct 02 02:18:14 crc kubenswrapper[4885]: I1002 02:18:14.086187 4885 scope.go:117] "RemoveContainer" containerID="fa1c8dfbc36fd1d1ad81f83d4fdff3abd1018ec697ea27a6ce4e03a4ed5a1ff7" Oct 02 02:18:21 crc kubenswrapper[4885]: I1002 02:18:21.091820 4885 generic.go:334] "Generic (PLEG): container finished" podID="3093cd67-4ea2-4256-bcf4-7e2533870ba6" containerID="8c856645b4c1b35cda2353b2458917621ade03436b25a2cd26b0e1cfc82122ff" exitCode=0 Oct 02 02:18:21 crc kubenswrapper[4885]: I1002 02:18:21.091887 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jnpmf" event={"ID":"3093cd67-4ea2-4256-bcf4-7e2533870ba6","Type":"ContainerDied","Data":"8c856645b4c1b35cda2353b2458917621ade03436b25a2cd26b0e1cfc82122ff"} Oct 02 02:18:22 crc kubenswrapper[4885]: I1002 02:18:22.604217 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jnpmf" Oct 02 02:18:22 crc kubenswrapper[4885]: I1002 02:18:22.641583 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3093cd67-4ea2-4256-bcf4-7e2533870ba6-inventory\") pod \"3093cd67-4ea2-4256-bcf4-7e2533870ba6\" (UID: \"3093cd67-4ea2-4256-bcf4-7e2533870ba6\") " Oct 02 02:18:22 crc kubenswrapper[4885]: I1002 02:18:22.641650 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3093cd67-4ea2-4256-bcf4-7e2533870ba6-ssh-key\") pod \"3093cd67-4ea2-4256-bcf4-7e2533870ba6\" (UID: \"3093cd67-4ea2-4256-bcf4-7e2533870ba6\") " Oct 02 02:18:22 crc kubenswrapper[4885]: I1002 02:18:22.673871 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3093cd67-4ea2-4256-bcf4-7e2533870ba6-inventory" (OuterVolumeSpecName: "inventory") pod "3093cd67-4ea2-4256-bcf4-7e2533870ba6" (UID: "3093cd67-4ea2-4256-bcf4-7e2533870ba6"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:18:22 crc kubenswrapper[4885]: I1002 02:18:22.686866 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3093cd67-4ea2-4256-bcf4-7e2533870ba6-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "3093cd67-4ea2-4256-bcf4-7e2533870ba6" (UID: "3093cd67-4ea2-4256-bcf4-7e2533870ba6"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:18:22 crc kubenswrapper[4885]: I1002 02:18:22.743717 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8mkx6\" (UniqueName: \"kubernetes.io/projected/3093cd67-4ea2-4256-bcf4-7e2533870ba6-kube-api-access-8mkx6\") pod \"3093cd67-4ea2-4256-bcf4-7e2533870ba6\" (UID: \"3093cd67-4ea2-4256-bcf4-7e2533870ba6\") " Oct 02 02:18:22 crc kubenswrapper[4885]: I1002 02:18:22.744844 4885 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3093cd67-4ea2-4256-bcf4-7e2533870ba6-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 02:18:22 crc kubenswrapper[4885]: I1002 02:18:22.744873 4885 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3093cd67-4ea2-4256-bcf4-7e2533870ba6-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 02:18:22 crc kubenswrapper[4885]: I1002 02:18:22.747769 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3093cd67-4ea2-4256-bcf4-7e2533870ba6-kube-api-access-8mkx6" (OuterVolumeSpecName: "kube-api-access-8mkx6") pod "3093cd67-4ea2-4256-bcf4-7e2533870ba6" (UID: "3093cd67-4ea2-4256-bcf4-7e2533870ba6"). InnerVolumeSpecName "kube-api-access-8mkx6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:18:22 crc kubenswrapper[4885]: I1002 02:18:22.846136 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8mkx6\" (UniqueName: \"kubernetes.io/projected/3093cd67-4ea2-4256-bcf4-7e2533870ba6-kube-api-access-8mkx6\") on node \"crc\" DevicePath \"\"" Oct 02 02:18:23 crc kubenswrapper[4885]: I1002 02:18:23.134948 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jnpmf" event={"ID":"3093cd67-4ea2-4256-bcf4-7e2533870ba6","Type":"ContainerDied","Data":"e5c0ce0915f170a120f902750b87b52f11f6a38595fc9b9eb0c5e49ea21e95c3"} Oct 02 02:18:23 crc kubenswrapper[4885]: I1002 02:18:23.135002 4885 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e5c0ce0915f170a120f902750b87b52f11f6a38595fc9b9eb0c5e49ea21e95c3" Oct 02 02:18:23 crc kubenswrapper[4885]: I1002 02:18:23.135009 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jnpmf" Oct 02 02:18:23 crc kubenswrapper[4885]: I1002 02:18:23.210209 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2rt6h"] Oct 02 02:18:23 crc kubenswrapper[4885]: E1002 02:18:23.211148 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3093cd67-4ea2-4256-bcf4-7e2533870ba6" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 02 02:18:23 crc kubenswrapper[4885]: I1002 02:18:23.211174 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="3093cd67-4ea2-4256-bcf4-7e2533870ba6" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 02 02:18:23 crc kubenswrapper[4885]: I1002 02:18:23.211467 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="3093cd67-4ea2-4256-bcf4-7e2533870ba6" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 02 02:18:23 crc kubenswrapper[4885]: I1002 02:18:23.212333 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2rt6h" Oct 02 02:18:23 crc kubenswrapper[4885]: I1002 02:18:23.214754 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-f8tvb" Oct 02 02:18:23 crc kubenswrapper[4885]: I1002 02:18:23.214967 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 02:18:23 crc kubenswrapper[4885]: I1002 02:18:23.215096 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 02:18:23 crc kubenswrapper[4885]: I1002 02:18:23.215538 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 02:18:23 crc kubenswrapper[4885]: I1002 02:18:23.223157 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2rt6h"] Oct 02 02:18:23 crc kubenswrapper[4885]: I1002 02:18:23.252451 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cnxzp\" (UniqueName: \"kubernetes.io/projected/2d16884c-61ae-4288-b629-2a7c3be23ea8-kube-api-access-cnxzp\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-2rt6h\" (UID: \"2d16884c-61ae-4288-b629-2a7c3be23ea8\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2rt6h" Oct 02 02:18:23 crc kubenswrapper[4885]: I1002 02:18:23.252858 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2d16884c-61ae-4288-b629-2a7c3be23ea8-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-2rt6h\" (UID: \"2d16884c-61ae-4288-b629-2a7c3be23ea8\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2rt6h" Oct 02 02:18:23 crc kubenswrapper[4885]: I1002 02:18:23.253018 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2d16884c-61ae-4288-b629-2a7c3be23ea8-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-2rt6h\" (UID: \"2d16884c-61ae-4288-b629-2a7c3be23ea8\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2rt6h" Oct 02 02:18:23 crc kubenswrapper[4885]: I1002 02:18:23.354126 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2d16884c-61ae-4288-b629-2a7c3be23ea8-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-2rt6h\" (UID: \"2d16884c-61ae-4288-b629-2a7c3be23ea8\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2rt6h" Oct 02 02:18:23 crc kubenswrapper[4885]: I1002 02:18:23.354206 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cnxzp\" (UniqueName: \"kubernetes.io/projected/2d16884c-61ae-4288-b629-2a7c3be23ea8-kube-api-access-cnxzp\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-2rt6h\" (UID: \"2d16884c-61ae-4288-b629-2a7c3be23ea8\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2rt6h" Oct 02 02:18:23 crc kubenswrapper[4885]: I1002 02:18:23.354345 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2d16884c-61ae-4288-b629-2a7c3be23ea8-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-2rt6h\" (UID: \"2d16884c-61ae-4288-b629-2a7c3be23ea8\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2rt6h" Oct 02 02:18:23 crc kubenswrapper[4885]: I1002 02:18:23.359593 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2d16884c-61ae-4288-b629-2a7c3be23ea8-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-2rt6h\" (UID: \"2d16884c-61ae-4288-b629-2a7c3be23ea8\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2rt6h" Oct 02 02:18:23 crc kubenswrapper[4885]: I1002 02:18:23.359648 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2d16884c-61ae-4288-b629-2a7c3be23ea8-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-2rt6h\" (UID: \"2d16884c-61ae-4288-b629-2a7c3be23ea8\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2rt6h" Oct 02 02:18:23 crc kubenswrapper[4885]: I1002 02:18:23.372766 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cnxzp\" (UniqueName: \"kubernetes.io/projected/2d16884c-61ae-4288-b629-2a7c3be23ea8-kube-api-access-cnxzp\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-2rt6h\" (UID: \"2d16884c-61ae-4288-b629-2a7c3be23ea8\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2rt6h" Oct 02 02:18:23 crc kubenswrapper[4885]: I1002 02:18:23.536146 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2rt6h" Oct 02 02:18:24 crc kubenswrapper[4885]: I1002 02:18:24.134922 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2rt6h"] Oct 02 02:18:24 crc kubenswrapper[4885]: I1002 02:18:24.151242 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2rt6h" event={"ID":"2d16884c-61ae-4288-b629-2a7c3be23ea8","Type":"ContainerStarted","Data":"8c5ff134d99b595ca94dbaa70f2401eef3621b96207f73a907782d3b6ca1c875"} Oct 02 02:18:25 crc kubenswrapper[4885]: I1002 02:18:25.161786 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2rt6h" event={"ID":"2d16884c-61ae-4288-b629-2a7c3be23ea8","Type":"ContainerStarted","Data":"a02801309dd9e67ce9ecc023b16a25b64345d9275628cea1f249a4af75b21fc3"} Oct 02 02:18:25 crc kubenswrapper[4885]: I1002 02:18:25.185300 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2rt6h" podStartSLOduration=1.780568899 podStartE2EDuration="2.185284501s" podCreationTimestamp="2025-10-02 02:18:23 +0000 UTC" firstStartedPulling="2025-10-02 02:18:24.135790603 +0000 UTC m=+1892.947538042" lastFinishedPulling="2025-10-02 02:18:24.540506205 +0000 UTC m=+1893.352253644" observedRunningTime="2025-10-02 02:18:25.183533907 +0000 UTC m=+1893.995281296" watchObservedRunningTime="2025-10-02 02:18:25.185284501 +0000 UTC m=+1893.997031910" Oct 02 02:18:34 crc kubenswrapper[4885]: I1002 02:18:34.283393 4885 generic.go:334] "Generic (PLEG): container finished" podID="2d16884c-61ae-4288-b629-2a7c3be23ea8" containerID="a02801309dd9e67ce9ecc023b16a25b64345d9275628cea1f249a4af75b21fc3" exitCode=0 Oct 02 02:18:34 crc kubenswrapper[4885]: I1002 02:18:34.283610 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2rt6h" event={"ID":"2d16884c-61ae-4288-b629-2a7c3be23ea8","Type":"ContainerDied","Data":"a02801309dd9e67ce9ecc023b16a25b64345d9275628cea1f249a4af75b21fc3"} Oct 02 02:18:35 crc kubenswrapper[4885]: I1002 02:18:35.815099 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2rt6h" Oct 02 02:18:35 crc kubenswrapper[4885]: I1002 02:18:35.953244 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2d16884c-61ae-4288-b629-2a7c3be23ea8-inventory\") pod \"2d16884c-61ae-4288-b629-2a7c3be23ea8\" (UID: \"2d16884c-61ae-4288-b629-2a7c3be23ea8\") " Oct 02 02:18:35 crc kubenswrapper[4885]: I1002 02:18:35.953483 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cnxzp\" (UniqueName: \"kubernetes.io/projected/2d16884c-61ae-4288-b629-2a7c3be23ea8-kube-api-access-cnxzp\") pod \"2d16884c-61ae-4288-b629-2a7c3be23ea8\" (UID: \"2d16884c-61ae-4288-b629-2a7c3be23ea8\") " Oct 02 02:18:35 crc kubenswrapper[4885]: I1002 02:18:35.953631 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2d16884c-61ae-4288-b629-2a7c3be23ea8-ssh-key\") pod \"2d16884c-61ae-4288-b629-2a7c3be23ea8\" (UID: \"2d16884c-61ae-4288-b629-2a7c3be23ea8\") " Oct 02 02:18:35 crc kubenswrapper[4885]: I1002 02:18:35.961244 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d16884c-61ae-4288-b629-2a7c3be23ea8-kube-api-access-cnxzp" (OuterVolumeSpecName: "kube-api-access-cnxzp") pod "2d16884c-61ae-4288-b629-2a7c3be23ea8" (UID: "2d16884c-61ae-4288-b629-2a7c3be23ea8"). InnerVolumeSpecName "kube-api-access-cnxzp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:18:35 crc kubenswrapper[4885]: I1002 02:18:35.982528 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d16884c-61ae-4288-b629-2a7c3be23ea8-inventory" (OuterVolumeSpecName: "inventory") pod "2d16884c-61ae-4288-b629-2a7c3be23ea8" (UID: "2d16884c-61ae-4288-b629-2a7c3be23ea8"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:18:35 crc kubenswrapper[4885]: I1002 02:18:35.989997 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d16884c-61ae-4288-b629-2a7c3be23ea8-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2d16884c-61ae-4288-b629-2a7c3be23ea8" (UID: "2d16884c-61ae-4288-b629-2a7c3be23ea8"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:18:36 crc kubenswrapper[4885]: I1002 02:18:36.056015 4885 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2d16884c-61ae-4288-b629-2a7c3be23ea8-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 02:18:36 crc kubenswrapper[4885]: I1002 02:18:36.056049 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cnxzp\" (UniqueName: \"kubernetes.io/projected/2d16884c-61ae-4288-b629-2a7c3be23ea8-kube-api-access-cnxzp\") on node \"crc\" DevicePath \"\"" Oct 02 02:18:36 crc kubenswrapper[4885]: I1002 02:18:36.056061 4885 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2d16884c-61ae-4288-b629-2a7c3be23ea8-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 02:18:36 crc kubenswrapper[4885]: I1002 02:18:36.344591 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2rt6h" event={"ID":"2d16884c-61ae-4288-b629-2a7c3be23ea8","Type":"ContainerDied","Data":"8c5ff134d99b595ca94dbaa70f2401eef3621b96207f73a907782d3b6ca1c875"} Oct 02 02:18:36 crc kubenswrapper[4885]: I1002 02:18:36.344668 4885 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8c5ff134d99b595ca94dbaa70f2401eef3621b96207f73a907782d3b6ca1c875" Oct 02 02:18:36 crc kubenswrapper[4885]: I1002 02:18:36.344673 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2rt6h" Oct 02 02:18:36 crc kubenswrapper[4885]: I1002 02:18:36.424250 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-prq96"] Oct 02 02:18:36 crc kubenswrapper[4885]: E1002 02:18:36.424682 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d16884c-61ae-4288-b629-2a7c3be23ea8" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 02 02:18:36 crc kubenswrapper[4885]: I1002 02:18:36.424700 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d16884c-61ae-4288-b629-2a7c3be23ea8" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 02 02:18:36 crc kubenswrapper[4885]: I1002 02:18:36.424877 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d16884c-61ae-4288-b629-2a7c3be23ea8" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 02 02:18:36 crc kubenswrapper[4885]: I1002 02:18:36.425489 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-prq96" Oct 02 02:18:36 crc kubenswrapper[4885]: I1002 02:18:36.437701 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-f8tvb" Oct 02 02:18:36 crc kubenswrapper[4885]: I1002 02:18:36.437968 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Oct 02 02:18:36 crc kubenswrapper[4885]: I1002 02:18:36.438226 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 02:18:36 crc kubenswrapper[4885]: I1002 02:18:36.438419 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 02:18:36 crc kubenswrapper[4885]: I1002 02:18:36.438601 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Oct 02 02:18:36 crc kubenswrapper[4885]: I1002 02:18:36.438764 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Oct 02 02:18:36 crc kubenswrapper[4885]: I1002 02:18:36.438902 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Oct 02 02:18:36 crc kubenswrapper[4885]: I1002 02:18:36.439526 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 02:18:36 crc kubenswrapper[4885]: I1002 02:18:36.447835 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-prq96"] Oct 02 02:18:36 crc kubenswrapper[4885]: I1002 02:18:36.567950 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd405392-0e1a-42c5-a81a-cab2d229ea55-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-prq96\" (UID: \"bd405392-0e1a-42c5-a81a-cab2d229ea55\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-prq96" Oct 02 02:18:36 crc kubenswrapper[4885]: I1002 02:18:36.568008 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd405392-0e1a-42c5-a81a-cab2d229ea55-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-prq96\" (UID: \"bd405392-0e1a-42c5-a81a-cab2d229ea55\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-prq96" Oct 02 02:18:36 crc kubenswrapper[4885]: I1002 02:18:36.568059 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c7mc7\" (UniqueName: \"kubernetes.io/projected/bd405392-0e1a-42c5-a81a-cab2d229ea55-kube-api-access-c7mc7\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-prq96\" (UID: \"bd405392-0e1a-42c5-a81a-cab2d229ea55\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-prq96" Oct 02 02:18:36 crc kubenswrapper[4885]: I1002 02:18:36.568085 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/bd405392-0e1a-42c5-a81a-cab2d229ea55-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-prq96\" (UID: \"bd405392-0e1a-42c5-a81a-cab2d229ea55\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-prq96" Oct 02 02:18:36 crc kubenswrapper[4885]: I1002 02:18:36.568117 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd405392-0e1a-42c5-a81a-cab2d229ea55-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-prq96\" (UID: \"bd405392-0e1a-42c5-a81a-cab2d229ea55\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-prq96" Oct 02 02:18:36 crc kubenswrapper[4885]: I1002 02:18:36.569129 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd405392-0e1a-42c5-a81a-cab2d229ea55-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-prq96\" (UID: \"bd405392-0e1a-42c5-a81a-cab2d229ea55\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-prq96" Oct 02 02:18:36 crc kubenswrapper[4885]: I1002 02:18:36.569206 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bd405392-0e1a-42c5-a81a-cab2d229ea55-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-prq96\" (UID: \"bd405392-0e1a-42c5-a81a-cab2d229ea55\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-prq96" Oct 02 02:18:36 crc kubenswrapper[4885]: I1002 02:18:36.569599 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bd405392-0e1a-42c5-a81a-cab2d229ea55-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-prq96\" (UID: \"bd405392-0e1a-42c5-a81a-cab2d229ea55\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-prq96" Oct 02 02:18:36 crc kubenswrapper[4885]: I1002 02:18:36.569703 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/bd405392-0e1a-42c5-a81a-cab2d229ea55-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-prq96\" (UID: \"bd405392-0e1a-42c5-a81a-cab2d229ea55\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-prq96" Oct 02 02:18:36 crc kubenswrapper[4885]: I1002 02:18:36.569798 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/bd405392-0e1a-42c5-a81a-cab2d229ea55-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-prq96\" (UID: \"bd405392-0e1a-42c5-a81a-cab2d229ea55\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-prq96" Oct 02 02:18:36 crc kubenswrapper[4885]: I1002 02:18:36.569898 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd405392-0e1a-42c5-a81a-cab2d229ea55-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-prq96\" (UID: \"bd405392-0e1a-42c5-a81a-cab2d229ea55\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-prq96" Oct 02 02:18:36 crc kubenswrapper[4885]: I1002 02:18:36.569983 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/bd405392-0e1a-42c5-a81a-cab2d229ea55-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-prq96\" (UID: \"bd405392-0e1a-42c5-a81a-cab2d229ea55\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-prq96" Oct 02 02:18:36 crc kubenswrapper[4885]: I1002 02:18:36.570090 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd405392-0e1a-42c5-a81a-cab2d229ea55-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-prq96\" (UID: \"bd405392-0e1a-42c5-a81a-cab2d229ea55\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-prq96" Oct 02 02:18:36 crc kubenswrapper[4885]: I1002 02:18:36.570402 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd405392-0e1a-42c5-a81a-cab2d229ea55-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-prq96\" (UID: \"bd405392-0e1a-42c5-a81a-cab2d229ea55\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-prq96" Oct 02 02:18:36 crc kubenswrapper[4885]: I1002 02:18:36.672425 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd405392-0e1a-42c5-a81a-cab2d229ea55-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-prq96\" (UID: \"bd405392-0e1a-42c5-a81a-cab2d229ea55\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-prq96" Oct 02 02:18:36 crc kubenswrapper[4885]: I1002 02:18:36.672587 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c7mc7\" (UniqueName: \"kubernetes.io/projected/bd405392-0e1a-42c5-a81a-cab2d229ea55-kube-api-access-c7mc7\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-prq96\" (UID: \"bd405392-0e1a-42c5-a81a-cab2d229ea55\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-prq96" Oct 02 02:18:36 crc kubenswrapper[4885]: I1002 02:18:36.672650 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/bd405392-0e1a-42c5-a81a-cab2d229ea55-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-prq96\" (UID: \"bd405392-0e1a-42c5-a81a-cab2d229ea55\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-prq96" Oct 02 02:18:36 crc kubenswrapper[4885]: I1002 02:18:36.672707 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd405392-0e1a-42c5-a81a-cab2d229ea55-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-prq96\" (UID: \"bd405392-0e1a-42c5-a81a-cab2d229ea55\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-prq96" Oct 02 02:18:36 crc kubenswrapper[4885]: I1002 02:18:36.672754 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd405392-0e1a-42c5-a81a-cab2d229ea55-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-prq96\" (UID: \"bd405392-0e1a-42c5-a81a-cab2d229ea55\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-prq96" Oct 02 02:18:36 crc kubenswrapper[4885]: I1002 02:18:36.675243 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bd405392-0e1a-42c5-a81a-cab2d229ea55-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-prq96\" (UID: \"bd405392-0e1a-42c5-a81a-cab2d229ea55\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-prq96" Oct 02 02:18:36 crc kubenswrapper[4885]: I1002 02:18:36.675367 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bd405392-0e1a-42c5-a81a-cab2d229ea55-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-prq96\" (UID: \"bd405392-0e1a-42c5-a81a-cab2d229ea55\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-prq96" Oct 02 02:18:36 crc kubenswrapper[4885]: I1002 02:18:36.675439 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/bd405392-0e1a-42c5-a81a-cab2d229ea55-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-prq96\" (UID: \"bd405392-0e1a-42c5-a81a-cab2d229ea55\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-prq96" Oct 02 02:18:36 crc kubenswrapper[4885]: I1002 02:18:36.675478 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/bd405392-0e1a-42c5-a81a-cab2d229ea55-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-prq96\" (UID: \"bd405392-0e1a-42c5-a81a-cab2d229ea55\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-prq96" Oct 02 02:18:36 crc kubenswrapper[4885]: I1002 02:18:36.675547 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd405392-0e1a-42c5-a81a-cab2d229ea55-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-prq96\" (UID: \"bd405392-0e1a-42c5-a81a-cab2d229ea55\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-prq96" Oct 02 02:18:36 crc kubenswrapper[4885]: I1002 02:18:36.675614 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/bd405392-0e1a-42c5-a81a-cab2d229ea55-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-prq96\" (UID: \"bd405392-0e1a-42c5-a81a-cab2d229ea55\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-prq96" Oct 02 02:18:36 crc kubenswrapper[4885]: I1002 02:18:36.675675 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd405392-0e1a-42c5-a81a-cab2d229ea55-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-prq96\" (UID: \"bd405392-0e1a-42c5-a81a-cab2d229ea55\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-prq96" Oct 02 02:18:36 crc kubenswrapper[4885]: I1002 02:18:36.675763 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd405392-0e1a-42c5-a81a-cab2d229ea55-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-prq96\" (UID: \"bd405392-0e1a-42c5-a81a-cab2d229ea55\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-prq96" Oct 02 02:18:36 crc kubenswrapper[4885]: I1002 02:18:36.680775 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd405392-0e1a-42c5-a81a-cab2d229ea55-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-prq96\" (UID: \"bd405392-0e1a-42c5-a81a-cab2d229ea55\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-prq96" Oct 02 02:18:36 crc kubenswrapper[4885]: I1002 02:18:36.681296 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd405392-0e1a-42c5-a81a-cab2d229ea55-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-prq96\" (UID: \"bd405392-0e1a-42c5-a81a-cab2d229ea55\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-prq96" Oct 02 02:18:36 crc kubenswrapper[4885]: I1002 02:18:36.681415 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/bd405392-0e1a-42c5-a81a-cab2d229ea55-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-prq96\" (UID: \"bd405392-0e1a-42c5-a81a-cab2d229ea55\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-prq96" Oct 02 02:18:36 crc kubenswrapper[4885]: I1002 02:18:36.681461 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd405392-0e1a-42c5-a81a-cab2d229ea55-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-prq96\" (UID: \"bd405392-0e1a-42c5-a81a-cab2d229ea55\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-prq96" Oct 02 02:18:36 crc kubenswrapper[4885]: I1002 02:18:36.681531 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd405392-0e1a-42c5-a81a-cab2d229ea55-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-prq96\" (UID: \"bd405392-0e1a-42c5-a81a-cab2d229ea55\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-prq96" Oct 02 02:18:36 crc kubenswrapper[4885]: I1002 02:18:36.683557 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bd405392-0e1a-42c5-a81a-cab2d229ea55-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-prq96\" (UID: \"bd405392-0e1a-42c5-a81a-cab2d229ea55\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-prq96" Oct 02 02:18:36 crc kubenswrapper[4885]: I1002 02:18:36.684057 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd405392-0e1a-42c5-a81a-cab2d229ea55-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-prq96\" (UID: \"bd405392-0e1a-42c5-a81a-cab2d229ea55\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-prq96" Oct 02 02:18:36 crc kubenswrapper[4885]: I1002 02:18:36.684422 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/bd405392-0e1a-42c5-a81a-cab2d229ea55-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-prq96\" (UID: \"bd405392-0e1a-42c5-a81a-cab2d229ea55\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-prq96" Oct 02 02:18:36 crc kubenswrapper[4885]: I1002 02:18:36.685076 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd405392-0e1a-42c5-a81a-cab2d229ea55-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-prq96\" (UID: \"bd405392-0e1a-42c5-a81a-cab2d229ea55\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-prq96" Oct 02 02:18:36 crc kubenswrapper[4885]: I1002 02:18:36.685333 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd405392-0e1a-42c5-a81a-cab2d229ea55-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-prq96\" (UID: \"bd405392-0e1a-42c5-a81a-cab2d229ea55\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-prq96" Oct 02 02:18:36 crc kubenswrapper[4885]: I1002 02:18:36.685571 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/bd405392-0e1a-42c5-a81a-cab2d229ea55-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-prq96\" (UID: \"bd405392-0e1a-42c5-a81a-cab2d229ea55\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-prq96" Oct 02 02:18:36 crc kubenswrapper[4885]: I1002 02:18:36.685986 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd405392-0e1a-42c5-a81a-cab2d229ea55-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-prq96\" (UID: \"bd405392-0e1a-42c5-a81a-cab2d229ea55\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-prq96" Oct 02 02:18:36 crc kubenswrapper[4885]: I1002 02:18:36.687161 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/bd405392-0e1a-42c5-a81a-cab2d229ea55-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-prq96\" (UID: \"bd405392-0e1a-42c5-a81a-cab2d229ea55\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-prq96" Oct 02 02:18:36 crc kubenswrapper[4885]: I1002 02:18:36.688422 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bd405392-0e1a-42c5-a81a-cab2d229ea55-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-prq96\" (UID: \"bd405392-0e1a-42c5-a81a-cab2d229ea55\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-prq96" Oct 02 02:18:36 crc kubenswrapper[4885]: I1002 02:18:36.703857 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c7mc7\" (UniqueName: \"kubernetes.io/projected/bd405392-0e1a-42c5-a81a-cab2d229ea55-kube-api-access-c7mc7\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-prq96\" (UID: \"bd405392-0e1a-42c5-a81a-cab2d229ea55\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-prq96" Oct 02 02:18:36 crc kubenswrapper[4885]: I1002 02:18:36.782241 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-prq96" Oct 02 02:18:37 crc kubenswrapper[4885]: I1002 02:18:37.457300 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-prq96"] Oct 02 02:18:38 crc kubenswrapper[4885]: I1002 02:18:38.370576 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-prq96" event={"ID":"bd405392-0e1a-42c5-a81a-cab2d229ea55","Type":"ContainerStarted","Data":"04ef92fef3eefaf306a56d0d3156d30a8a129310fd1a06613979cabcf3357d68"} Oct 02 02:18:38 crc kubenswrapper[4885]: I1002 02:18:38.370823 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-prq96" event={"ID":"bd405392-0e1a-42c5-a81a-cab2d229ea55","Type":"ContainerStarted","Data":"5f9d054aa0866f4a0b2d83cf8298dcb5f0a9ed460f0c134702f9a578d4882ddf"} Oct 02 02:18:38 crc kubenswrapper[4885]: I1002 02:18:38.401839 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-prq96" podStartSLOduration=2.003722566 podStartE2EDuration="2.401818315s" podCreationTimestamp="2025-10-02 02:18:36 +0000 UTC" firstStartedPulling="2025-10-02 02:18:37.471104562 +0000 UTC m=+1906.282852001" lastFinishedPulling="2025-10-02 02:18:37.869200351 +0000 UTC m=+1906.680947750" observedRunningTime="2025-10-02 02:18:38.395189183 +0000 UTC m=+1907.206936622" watchObservedRunningTime="2025-10-02 02:18:38.401818315 +0000 UTC m=+1907.213565734" Oct 02 02:19:20 crc kubenswrapper[4885]: I1002 02:19:20.828368 4885 generic.go:334] "Generic (PLEG): container finished" podID="bd405392-0e1a-42c5-a81a-cab2d229ea55" containerID="04ef92fef3eefaf306a56d0d3156d30a8a129310fd1a06613979cabcf3357d68" exitCode=0 Oct 02 02:19:20 crc kubenswrapper[4885]: I1002 02:19:20.828450 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-prq96" event={"ID":"bd405392-0e1a-42c5-a81a-cab2d229ea55","Type":"ContainerDied","Data":"04ef92fef3eefaf306a56d0d3156d30a8a129310fd1a06613979cabcf3357d68"} Oct 02 02:19:22 crc kubenswrapper[4885]: I1002 02:19:22.336456 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-prq96" Oct 02 02:19:22 crc kubenswrapper[4885]: I1002 02:19:22.493195 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bd405392-0e1a-42c5-a81a-cab2d229ea55-inventory\") pod \"bd405392-0e1a-42c5-a81a-cab2d229ea55\" (UID: \"bd405392-0e1a-42c5-a81a-cab2d229ea55\") " Oct 02 02:19:22 crc kubenswrapper[4885]: I1002 02:19:22.493315 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/bd405392-0e1a-42c5-a81a-cab2d229ea55-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"bd405392-0e1a-42c5-a81a-cab2d229ea55\" (UID: \"bd405392-0e1a-42c5-a81a-cab2d229ea55\") " Oct 02 02:19:22 crc kubenswrapper[4885]: I1002 02:19:22.493424 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/bd405392-0e1a-42c5-a81a-cab2d229ea55-openstack-edpm-ipam-ovn-default-certs-0\") pod \"bd405392-0e1a-42c5-a81a-cab2d229ea55\" (UID: \"bd405392-0e1a-42c5-a81a-cab2d229ea55\") " Oct 02 02:19:22 crc kubenswrapper[4885]: I1002 02:19:22.493508 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bd405392-0e1a-42c5-a81a-cab2d229ea55-ssh-key\") pod \"bd405392-0e1a-42c5-a81a-cab2d229ea55\" (UID: \"bd405392-0e1a-42c5-a81a-cab2d229ea55\") " Oct 02 02:19:22 crc kubenswrapper[4885]: I1002 02:19:22.493597 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd405392-0e1a-42c5-a81a-cab2d229ea55-telemetry-combined-ca-bundle\") pod \"bd405392-0e1a-42c5-a81a-cab2d229ea55\" (UID: \"bd405392-0e1a-42c5-a81a-cab2d229ea55\") " Oct 02 02:19:22 crc kubenswrapper[4885]: I1002 02:19:22.493643 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd405392-0e1a-42c5-a81a-cab2d229ea55-neutron-metadata-combined-ca-bundle\") pod \"bd405392-0e1a-42c5-a81a-cab2d229ea55\" (UID: \"bd405392-0e1a-42c5-a81a-cab2d229ea55\") " Oct 02 02:19:22 crc kubenswrapper[4885]: I1002 02:19:22.493703 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c7mc7\" (UniqueName: \"kubernetes.io/projected/bd405392-0e1a-42c5-a81a-cab2d229ea55-kube-api-access-c7mc7\") pod \"bd405392-0e1a-42c5-a81a-cab2d229ea55\" (UID: \"bd405392-0e1a-42c5-a81a-cab2d229ea55\") " Oct 02 02:19:22 crc kubenswrapper[4885]: I1002 02:19:22.493736 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd405392-0e1a-42c5-a81a-cab2d229ea55-repo-setup-combined-ca-bundle\") pod \"bd405392-0e1a-42c5-a81a-cab2d229ea55\" (UID: \"bd405392-0e1a-42c5-a81a-cab2d229ea55\") " Oct 02 02:19:22 crc kubenswrapper[4885]: I1002 02:19:22.493816 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd405392-0e1a-42c5-a81a-cab2d229ea55-ovn-combined-ca-bundle\") pod \"bd405392-0e1a-42c5-a81a-cab2d229ea55\" (UID: \"bd405392-0e1a-42c5-a81a-cab2d229ea55\") " Oct 02 02:19:22 crc kubenswrapper[4885]: I1002 02:19:22.493855 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/bd405392-0e1a-42c5-a81a-cab2d229ea55-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"bd405392-0e1a-42c5-a81a-cab2d229ea55\" (UID: \"bd405392-0e1a-42c5-a81a-cab2d229ea55\") " Oct 02 02:19:22 crc kubenswrapper[4885]: I1002 02:19:22.493893 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd405392-0e1a-42c5-a81a-cab2d229ea55-libvirt-combined-ca-bundle\") pod \"bd405392-0e1a-42c5-a81a-cab2d229ea55\" (UID: \"bd405392-0e1a-42c5-a81a-cab2d229ea55\") " Oct 02 02:19:22 crc kubenswrapper[4885]: I1002 02:19:22.494084 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd405392-0e1a-42c5-a81a-cab2d229ea55-nova-combined-ca-bundle\") pod \"bd405392-0e1a-42c5-a81a-cab2d229ea55\" (UID: \"bd405392-0e1a-42c5-a81a-cab2d229ea55\") " Oct 02 02:19:22 crc kubenswrapper[4885]: I1002 02:19:22.494528 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd405392-0e1a-42c5-a81a-cab2d229ea55-bootstrap-combined-ca-bundle\") pod \"bd405392-0e1a-42c5-a81a-cab2d229ea55\" (UID: \"bd405392-0e1a-42c5-a81a-cab2d229ea55\") " Oct 02 02:19:22 crc kubenswrapper[4885]: I1002 02:19:22.494588 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/bd405392-0e1a-42c5-a81a-cab2d229ea55-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"bd405392-0e1a-42c5-a81a-cab2d229ea55\" (UID: \"bd405392-0e1a-42c5-a81a-cab2d229ea55\") " Oct 02 02:19:22 crc kubenswrapper[4885]: I1002 02:19:22.501339 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd405392-0e1a-42c5-a81a-cab2d229ea55-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "bd405392-0e1a-42c5-a81a-cab2d229ea55" (UID: "bd405392-0e1a-42c5-a81a-cab2d229ea55"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:19:22 crc kubenswrapper[4885]: I1002 02:19:22.502883 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd405392-0e1a-42c5-a81a-cab2d229ea55-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "bd405392-0e1a-42c5-a81a-cab2d229ea55" (UID: "bd405392-0e1a-42c5-a81a-cab2d229ea55"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:19:22 crc kubenswrapper[4885]: I1002 02:19:22.504352 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd405392-0e1a-42c5-a81a-cab2d229ea55-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "bd405392-0e1a-42c5-a81a-cab2d229ea55" (UID: "bd405392-0e1a-42c5-a81a-cab2d229ea55"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:19:22 crc kubenswrapper[4885]: I1002 02:19:22.505323 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd405392-0e1a-42c5-a81a-cab2d229ea55-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "bd405392-0e1a-42c5-a81a-cab2d229ea55" (UID: "bd405392-0e1a-42c5-a81a-cab2d229ea55"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:19:22 crc kubenswrapper[4885]: I1002 02:19:22.505670 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd405392-0e1a-42c5-a81a-cab2d229ea55-kube-api-access-c7mc7" (OuterVolumeSpecName: "kube-api-access-c7mc7") pod "bd405392-0e1a-42c5-a81a-cab2d229ea55" (UID: "bd405392-0e1a-42c5-a81a-cab2d229ea55"). InnerVolumeSpecName "kube-api-access-c7mc7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:19:22 crc kubenswrapper[4885]: I1002 02:19:22.506216 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd405392-0e1a-42c5-a81a-cab2d229ea55-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "bd405392-0e1a-42c5-a81a-cab2d229ea55" (UID: "bd405392-0e1a-42c5-a81a-cab2d229ea55"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:19:22 crc kubenswrapper[4885]: I1002 02:19:22.506294 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd405392-0e1a-42c5-a81a-cab2d229ea55-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "bd405392-0e1a-42c5-a81a-cab2d229ea55" (UID: "bd405392-0e1a-42c5-a81a-cab2d229ea55"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:19:22 crc kubenswrapper[4885]: I1002 02:19:22.510886 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd405392-0e1a-42c5-a81a-cab2d229ea55-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "bd405392-0e1a-42c5-a81a-cab2d229ea55" (UID: "bd405392-0e1a-42c5-a81a-cab2d229ea55"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:19:22 crc kubenswrapper[4885]: I1002 02:19:22.513506 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd405392-0e1a-42c5-a81a-cab2d229ea55-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "bd405392-0e1a-42c5-a81a-cab2d229ea55" (UID: "bd405392-0e1a-42c5-a81a-cab2d229ea55"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:19:22 crc kubenswrapper[4885]: I1002 02:19:22.517924 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd405392-0e1a-42c5-a81a-cab2d229ea55-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "bd405392-0e1a-42c5-a81a-cab2d229ea55" (UID: "bd405392-0e1a-42c5-a81a-cab2d229ea55"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:19:22 crc kubenswrapper[4885]: I1002 02:19:22.523936 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd405392-0e1a-42c5-a81a-cab2d229ea55-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "bd405392-0e1a-42c5-a81a-cab2d229ea55" (UID: "bd405392-0e1a-42c5-a81a-cab2d229ea55"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:19:22 crc kubenswrapper[4885]: I1002 02:19:22.530826 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd405392-0e1a-42c5-a81a-cab2d229ea55-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "bd405392-0e1a-42c5-a81a-cab2d229ea55" (UID: "bd405392-0e1a-42c5-a81a-cab2d229ea55"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:19:22 crc kubenswrapper[4885]: I1002 02:19:22.542413 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd405392-0e1a-42c5-a81a-cab2d229ea55-inventory" (OuterVolumeSpecName: "inventory") pod "bd405392-0e1a-42c5-a81a-cab2d229ea55" (UID: "bd405392-0e1a-42c5-a81a-cab2d229ea55"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:19:22 crc kubenswrapper[4885]: I1002 02:19:22.545817 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd405392-0e1a-42c5-a81a-cab2d229ea55-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "bd405392-0e1a-42c5-a81a-cab2d229ea55" (UID: "bd405392-0e1a-42c5-a81a-cab2d229ea55"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:19:22 crc kubenswrapper[4885]: I1002 02:19:22.596917 4885 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd405392-0e1a-42c5-a81a-cab2d229ea55-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:19:22 crc kubenswrapper[4885]: I1002 02:19:22.596954 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c7mc7\" (UniqueName: \"kubernetes.io/projected/bd405392-0e1a-42c5-a81a-cab2d229ea55-kube-api-access-c7mc7\") on node \"crc\" DevicePath \"\"" Oct 02 02:19:22 crc kubenswrapper[4885]: I1002 02:19:22.596963 4885 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd405392-0e1a-42c5-a81a-cab2d229ea55-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:19:22 crc kubenswrapper[4885]: I1002 02:19:22.596973 4885 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/bd405392-0e1a-42c5-a81a-cab2d229ea55-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 02 02:19:22 crc kubenswrapper[4885]: I1002 02:19:22.596984 4885 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd405392-0e1a-42c5-a81a-cab2d229ea55-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:19:22 crc kubenswrapper[4885]: I1002 02:19:22.596994 4885 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd405392-0e1a-42c5-a81a-cab2d229ea55-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:19:22 crc kubenswrapper[4885]: I1002 02:19:22.597002 4885 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd405392-0e1a-42c5-a81a-cab2d229ea55-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:19:22 crc kubenswrapper[4885]: I1002 02:19:22.597012 4885 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/bd405392-0e1a-42c5-a81a-cab2d229ea55-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 02 02:19:22 crc kubenswrapper[4885]: I1002 02:19:22.597024 4885 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bd405392-0e1a-42c5-a81a-cab2d229ea55-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 02:19:22 crc kubenswrapper[4885]: I1002 02:19:22.597032 4885 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/bd405392-0e1a-42c5-a81a-cab2d229ea55-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 02 02:19:22 crc kubenswrapper[4885]: I1002 02:19:22.597042 4885 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/bd405392-0e1a-42c5-a81a-cab2d229ea55-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 02 02:19:22 crc kubenswrapper[4885]: I1002 02:19:22.597050 4885 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bd405392-0e1a-42c5-a81a-cab2d229ea55-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 02:19:22 crc kubenswrapper[4885]: I1002 02:19:22.597059 4885 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd405392-0e1a-42c5-a81a-cab2d229ea55-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:19:22 crc kubenswrapper[4885]: I1002 02:19:22.597068 4885 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd405392-0e1a-42c5-a81a-cab2d229ea55-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:19:22 crc kubenswrapper[4885]: I1002 02:19:22.848326 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-prq96" event={"ID":"bd405392-0e1a-42c5-a81a-cab2d229ea55","Type":"ContainerDied","Data":"5f9d054aa0866f4a0b2d83cf8298dcb5f0a9ed460f0c134702f9a578d4882ddf"} Oct 02 02:19:22 crc kubenswrapper[4885]: I1002 02:19:22.849192 4885 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5f9d054aa0866f4a0b2d83cf8298dcb5f0a9ed460f0c134702f9a578d4882ddf" Oct 02 02:19:22 crc kubenswrapper[4885]: I1002 02:19:22.848372 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-prq96" Oct 02 02:19:22 crc kubenswrapper[4885]: I1002 02:19:22.973666 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-xkwk4"] Oct 02 02:19:22 crc kubenswrapper[4885]: E1002 02:19:22.974063 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd405392-0e1a-42c5-a81a-cab2d229ea55" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Oct 02 02:19:22 crc kubenswrapper[4885]: I1002 02:19:22.974080 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd405392-0e1a-42c5-a81a-cab2d229ea55" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Oct 02 02:19:22 crc kubenswrapper[4885]: I1002 02:19:22.974312 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd405392-0e1a-42c5-a81a-cab2d229ea55" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Oct 02 02:19:22 crc kubenswrapper[4885]: I1002 02:19:22.975028 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xkwk4" Oct 02 02:19:22 crc kubenswrapper[4885]: I1002 02:19:22.982473 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-f8tvb" Oct 02 02:19:22 crc kubenswrapper[4885]: I1002 02:19:22.982560 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Oct 02 02:19:22 crc kubenswrapper[4885]: I1002 02:19:22.982595 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 02:19:22 crc kubenswrapper[4885]: I1002 02:19:22.982745 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 02:19:22 crc kubenswrapper[4885]: I1002 02:19:22.982800 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 02:19:22 crc kubenswrapper[4885]: I1002 02:19:22.989079 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-xkwk4"] Oct 02 02:19:23 crc kubenswrapper[4885]: I1002 02:19:23.104488 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d8d7cf22-9bf7-417b-b4e2-d199cfccf949-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xkwk4\" (UID: \"d8d7cf22-9bf7-417b-b4e2-d199cfccf949\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xkwk4" Oct 02 02:19:23 crc kubenswrapper[4885]: I1002 02:19:23.104661 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/d8d7cf22-9bf7-417b-b4e2-d199cfccf949-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xkwk4\" (UID: \"d8d7cf22-9bf7-417b-b4e2-d199cfccf949\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xkwk4" Oct 02 02:19:23 crc kubenswrapper[4885]: I1002 02:19:23.104793 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d8d7cf22-9bf7-417b-b4e2-d199cfccf949-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xkwk4\" (UID: \"d8d7cf22-9bf7-417b-b4e2-d199cfccf949\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xkwk4" Oct 02 02:19:23 crc kubenswrapper[4885]: I1002 02:19:23.105046 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zscrp\" (UniqueName: \"kubernetes.io/projected/d8d7cf22-9bf7-417b-b4e2-d199cfccf949-kube-api-access-zscrp\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xkwk4\" (UID: \"d8d7cf22-9bf7-417b-b4e2-d199cfccf949\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xkwk4" Oct 02 02:19:23 crc kubenswrapper[4885]: I1002 02:19:23.105102 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8d7cf22-9bf7-417b-b4e2-d199cfccf949-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xkwk4\" (UID: \"d8d7cf22-9bf7-417b-b4e2-d199cfccf949\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xkwk4" Oct 02 02:19:23 crc kubenswrapper[4885]: I1002 02:19:23.206681 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d8d7cf22-9bf7-417b-b4e2-d199cfccf949-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xkwk4\" (UID: \"d8d7cf22-9bf7-417b-b4e2-d199cfccf949\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xkwk4" Oct 02 02:19:23 crc kubenswrapper[4885]: I1002 02:19:23.207369 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zscrp\" (UniqueName: \"kubernetes.io/projected/d8d7cf22-9bf7-417b-b4e2-d199cfccf949-kube-api-access-zscrp\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xkwk4\" (UID: \"d8d7cf22-9bf7-417b-b4e2-d199cfccf949\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xkwk4" Oct 02 02:19:23 crc kubenswrapper[4885]: I1002 02:19:23.207402 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8d7cf22-9bf7-417b-b4e2-d199cfccf949-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xkwk4\" (UID: \"d8d7cf22-9bf7-417b-b4e2-d199cfccf949\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xkwk4" Oct 02 02:19:23 crc kubenswrapper[4885]: I1002 02:19:23.207479 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d8d7cf22-9bf7-417b-b4e2-d199cfccf949-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xkwk4\" (UID: \"d8d7cf22-9bf7-417b-b4e2-d199cfccf949\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xkwk4" Oct 02 02:19:23 crc kubenswrapper[4885]: I1002 02:19:23.207516 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/d8d7cf22-9bf7-417b-b4e2-d199cfccf949-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xkwk4\" (UID: \"d8d7cf22-9bf7-417b-b4e2-d199cfccf949\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xkwk4" Oct 02 02:19:23 crc kubenswrapper[4885]: I1002 02:19:23.208251 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/d8d7cf22-9bf7-417b-b4e2-d199cfccf949-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xkwk4\" (UID: \"d8d7cf22-9bf7-417b-b4e2-d199cfccf949\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xkwk4" Oct 02 02:19:23 crc kubenswrapper[4885]: I1002 02:19:23.212378 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d8d7cf22-9bf7-417b-b4e2-d199cfccf949-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xkwk4\" (UID: \"d8d7cf22-9bf7-417b-b4e2-d199cfccf949\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xkwk4" Oct 02 02:19:23 crc kubenswrapper[4885]: I1002 02:19:23.212768 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d8d7cf22-9bf7-417b-b4e2-d199cfccf949-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xkwk4\" (UID: \"d8d7cf22-9bf7-417b-b4e2-d199cfccf949\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xkwk4" Oct 02 02:19:23 crc kubenswrapper[4885]: I1002 02:19:23.222211 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8d7cf22-9bf7-417b-b4e2-d199cfccf949-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xkwk4\" (UID: \"d8d7cf22-9bf7-417b-b4e2-d199cfccf949\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xkwk4" Oct 02 02:19:23 crc kubenswrapper[4885]: I1002 02:19:23.228059 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zscrp\" (UniqueName: \"kubernetes.io/projected/d8d7cf22-9bf7-417b-b4e2-d199cfccf949-kube-api-access-zscrp\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xkwk4\" (UID: \"d8d7cf22-9bf7-417b-b4e2-d199cfccf949\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xkwk4" Oct 02 02:19:23 crc kubenswrapper[4885]: I1002 02:19:23.296719 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xkwk4" Oct 02 02:19:23 crc kubenswrapper[4885]: I1002 02:19:23.853412 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-xkwk4"] Oct 02 02:19:24 crc kubenswrapper[4885]: I1002 02:19:24.867458 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xkwk4" event={"ID":"d8d7cf22-9bf7-417b-b4e2-d199cfccf949","Type":"ContainerStarted","Data":"3680971d810b937dc0a88eb5f0d4cb6a4ae292a4d06af9ca7a6cb03548b96a84"} Oct 02 02:19:25 crc kubenswrapper[4885]: I1002 02:19:25.880148 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xkwk4" event={"ID":"d8d7cf22-9bf7-417b-b4e2-d199cfccf949","Type":"ContainerStarted","Data":"391d87e824c65f8ba8760f8c6e88849b359bb2c38bd9bb1cedfd195693823c26"} Oct 02 02:19:26 crc kubenswrapper[4885]: I1002 02:19:26.909439 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xkwk4" podStartSLOduration=3.325749518 podStartE2EDuration="4.909420548s" podCreationTimestamp="2025-10-02 02:19:22 +0000 UTC" firstStartedPulling="2025-10-02 02:19:23.870343103 +0000 UTC m=+1952.682090502" lastFinishedPulling="2025-10-02 02:19:25.454014093 +0000 UTC m=+1954.265761532" observedRunningTime="2025-10-02 02:19:26.904604921 +0000 UTC m=+1955.716352320" watchObservedRunningTime="2025-10-02 02:19:26.909420548 +0000 UTC m=+1955.721167947" Oct 02 02:20:13 crc kubenswrapper[4885]: I1002 02:20:13.265659 4885 patch_prober.go:28] interesting pod/machine-config-daemon-rttx8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 02:20:13 crc kubenswrapper[4885]: I1002 02:20:13.266222 4885 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 02:20:32 crc kubenswrapper[4885]: I1002 02:20:32.633813 4885 generic.go:334] "Generic (PLEG): container finished" podID="d8d7cf22-9bf7-417b-b4e2-d199cfccf949" containerID="391d87e824c65f8ba8760f8c6e88849b359bb2c38bd9bb1cedfd195693823c26" exitCode=0 Oct 02 02:20:32 crc kubenswrapper[4885]: I1002 02:20:32.633914 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xkwk4" event={"ID":"d8d7cf22-9bf7-417b-b4e2-d199cfccf949","Type":"ContainerDied","Data":"391d87e824c65f8ba8760f8c6e88849b359bb2c38bd9bb1cedfd195693823c26"} Oct 02 02:20:34 crc kubenswrapper[4885]: I1002 02:20:34.113191 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xkwk4" Oct 02 02:20:34 crc kubenswrapper[4885]: I1002 02:20:34.271275 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d8d7cf22-9bf7-417b-b4e2-d199cfccf949-inventory\") pod \"d8d7cf22-9bf7-417b-b4e2-d199cfccf949\" (UID: \"d8d7cf22-9bf7-417b-b4e2-d199cfccf949\") " Oct 02 02:20:34 crc kubenswrapper[4885]: I1002 02:20:34.271327 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zscrp\" (UniqueName: \"kubernetes.io/projected/d8d7cf22-9bf7-417b-b4e2-d199cfccf949-kube-api-access-zscrp\") pod \"d8d7cf22-9bf7-417b-b4e2-d199cfccf949\" (UID: \"d8d7cf22-9bf7-417b-b4e2-d199cfccf949\") " Oct 02 02:20:34 crc kubenswrapper[4885]: I1002 02:20:34.271421 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d8d7cf22-9bf7-417b-b4e2-d199cfccf949-ssh-key\") pod \"d8d7cf22-9bf7-417b-b4e2-d199cfccf949\" (UID: \"d8d7cf22-9bf7-417b-b4e2-d199cfccf949\") " Oct 02 02:20:34 crc kubenswrapper[4885]: I1002 02:20:34.271458 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8d7cf22-9bf7-417b-b4e2-d199cfccf949-ovn-combined-ca-bundle\") pod \"d8d7cf22-9bf7-417b-b4e2-d199cfccf949\" (UID: \"d8d7cf22-9bf7-417b-b4e2-d199cfccf949\") " Oct 02 02:20:34 crc kubenswrapper[4885]: I1002 02:20:34.271528 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/d8d7cf22-9bf7-417b-b4e2-d199cfccf949-ovncontroller-config-0\") pod \"d8d7cf22-9bf7-417b-b4e2-d199cfccf949\" (UID: \"d8d7cf22-9bf7-417b-b4e2-d199cfccf949\") " Oct 02 02:20:34 crc kubenswrapper[4885]: I1002 02:20:34.277280 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d8d7cf22-9bf7-417b-b4e2-d199cfccf949-kube-api-access-zscrp" (OuterVolumeSpecName: "kube-api-access-zscrp") pod "d8d7cf22-9bf7-417b-b4e2-d199cfccf949" (UID: "d8d7cf22-9bf7-417b-b4e2-d199cfccf949"). InnerVolumeSpecName "kube-api-access-zscrp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:20:34 crc kubenswrapper[4885]: I1002 02:20:34.277499 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8d7cf22-9bf7-417b-b4e2-d199cfccf949-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "d8d7cf22-9bf7-417b-b4e2-d199cfccf949" (UID: "d8d7cf22-9bf7-417b-b4e2-d199cfccf949"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:20:34 crc kubenswrapper[4885]: I1002 02:20:34.304059 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d8d7cf22-9bf7-417b-b4e2-d199cfccf949-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "d8d7cf22-9bf7-417b-b4e2-d199cfccf949" (UID: "d8d7cf22-9bf7-417b-b4e2-d199cfccf949"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:20:34 crc kubenswrapper[4885]: I1002 02:20:34.304572 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8d7cf22-9bf7-417b-b4e2-d199cfccf949-inventory" (OuterVolumeSpecName: "inventory") pod "d8d7cf22-9bf7-417b-b4e2-d199cfccf949" (UID: "d8d7cf22-9bf7-417b-b4e2-d199cfccf949"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:20:34 crc kubenswrapper[4885]: I1002 02:20:34.327679 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8d7cf22-9bf7-417b-b4e2-d199cfccf949-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d8d7cf22-9bf7-417b-b4e2-d199cfccf949" (UID: "d8d7cf22-9bf7-417b-b4e2-d199cfccf949"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:20:34 crc kubenswrapper[4885]: I1002 02:20:34.374181 4885 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d8d7cf22-9bf7-417b-b4e2-d199cfccf949-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 02:20:34 crc kubenswrapper[4885]: I1002 02:20:34.374408 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zscrp\" (UniqueName: \"kubernetes.io/projected/d8d7cf22-9bf7-417b-b4e2-d199cfccf949-kube-api-access-zscrp\") on node \"crc\" DevicePath \"\"" Oct 02 02:20:34 crc kubenswrapper[4885]: I1002 02:20:34.374542 4885 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d8d7cf22-9bf7-417b-b4e2-d199cfccf949-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 02:20:34 crc kubenswrapper[4885]: I1002 02:20:34.374641 4885 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8d7cf22-9bf7-417b-b4e2-d199cfccf949-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:20:34 crc kubenswrapper[4885]: I1002 02:20:34.374734 4885 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/d8d7cf22-9bf7-417b-b4e2-d199cfccf949-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Oct 02 02:20:34 crc kubenswrapper[4885]: I1002 02:20:34.653249 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xkwk4" event={"ID":"d8d7cf22-9bf7-417b-b4e2-d199cfccf949","Type":"ContainerDied","Data":"3680971d810b937dc0a88eb5f0d4cb6a4ae292a4d06af9ca7a6cb03548b96a84"} Oct 02 02:20:34 crc kubenswrapper[4885]: I1002 02:20:34.653357 4885 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3680971d810b937dc0a88eb5f0d4cb6a4ae292a4d06af9ca7a6cb03548b96a84" Oct 02 02:20:34 crc kubenswrapper[4885]: I1002 02:20:34.653426 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xkwk4" Oct 02 02:20:34 crc kubenswrapper[4885]: I1002 02:20:34.764145 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-2hp4t"] Oct 02 02:20:34 crc kubenswrapper[4885]: E1002 02:20:34.764855 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8d7cf22-9bf7-417b-b4e2-d199cfccf949" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 02 02:20:34 crc kubenswrapper[4885]: I1002 02:20:34.764875 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8d7cf22-9bf7-417b-b4e2-d199cfccf949" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 02 02:20:34 crc kubenswrapper[4885]: I1002 02:20:34.765131 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8d7cf22-9bf7-417b-b4e2-d199cfccf949" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 02 02:20:34 crc kubenswrapper[4885]: I1002 02:20:34.765909 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-2hp4t" Oct 02 02:20:34 crc kubenswrapper[4885]: I1002 02:20:34.774838 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 02:20:34 crc kubenswrapper[4885]: I1002 02:20:34.775425 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 02:20:34 crc kubenswrapper[4885]: I1002 02:20:34.777074 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Oct 02 02:20:34 crc kubenswrapper[4885]: I1002 02:20:34.777282 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 02:20:34 crc kubenswrapper[4885]: I1002 02:20:34.777435 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Oct 02 02:20:34 crc kubenswrapper[4885]: I1002 02:20:34.777560 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-f8tvb" Oct 02 02:20:34 crc kubenswrapper[4885]: I1002 02:20:34.791880 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-2hp4t"] Oct 02 02:20:34 crc kubenswrapper[4885]: I1002 02:20:34.887409 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f6edca36-b630-4682-93ed-2c0c2dcadecb-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-2hp4t\" (UID: \"f6edca36-b630-4682-93ed-2c0c2dcadecb\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-2hp4t" Oct 02 02:20:34 crc kubenswrapper[4885]: I1002 02:20:34.887703 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k5bj9\" (UniqueName: \"kubernetes.io/projected/f6edca36-b630-4682-93ed-2c0c2dcadecb-kube-api-access-k5bj9\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-2hp4t\" (UID: \"f6edca36-b630-4682-93ed-2c0c2dcadecb\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-2hp4t" Oct 02 02:20:34 crc kubenswrapper[4885]: I1002 02:20:34.887764 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/f6edca36-b630-4682-93ed-2c0c2dcadecb-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-2hp4t\" (UID: \"f6edca36-b630-4682-93ed-2c0c2dcadecb\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-2hp4t" Oct 02 02:20:34 crc kubenswrapper[4885]: I1002 02:20:34.887968 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6edca36-b630-4682-93ed-2c0c2dcadecb-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-2hp4t\" (UID: \"f6edca36-b630-4682-93ed-2c0c2dcadecb\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-2hp4t" Oct 02 02:20:34 crc kubenswrapper[4885]: I1002 02:20:34.888035 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f6edca36-b630-4682-93ed-2c0c2dcadecb-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-2hp4t\" (UID: \"f6edca36-b630-4682-93ed-2c0c2dcadecb\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-2hp4t" Oct 02 02:20:34 crc kubenswrapper[4885]: I1002 02:20:34.889728 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/f6edca36-b630-4682-93ed-2c0c2dcadecb-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-2hp4t\" (UID: \"f6edca36-b630-4682-93ed-2c0c2dcadecb\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-2hp4t" Oct 02 02:20:34 crc kubenswrapper[4885]: I1002 02:20:34.991650 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/f6edca36-b630-4682-93ed-2c0c2dcadecb-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-2hp4t\" (UID: \"f6edca36-b630-4682-93ed-2c0c2dcadecb\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-2hp4t" Oct 02 02:20:34 crc kubenswrapper[4885]: I1002 02:20:34.991764 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f6edca36-b630-4682-93ed-2c0c2dcadecb-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-2hp4t\" (UID: \"f6edca36-b630-4682-93ed-2c0c2dcadecb\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-2hp4t" Oct 02 02:20:34 crc kubenswrapper[4885]: I1002 02:20:34.991858 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k5bj9\" (UniqueName: \"kubernetes.io/projected/f6edca36-b630-4682-93ed-2c0c2dcadecb-kube-api-access-k5bj9\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-2hp4t\" (UID: \"f6edca36-b630-4682-93ed-2c0c2dcadecb\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-2hp4t" Oct 02 02:20:34 crc kubenswrapper[4885]: I1002 02:20:34.991907 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/f6edca36-b630-4682-93ed-2c0c2dcadecb-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-2hp4t\" (UID: \"f6edca36-b630-4682-93ed-2c0c2dcadecb\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-2hp4t" Oct 02 02:20:34 crc kubenswrapper[4885]: I1002 02:20:34.992033 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6edca36-b630-4682-93ed-2c0c2dcadecb-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-2hp4t\" (UID: \"f6edca36-b630-4682-93ed-2c0c2dcadecb\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-2hp4t" Oct 02 02:20:34 crc kubenswrapper[4885]: I1002 02:20:34.992092 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f6edca36-b630-4682-93ed-2c0c2dcadecb-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-2hp4t\" (UID: \"f6edca36-b630-4682-93ed-2c0c2dcadecb\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-2hp4t" Oct 02 02:20:34 crc kubenswrapper[4885]: I1002 02:20:34.997056 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f6edca36-b630-4682-93ed-2c0c2dcadecb-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-2hp4t\" (UID: \"f6edca36-b630-4682-93ed-2c0c2dcadecb\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-2hp4t" Oct 02 02:20:34 crc kubenswrapper[4885]: I1002 02:20:34.997456 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/f6edca36-b630-4682-93ed-2c0c2dcadecb-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-2hp4t\" (UID: \"f6edca36-b630-4682-93ed-2c0c2dcadecb\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-2hp4t" Oct 02 02:20:34 crc kubenswrapper[4885]: I1002 02:20:34.997755 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f6edca36-b630-4682-93ed-2c0c2dcadecb-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-2hp4t\" (UID: \"f6edca36-b630-4682-93ed-2c0c2dcadecb\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-2hp4t" Oct 02 02:20:34 crc kubenswrapper[4885]: I1002 02:20:34.998814 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/f6edca36-b630-4682-93ed-2c0c2dcadecb-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-2hp4t\" (UID: \"f6edca36-b630-4682-93ed-2c0c2dcadecb\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-2hp4t" Oct 02 02:20:35 crc kubenswrapper[4885]: I1002 02:20:35.004976 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6edca36-b630-4682-93ed-2c0c2dcadecb-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-2hp4t\" (UID: \"f6edca36-b630-4682-93ed-2c0c2dcadecb\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-2hp4t" Oct 02 02:20:35 crc kubenswrapper[4885]: I1002 02:20:35.021230 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k5bj9\" (UniqueName: \"kubernetes.io/projected/f6edca36-b630-4682-93ed-2c0c2dcadecb-kube-api-access-k5bj9\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-2hp4t\" (UID: \"f6edca36-b630-4682-93ed-2c0c2dcadecb\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-2hp4t" Oct 02 02:20:35 crc kubenswrapper[4885]: I1002 02:20:35.091776 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-2hp4t" Oct 02 02:20:35 crc kubenswrapper[4885]: I1002 02:20:35.681228 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-2hp4t"] Oct 02 02:20:36 crc kubenswrapper[4885]: I1002 02:20:36.673782 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-2hp4t" event={"ID":"f6edca36-b630-4682-93ed-2c0c2dcadecb","Type":"ContainerStarted","Data":"71af0ebf476ced4ef2578ffe478331e575a27eb964f3d2cb9f2f995871786bcb"} Oct 02 02:20:37 crc kubenswrapper[4885]: I1002 02:20:37.689793 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-2hp4t" event={"ID":"f6edca36-b630-4682-93ed-2c0c2dcadecb","Type":"ContainerStarted","Data":"17569acece96c4ce614a775ba915681df112f240dcb323f337c1e4756ef42e7e"} Oct 02 02:20:37 crc kubenswrapper[4885]: I1002 02:20:37.728378 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-2hp4t" podStartSLOduration=2.922086815 podStartE2EDuration="3.72834164s" podCreationTimestamp="2025-10-02 02:20:34 +0000 UTC" firstStartedPulling="2025-10-02 02:20:35.682816525 +0000 UTC m=+2024.494563924" lastFinishedPulling="2025-10-02 02:20:36.48907131 +0000 UTC m=+2025.300818749" observedRunningTime="2025-10-02 02:20:37.71460237 +0000 UTC m=+2026.526349799" watchObservedRunningTime="2025-10-02 02:20:37.72834164 +0000 UTC m=+2026.540089079" Oct 02 02:20:43 crc kubenswrapper[4885]: I1002 02:20:43.265634 4885 patch_prober.go:28] interesting pod/machine-config-daemon-rttx8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 02:20:43 crc kubenswrapper[4885]: I1002 02:20:43.266029 4885 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 02:20:57 crc kubenswrapper[4885]: I1002 02:20:57.822631 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-85css"] Oct 02 02:20:57 crc kubenswrapper[4885]: I1002 02:20:57.827201 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-85css" Oct 02 02:20:57 crc kubenswrapper[4885]: I1002 02:20:57.848468 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-85css"] Oct 02 02:20:57 crc kubenswrapper[4885]: I1002 02:20:57.901569 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gwvfn\" (UniqueName: \"kubernetes.io/projected/ecc07b99-f24f-408f-92fb-27e716d87e86-kube-api-access-gwvfn\") pod \"redhat-operators-85css\" (UID: \"ecc07b99-f24f-408f-92fb-27e716d87e86\") " pod="openshift-marketplace/redhat-operators-85css" Oct 02 02:20:57 crc kubenswrapper[4885]: I1002 02:20:57.901720 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ecc07b99-f24f-408f-92fb-27e716d87e86-utilities\") pod \"redhat-operators-85css\" (UID: \"ecc07b99-f24f-408f-92fb-27e716d87e86\") " pod="openshift-marketplace/redhat-operators-85css" Oct 02 02:20:57 crc kubenswrapper[4885]: I1002 02:20:57.901819 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ecc07b99-f24f-408f-92fb-27e716d87e86-catalog-content\") pod \"redhat-operators-85css\" (UID: \"ecc07b99-f24f-408f-92fb-27e716d87e86\") " pod="openshift-marketplace/redhat-operators-85css" Oct 02 02:20:58 crc kubenswrapper[4885]: I1002 02:20:58.003628 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ecc07b99-f24f-408f-92fb-27e716d87e86-catalog-content\") pod \"redhat-operators-85css\" (UID: \"ecc07b99-f24f-408f-92fb-27e716d87e86\") " pod="openshift-marketplace/redhat-operators-85css" Oct 02 02:20:58 crc kubenswrapper[4885]: I1002 02:20:58.004145 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gwvfn\" (UniqueName: \"kubernetes.io/projected/ecc07b99-f24f-408f-92fb-27e716d87e86-kube-api-access-gwvfn\") pod \"redhat-operators-85css\" (UID: \"ecc07b99-f24f-408f-92fb-27e716d87e86\") " pod="openshift-marketplace/redhat-operators-85css" Oct 02 02:20:58 crc kubenswrapper[4885]: I1002 02:20:58.004438 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ecc07b99-f24f-408f-92fb-27e716d87e86-catalog-content\") pod \"redhat-operators-85css\" (UID: \"ecc07b99-f24f-408f-92fb-27e716d87e86\") " pod="openshift-marketplace/redhat-operators-85css" Oct 02 02:20:58 crc kubenswrapper[4885]: I1002 02:20:58.004843 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ecc07b99-f24f-408f-92fb-27e716d87e86-utilities\") pod \"redhat-operators-85css\" (UID: \"ecc07b99-f24f-408f-92fb-27e716d87e86\") " pod="openshift-marketplace/redhat-operators-85css" Oct 02 02:20:58 crc kubenswrapper[4885]: I1002 02:20:58.005378 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ecc07b99-f24f-408f-92fb-27e716d87e86-utilities\") pod \"redhat-operators-85css\" (UID: \"ecc07b99-f24f-408f-92fb-27e716d87e86\") " pod="openshift-marketplace/redhat-operators-85css" Oct 02 02:20:58 crc kubenswrapper[4885]: I1002 02:20:58.026399 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gwvfn\" (UniqueName: \"kubernetes.io/projected/ecc07b99-f24f-408f-92fb-27e716d87e86-kube-api-access-gwvfn\") pod \"redhat-operators-85css\" (UID: \"ecc07b99-f24f-408f-92fb-27e716d87e86\") " pod="openshift-marketplace/redhat-operators-85css" Oct 02 02:20:58 crc kubenswrapper[4885]: I1002 02:20:58.154123 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-85css" Oct 02 02:20:58 crc kubenswrapper[4885]: I1002 02:20:58.440818 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-85css"] Oct 02 02:20:58 crc kubenswrapper[4885]: I1002 02:20:58.937084 4885 generic.go:334] "Generic (PLEG): container finished" podID="ecc07b99-f24f-408f-92fb-27e716d87e86" containerID="0c14f5e86a6bcd1e2a0cdda41a273c36c2f744fc21815a1dd968817a11ea75df" exitCode=0 Oct 02 02:20:58 crc kubenswrapper[4885]: I1002 02:20:58.937309 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-85css" event={"ID":"ecc07b99-f24f-408f-92fb-27e716d87e86","Type":"ContainerDied","Data":"0c14f5e86a6bcd1e2a0cdda41a273c36c2f744fc21815a1dd968817a11ea75df"} Oct 02 02:20:58 crc kubenswrapper[4885]: I1002 02:20:58.937384 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-85css" event={"ID":"ecc07b99-f24f-408f-92fb-27e716d87e86","Type":"ContainerStarted","Data":"ff02bcd456c780308e4bcbe3e9d7f88de26fa5b745f287b5d9c7dac8ba18a180"} Oct 02 02:20:59 crc kubenswrapper[4885]: I1002 02:20:59.949095 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-85css" event={"ID":"ecc07b99-f24f-408f-92fb-27e716d87e86","Type":"ContainerStarted","Data":"7a2013c03fdb36e9287fe58e38e71fb717b3f3a5e0759ed6ed57d678daa16b4f"} Oct 02 02:21:00 crc kubenswrapper[4885]: I1002 02:21:00.962759 4885 generic.go:334] "Generic (PLEG): container finished" podID="ecc07b99-f24f-408f-92fb-27e716d87e86" containerID="7a2013c03fdb36e9287fe58e38e71fb717b3f3a5e0759ed6ed57d678daa16b4f" exitCode=0 Oct 02 02:21:00 crc kubenswrapper[4885]: I1002 02:21:00.963007 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-85css" event={"ID":"ecc07b99-f24f-408f-92fb-27e716d87e86","Type":"ContainerDied","Data":"7a2013c03fdb36e9287fe58e38e71fb717b3f3a5e0759ed6ed57d678daa16b4f"} Oct 02 02:21:01 crc kubenswrapper[4885]: I1002 02:21:01.981005 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-85css" event={"ID":"ecc07b99-f24f-408f-92fb-27e716d87e86","Type":"ContainerStarted","Data":"94b3fff883c6598ea231209738af0b8a0412db69c875ffa71ff29aa3a32e60a1"} Oct 02 02:21:02 crc kubenswrapper[4885]: I1002 02:21:02.018638 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-85css" podStartSLOduration=2.431540354 podStartE2EDuration="5.018614665s" podCreationTimestamp="2025-10-02 02:20:57 +0000 UTC" firstStartedPulling="2025-10-02 02:20:58.938894009 +0000 UTC m=+2047.750641398" lastFinishedPulling="2025-10-02 02:21:01.52596827 +0000 UTC m=+2050.337715709" observedRunningTime="2025-10-02 02:21:02.013448067 +0000 UTC m=+2050.825195486" watchObservedRunningTime="2025-10-02 02:21:02.018614665 +0000 UTC m=+2050.830362064" Oct 02 02:21:08 crc kubenswrapper[4885]: I1002 02:21:08.154466 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-85css" Oct 02 02:21:08 crc kubenswrapper[4885]: I1002 02:21:08.155072 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-85css" Oct 02 02:21:08 crc kubenswrapper[4885]: I1002 02:21:08.215126 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-85css" Oct 02 02:21:09 crc kubenswrapper[4885]: I1002 02:21:09.123930 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-85css" Oct 02 02:21:09 crc kubenswrapper[4885]: I1002 02:21:09.190906 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-85css"] Oct 02 02:21:11 crc kubenswrapper[4885]: I1002 02:21:11.084675 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-85css" podUID="ecc07b99-f24f-408f-92fb-27e716d87e86" containerName="registry-server" containerID="cri-o://94b3fff883c6598ea231209738af0b8a0412db69c875ffa71ff29aa3a32e60a1" gracePeriod=2 Oct 02 02:21:11 crc kubenswrapper[4885]: I1002 02:21:11.649818 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-85css" Oct 02 02:21:11 crc kubenswrapper[4885]: I1002 02:21:11.708289 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gwvfn\" (UniqueName: \"kubernetes.io/projected/ecc07b99-f24f-408f-92fb-27e716d87e86-kube-api-access-gwvfn\") pod \"ecc07b99-f24f-408f-92fb-27e716d87e86\" (UID: \"ecc07b99-f24f-408f-92fb-27e716d87e86\") " Oct 02 02:21:11 crc kubenswrapper[4885]: I1002 02:21:11.708391 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ecc07b99-f24f-408f-92fb-27e716d87e86-utilities\") pod \"ecc07b99-f24f-408f-92fb-27e716d87e86\" (UID: \"ecc07b99-f24f-408f-92fb-27e716d87e86\") " Oct 02 02:21:11 crc kubenswrapper[4885]: I1002 02:21:11.708733 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ecc07b99-f24f-408f-92fb-27e716d87e86-catalog-content\") pod \"ecc07b99-f24f-408f-92fb-27e716d87e86\" (UID: \"ecc07b99-f24f-408f-92fb-27e716d87e86\") " Oct 02 02:21:11 crc kubenswrapper[4885]: I1002 02:21:11.711991 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ecc07b99-f24f-408f-92fb-27e716d87e86-utilities" (OuterVolumeSpecName: "utilities") pod "ecc07b99-f24f-408f-92fb-27e716d87e86" (UID: "ecc07b99-f24f-408f-92fb-27e716d87e86"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:21:11 crc kubenswrapper[4885]: I1002 02:21:11.715235 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ecc07b99-f24f-408f-92fb-27e716d87e86-kube-api-access-gwvfn" (OuterVolumeSpecName: "kube-api-access-gwvfn") pod "ecc07b99-f24f-408f-92fb-27e716d87e86" (UID: "ecc07b99-f24f-408f-92fb-27e716d87e86"). InnerVolumeSpecName "kube-api-access-gwvfn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:21:11 crc kubenswrapper[4885]: I1002 02:21:11.791415 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ecc07b99-f24f-408f-92fb-27e716d87e86-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ecc07b99-f24f-408f-92fb-27e716d87e86" (UID: "ecc07b99-f24f-408f-92fb-27e716d87e86"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:21:11 crc kubenswrapper[4885]: I1002 02:21:11.811229 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gwvfn\" (UniqueName: \"kubernetes.io/projected/ecc07b99-f24f-408f-92fb-27e716d87e86-kube-api-access-gwvfn\") on node \"crc\" DevicePath \"\"" Oct 02 02:21:11 crc kubenswrapper[4885]: I1002 02:21:11.811286 4885 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ecc07b99-f24f-408f-92fb-27e716d87e86-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 02:21:11 crc kubenswrapper[4885]: I1002 02:21:11.811297 4885 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ecc07b99-f24f-408f-92fb-27e716d87e86-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 02:21:12 crc kubenswrapper[4885]: I1002 02:21:12.099846 4885 generic.go:334] "Generic (PLEG): container finished" podID="ecc07b99-f24f-408f-92fb-27e716d87e86" containerID="94b3fff883c6598ea231209738af0b8a0412db69c875ffa71ff29aa3a32e60a1" exitCode=0 Oct 02 02:21:12 crc kubenswrapper[4885]: I1002 02:21:12.099914 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-85css" event={"ID":"ecc07b99-f24f-408f-92fb-27e716d87e86","Type":"ContainerDied","Data":"94b3fff883c6598ea231209738af0b8a0412db69c875ffa71ff29aa3a32e60a1"} Oct 02 02:21:12 crc kubenswrapper[4885]: I1002 02:21:12.099920 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-85css" Oct 02 02:21:12 crc kubenswrapper[4885]: I1002 02:21:12.099970 4885 scope.go:117] "RemoveContainer" containerID="94b3fff883c6598ea231209738af0b8a0412db69c875ffa71ff29aa3a32e60a1" Oct 02 02:21:12 crc kubenswrapper[4885]: I1002 02:21:12.099955 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-85css" event={"ID":"ecc07b99-f24f-408f-92fb-27e716d87e86","Type":"ContainerDied","Data":"ff02bcd456c780308e4bcbe3e9d7f88de26fa5b745f287b5d9c7dac8ba18a180"} Oct 02 02:21:12 crc kubenswrapper[4885]: I1002 02:21:12.139608 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-85css"] Oct 02 02:21:12 crc kubenswrapper[4885]: I1002 02:21:12.151362 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-85css"] Oct 02 02:21:12 crc kubenswrapper[4885]: I1002 02:21:12.165888 4885 scope.go:117] "RemoveContainer" containerID="7a2013c03fdb36e9287fe58e38e71fb717b3f3a5e0759ed6ed57d678daa16b4f" Oct 02 02:21:12 crc kubenswrapper[4885]: I1002 02:21:12.205716 4885 scope.go:117] "RemoveContainer" containerID="0c14f5e86a6bcd1e2a0cdda41a273c36c2f744fc21815a1dd968817a11ea75df" Oct 02 02:21:12 crc kubenswrapper[4885]: I1002 02:21:12.249303 4885 scope.go:117] "RemoveContainer" containerID="94b3fff883c6598ea231209738af0b8a0412db69c875ffa71ff29aa3a32e60a1" Oct 02 02:21:12 crc kubenswrapper[4885]: E1002 02:21:12.250048 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"94b3fff883c6598ea231209738af0b8a0412db69c875ffa71ff29aa3a32e60a1\": container with ID starting with 94b3fff883c6598ea231209738af0b8a0412db69c875ffa71ff29aa3a32e60a1 not found: ID does not exist" containerID="94b3fff883c6598ea231209738af0b8a0412db69c875ffa71ff29aa3a32e60a1" Oct 02 02:21:12 crc kubenswrapper[4885]: I1002 02:21:12.250130 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"94b3fff883c6598ea231209738af0b8a0412db69c875ffa71ff29aa3a32e60a1"} err="failed to get container status \"94b3fff883c6598ea231209738af0b8a0412db69c875ffa71ff29aa3a32e60a1\": rpc error: code = NotFound desc = could not find container \"94b3fff883c6598ea231209738af0b8a0412db69c875ffa71ff29aa3a32e60a1\": container with ID starting with 94b3fff883c6598ea231209738af0b8a0412db69c875ffa71ff29aa3a32e60a1 not found: ID does not exist" Oct 02 02:21:12 crc kubenswrapper[4885]: I1002 02:21:12.250178 4885 scope.go:117] "RemoveContainer" containerID="7a2013c03fdb36e9287fe58e38e71fb717b3f3a5e0759ed6ed57d678daa16b4f" Oct 02 02:21:12 crc kubenswrapper[4885]: E1002 02:21:12.250627 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7a2013c03fdb36e9287fe58e38e71fb717b3f3a5e0759ed6ed57d678daa16b4f\": container with ID starting with 7a2013c03fdb36e9287fe58e38e71fb717b3f3a5e0759ed6ed57d678daa16b4f not found: ID does not exist" containerID="7a2013c03fdb36e9287fe58e38e71fb717b3f3a5e0759ed6ed57d678daa16b4f" Oct 02 02:21:12 crc kubenswrapper[4885]: I1002 02:21:12.250673 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a2013c03fdb36e9287fe58e38e71fb717b3f3a5e0759ed6ed57d678daa16b4f"} err="failed to get container status \"7a2013c03fdb36e9287fe58e38e71fb717b3f3a5e0759ed6ed57d678daa16b4f\": rpc error: code = NotFound desc = could not find container \"7a2013c03fdb36e9287fe58e38e71fb717b3f3a5e0759ed6ed57d678daa16b4f\": container with ID starting with 7a2013c03fdb36e9287fe58e38e71fb717b3f3a5e0759ed6ed57d678daa16b4f not found: ID does not exist" Oct 02 02:21:12 crc kubenswrapper[4885]: I1002 02:21:12.250703 4885 scope.go:117] "RemoveContainer" containerID="0c14f5e86a6bcd1e2a0cdda41a273c36c2f744fc21815a1dd968817a11ea75df" Oct 02 02:21:12 crc kubenswrapper[4885]: E1002 02:21:12.251073 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0c14f5e86a6bcd1e2a0cdda41a273c36c2f744fc21815a1dd968817a11ea75df\": container with ID starting with 0c14f5e86a6bcd1e2a0cdda41a273c36c2f744fc21815a1dd968817a11ea75df not found: ID does not exist" containerID="0c14f5e86a6bcd1e2a0cdda41a273c36c2f744fc21815a1dd968817a11ea75df" Oct 02 02:21:12 crc kubenswrapper[4885]: I1002 02:21:12.251118 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c14f5e86a6bcd1e2a0cdda41a273c36c2f744fc21815a1dd968817a11ea75df"} err="failed to get container status \"0c14f5e86a6bcd1e2a0cdda41a273c36c2f744fc21815a1dd968817a11ea75df\": rpc error: code = NotFound desc = could not find container \"0c14f5e86a6bcd1e2a0cdda41a273c36c2f744fc21815a1dd968817a11ea75df\": container with ID starting with 0c14f5e86a6bcd1e2a0cdda41a273c36c2f744fc21815a1dd968817a11ea75df not found: ID does not exist" Oct 02 02:21:13 crc kubenswrapper[4885]: I1002 02:21:13.266147 4885 patch_prober.go:28] interesting pod/machine-config-daemon-rttx8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 02:21:13 crc kubenswrapper[4885]: I1002 02:21:13.266240 4885 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 02:21:13 crc kubenswrapper[4885]: I1002 02:21:13.266384 4885 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" Oct 02 02:21:13 crc kubenswrapper[4885]: I1002 02:21:13.267183 4885 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"562009e16fbeefe72a46fc9d563e2da31ec7f209f7fd3da12e264d10773a1c88"} pod="openshift-machine-config-operator/machine-config-daemon-rttx8" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 02:21:13 crc kubenswrapper[4885]: I1002 02:21:13.267313 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" containerName="machine-config-daemon" containerID="cri-o://562009e16fbeefe72a46fc9d563e2da31ec7f209f7fd3da12e264d10773a1c88" gracePeriod=600 Oct 02 02:21:14 crc kubenswrapper[4885]: I1002 02:21:14.059730 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ecc07b99-f24f-408f-92fb-27e716d87e86" path="/var/lib/kubelet/pods/ecc07b99-f24f-408f-92fb-27e716d87e86/volumes" Oct 02 02:21:14 crc kubenswrapper[4885]: I1002 02:21:14.121383 4885 generic.go:334] "Generic (PLEG): container finished" podID="d169c8ba-000a-4218-8e3b-4ae53035b110" containerID="562009e16fbeefe72a46fc9d563e2da31ec7f209f7fd3da12e264d10773a1c88" exitCode=0 Oct 02 02:21:14 crc kubenswrapper[4885]: I1002 02:21:14.121426 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" event={"ID":"d169c8ba-000a-4218-8e3b-4ae53035b110","Type":"ContainerDied","Data":"562009e16fbeefe72a46fc9d563e2da31ec7f209f7fd3da12e264d10773a1c88"} Oct 02 02:21:14 crc kubenswrapper[4885]: I1002 02:21:14.121456 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" event={"ID":"d169c8ba-000a-4218-8e3b-4ae53035b110","Type":"ContainerStarted","Data":"7d1be267228233dc9433f75e76a5a805aa23d98aa3b0edbd385a1a565a0d3231"} Oct 02 02:21:14 crc kubenswrapper[4885]: I1002 02:21:14.121473 4885 scope.go:117] "RemoveContainer" containerID="be0fc35a7f63d87317c3e5145a3df69700656f283328299a3b8545263cc22494" Oct 02 02:21:30 crc kubenswrapper[4885]: I1002 02:21:30.309320 4885 generic.go:334] "Generic (PLEG): container finished" podID="f6edca36-b630-4682-93ed-2c0c2dcadecb" containerID="17569acece96c4ce614a775ba915681df112f240dcb323f337c1e4756ef42e7e" exitCode=0 Oct 02 02:21:30 crc kubenswrapper[4885]: I1002 02:21:30.309426 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-2hp4t" event={"ID":"f6edca36-b630-4682-93ed-2c0c2dcadecb","Type":"ContainerDied","Data":"17569acece96c4ce614a775ba915681df112f240dcb323f337c1e4756ef42e7e"} Oct 02 02:21:31 crc kubenswrapper[4885]: I1002 02:21:31.787785 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-2hp4t" Oct 02 02:21:31 crc kubenswrapper[4885]: I1002 02:21:31.880665 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k5bj9\" (UniqueName: \"kubernetes.io/projected/f6edca36-b630-4682-93ed-2c0c2dcadecb-kube-api-access-k5bj9\") pod \"f6edca36-b630-4682-93ed-2c0c2dcadecb\" (UID: \"f6edca36-b630-4682-93ed-2c0c2dcadecb\") " Oct 02 02:21:31 crc kubenswrapper[4885]: I1002 02:21:31.881080 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6edca36-b630-4682-93ed-2c0c2dcadecb-neutron-metadata-combined-ca-bundle\") pod \"f6edca36-b630-4682-93ed-2c0c2dcadecb\" (UID: \"f6edca36-b630-4682-93ed-2c0c2dcadecb\") " Oct 02 02:21:31 crc kubenswrapper[4885]: I1002 02:21:31.881139 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f6edca36-b630-4682-93ed-2c0c2dcadecb-ssh-key\") pod \"f6edca36-b630-4682-93ed-2c0c2dcadecb\" (UID: \"f6edca36-b630-4682-93ed-2c0c2dcadecb\") " Oct 02 02:21:31 crc kubenswrapper[4885]: I1002 02:21:31.881215 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f6edca36-b630-4682-93ed-2c0c2dcadecb-inventory\") pod \"f6edca36-b630-4682-93ed-2c0c2dcadecb\" (UID: \"f6edca36-b630-4682-93ed-2c0c2dcadecb\") " Oct 02 02:21:31 crc kubenswrapper[4885]: I1002 02:21:31.881241 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/f6edca36-b630-4682-93ed-2c0c2dcadecb-neutron-ovn-metadata-agent-neutron-config-0\") pod \"f6edca36-b630-4682-93ed-2c0c2dcadecb\" (UID: \"f6edca36-b630-4682-93ed-2c0c2dcadecb\") " Oct 02 02:21:31 crc kubenswrapper[4885]: I1002 02:21:31.881321 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/f6edca36-b630-4682-93ed-2c0c2dcadecb-nova-metadata-neutron-config-0\") pod \"f6edca36-b630-4682-93ed-2c0c2dcadecb\" (UID: \"f6edca36-b630-4682-93ed-2c0c2dcadecb\") " Oct 02 02:21:31 crc kubenswrapper[4885]: I1002 02:21:31.892976 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6edca36-b630-4682-93ed-2c0c2dcadecb-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "f6edca36-b630-4682-93ed-2c0c2dcadecb" (UID: "f6edca36-b630-4682-93ed-2c0c2dcadecb"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:21:31 crc kubenswrapper[4885]: I1002 02:21:31.893388 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6edca36-b630-4682-93ed-2c0c2dcadecb-kube-api-access-k5bj9" (OuterVolumeSpecName: "kube-api-access-k5bj9") pod "f6edca36-b630-4682-93ed-2c0c2dcadecb" (UID: "f6edca36-b630-4682-93ed-2c0c2dcadecb"). InnerVolumeSpecName "kube-api-access-k5bj9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:21:31 crc kubenswrapper[4885]: I1002 02:21:31.909405 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6edca36-b630-4682-93ed-2c0c2dcadecb-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f6edca36-b630-4682-93ed-2c0c2dcadecb" (UID: "f6edca36-b630-4682-93ed-2c0c2dcadecb"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:21:31 crc kubenswrapper[4885]: I1002 02:21:31.911899 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6edca36-b630-4682-93ed-2c0c2dcadecb-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "f6edca36-b630-4682-93ed-2c0c2dcadecb" (UID: "f6edca36-b630-4682-93ed-2c0c2dcadecb"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:21:31 crc kubenswrapper[4885]: I1002 02:21:31.921352 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6edca36-b630-4682-93ed-2c0c2dcadecb-inventory" (OuterVolumeSpecName: "inventory") pod "f6edca36-b630-4682-93ed-2c0c2dcadecb" (UID: "f6edca36-b630-4682-93ed-2c0c2dcadecb"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:21:31 crc kubenswrapper[4885]: I1002 02:21:31.932033 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6edca36-b630-4682-93ed-2c0c2dcadecb-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "f6edca36-b630-4682-93ed-2c0c2dcadecb" (UID: "f6edca36-b630-4682-93ed-2c0c2dcadecb"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:21:31 crc kubenswrapper[4885]: I1002 02:21:31.984091 4885 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f6edca36-b630-4682-93ed-2c0c2dcadecb-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 02:21:31 crc kubenswrapper[4885]: I1002 02:21:31.984128 4885 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/f6edca36-b630-4682-93ed-2c0c2dcadecb-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 02 02:21:31 crc kubenswrapper[4885]: I1002 02:21:31.984144 4885 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/f6edca36-b630-4682-93ed-2c0c2dcadecb-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 02 02:21:31 crc kubenswrapper[4885]: I1002 02:21:31.984157 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k5bj9\" (UniqueName: \"kubernetes.io/projected/f6edca36-b630-4682-93ed-2c0c2dcadecb-kube-api-access-k5bj9\") on node \"crc\" DevicePath \"\"" Oct 02 02:21:31 crc kubenswrapper[4885]: I1002 02:21:31.984171 4885 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6edca36-b630-4682-93ed-2c0c2dcadecb-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:21:31 crc kubenswrapper[4885]: I1002 02:21:31.984183 4885 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f6edca36-b630-4682-93ed-2c0c2dcadecb-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 02:21:32 crc kubenswrapper[4885]: I1002 02:21:32.331821 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-2hp4t" event={"ID":"f6edca36-b630-4682-93ed-2c0c2dcadecb","Type":"ContainerDied","Data":"71af0ebf476ced4ef2578ffe478331e575a27eb964f3d2cb9f2f995871786bcb"} Oct 02 02:21:32 crc kubenswrapper[4885]: I1002 02:21:32.331894 4885 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="71af0ebf476ced4ef2578ffe478331e575a27eb964f3d2cb9f2f995871786bcb" Oct 02 02:21:32 crc kubenswrapper[4885]: I1002 02:21:32.331955 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-2hp4t" Oct 02 02:21:32 crc kubenswrapper[4885]: I1002 02:21:32.481338 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-x47dm"] Oct 02 02:21:32 crc kubenswrapper[4885]: E1002 02:21:32.481997 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ecc07b99-f24f-408f-92fb-27e716d87e86" containerName="extract-utilities" Oct 02 02:21:32 crc kubenswrapper[4885]: I1002 02:21:32.482031 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="ecc07b99-f24f-408f-92fb-27e716d87e86" containerName="extract-utilities" Oct 02 02:21:32 crc kubenswrapper[4885]: E1002 02:21:32.482061 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6edca36-b630-4682-93ed-2c0c2dcadecb" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Oct 02 02:21:32 crc kubenswrapper[4885]: I1002 02:21:32.482079 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6edca36-b630-4682-93ed-2c0c2dcadecb" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Oct 02 02:21:32 crc kubenswrapper[4885]: E1002 02:21:32.482110 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ecc07b99-f24f-408f-92fb-27e716d87e86" containerName="extract-content" Oct 02 02:21:32 crc kubenswrapper[4885]: I1002 02:21:32.482123 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="ecc07b99-f24f-408f-92fb-27e716d87e86" containerName="extract-content" Oct 02 02:21:32 crc kubenswrapper[4885]: E1002 02:21:32.482151 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ecc07b99-f24f-408f-92fb-27e716d87e86" containerName="registry-server" Oct 02 02:21:32 crc kubenswrapper[4885]: I1002 02:21:32.482166 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="ecc07b99-f24f-408f-92fb-27e716d87e86" containerName="registry-server" Oct 02 02:21:32 crc kubenswrapper[4885]: I1002 02:21:32.482523 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="ecc07b99-f24f-408f-92fb-27e716d87e86" containerName="registry-server" Oct 02 02:21:32 crc kubenswrapper[4885]: I1002 02:21:32.482550 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6edca36-b630-4682-93ed-2c0c2dcadecb" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Oct 02 02:21:32 crc kubenswrapper[4885]: I1002 02:21:32.483663 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-x47dm" Oct 02 02:21:32 crc kubenswrapper[4885]: I1002 02:21:32.487155 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-f8tvb" Oct 02 02:21:32 crc kubenswrapper[4885]: I1002 02:21:32.487610 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 02:21:32 crc kubenswrapper[4885]: I1002 02:21:32.488080 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 02:21:32 crc kubenswrapper[4885]: I1002 02:21:32.488428 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 02:21:32 crc kubenswrapper[4885]: I1002 02:21:32.488790 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Oct 02 02:21:32 crc kubenswrapper[4885]: I1002 02:21:32.494850 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-x47dm"] Oct 02 02:21:32 crc kubenswrapper[4885]: I1002 02:21:32.597581 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54cf89d0-a777-4ac9-968b-49836d0b13b1-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-x47dm\" (UID: \"54cf89d0-a777-4ac9-968b-49836d0b13b1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-x47dm" Oct 02 02:21:32 crc kubenswrapper[4885]: I1002 02:21:32.597658 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/54cf89d0-a777-4ac9-968b-49836d0b13b1-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-x47dm\" (UID: \"54cf89d0-a777-4ac9-968b-49836d0b13b1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-x47dm" Oct 02 02:21:32 crc kubenswrapper[4885]: I1002 02:21:32.597955 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/54cf89d0-a777-4ac9-968b-49836d0b13b1-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-x47dm\" (UID: \"54cf89d0-a777-4ac9-968b-49836d0b13b1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-x47dm" Oct 02 02:21:32 crc kubenswrapper[4885]: I1002 02:21:32.598093 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/54cf89d0-a777-4ac9-968b-49836d0b13b1-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-x47dm\" (UID: \"54cf89d0-a777-4ac9-968b-49836d0b13b1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-x47dm" Oct 02 02:21:32 crc kubenswrapper[4885]: I1002 02:21:32.598154 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ph8fc\" (UniqueName: \"kubernetes.io/projected/54cf89d0-a777-4ac9-968b-49836d0b13b1-kube-api-access-ph8fc\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-x47dm\" (UID: \"54cf89d0-a777-4ac9-968b-49836d0b13b1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-x47dm" Oct 02 02:21:32 crc kubenswrapper[4885]: I1002 02:21:32.700383 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/54cf89d0-a777-4ac9-968b-49836d0b13b1-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-x47dm\" (UID: \"54cf89d0-a777-4ac9-968b-49836d0b13b1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-x47dm" Oct 02 02:21:32 crc kubenswrapper[4885]: I1002 02:21:32.700734 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/54cf89d0-a777-4ac9-968b-49836d0b13b1-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-x47dm\" (UID: \"54cf89d0-a777-4ac9-968b-49836d0b13b1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-x47dm" Oct 02 02:21:32 crc kubenswrapper[4885]: I1002 02:21:32.700785 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/54cf89d0-a777-4ac9-968b-49836d0b13b1-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-x47dm\" (UID: \"54cf89d0-a777-4ac9-968b-49836d0b13b1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-x47dm" Oct 02 02:21:32 crc kubenswrapper[4885]: I1002 02:21:32.700816 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ph8fc\" (UniqueName: \"kubernetes.io/projected/54cf89d0-a777-4ac9-968b-49836d0b13b1-kube-api-access-ph8fc\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-x47dm\" (UID: \"54cf89d0-a777-4ac9-968b-49836d0b13b1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-x47dm" Oct 02 02:21:32 crc kubenswrapper[4885]: I1002 02:21:32.700910 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54cf89d0-a777-4ac9-968b-49836d0b13b1-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-x47dm\" (UID: \"54cf89d0-a777-4ac9-968b-49836d0b13b1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-x47dm" Oct 02 02:21:32 crc kubenswrapper[4885]: I1002 02:21:32.710129 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/54cf89d0-a777-4ac9-968b-49836d0b13b1-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-x47dm\" (UID: \"54cf89d0-a777-4ac9-968b-49836d0b13b1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-x47dm" Oct 02 02:21:32 crc kubenswrapper[4885]: I1002 02:21:32.710438 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54cf89d0-a777-4ac9-968b-49836d0b13b1-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-x47dm\" (UID: \"54cf89d0-a777-4ac9-968b-49836d0b13b1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-x47dm" Oct 02 02:21:32 crc kubenswrapper[4885]: I1002 02:21:32.711125 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/54cf89d0-a777-4ac9-968b-49836d0b13b1-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-x47dm\" (UID: \"54cf89d0-a777-4ac9-968b-49836d0b13b1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-x47dm" Oct 02 02:21:32 crc kubenswrapper[4885]: I1002 02:21:32.711184 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/54cf89d0-a777-4ac9-968b-49836d0b13b1-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-x47dm\" (UID: \"54cf89d0-a777-4ac9-968b-49836d0b13b1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-x47dm" Oct 02 02:21:32 crc kubenswrapper[4885]: I1002 02:21:32.720714 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ph8fc\" (UniqueName: \"kubernetes.io/projected/54cf89d0-a777-4ac9-968b-49836d0b13b1-kube-api-access-ph8fc\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-x47dm\" (UID: \"54cf89d0-a777-4ac9-968b-49836d0b13b1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-x47dm" Oct 02 02:21:32 crc kubenswrapper[4885]: I1002 02:21:32.809727 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-x47dm" Oct 02 02:21:33 crc kubenswrapper[4885]: I1002 02:21:33.351491 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-x47dm"] Oct 02 02:21:34 crc kubenswrapper[4885]: I1002 02:21:34.363256 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-x47dm" event={"ID":"54cf89d0-a777-4ac9-968b-49836d0b13b1","Type":"ContainerStarted","Data":"62a6c50591117ca89f8931562ac58e8042be832ab8a0c335ce598b0b7c4d4c2e"} Oct 02 02:21:34 crc kubenswrapper[4885]: I1002 02:21:34.364128 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-x47dm" event={"ID":"54cf89d0-a777-4ac9-968b-49836d0b13b1","Type":"ContainerStarted","Data":"978e2d4949a047641300bba9c41a77527e0736cefc5bb06ca536910885095780"} Oct 02 02:21:34 crc kubenswrapper[4885]: I1002 02:21:34.393729 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-x47dm" podStartSLOduration=1.712764857 podStartE2EDuration="2.393702197s" podCreationTimestamp="2025-10-02 02:21:32 +0000 UTC" firstStartedPulling="2025-10-02 02:21:33.364623553 +0000 UTC m=+2082.176370952" lastFinishedPulling="2025-10-02 02:21:34.045560853 +0000 UTC m=+2082.857308292" observedRunningTime="2025-10-02 02:21:34.385003252 +0000 UTC m=+2083.196750741" watchObservedRunningTime="2025-10-02 02:21:34.393702197 +0000 UTC m=+2083.205449626" Oct 02 02:22:59 crc kubenswrapper[4885]: I1002 02:22:59.259368 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-kl8hf"] Oct 02 02:22:59 crc kubenswrapper[4885]: I1002 02:22:59.262708 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kl8hf" Oct 02 02:22:59 crc kubenswrapper[4885]: I1002 02:22:59.277020 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-kl8hf"] Oct 02 02:22:59 crc kubenswrapper[4885]: I1002 02:22:59.347879 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wplzh\" (UniqueName: \"kubernetes.io/projected/0c38718b-d4ec-492a-a183-cee6b6a59bfe-kube-api-access-wplzh\") pod \"community-operators-kl8hf\" (UID: \"0c38718b-d4ec-492a-a183-cee6b6a59bfe\") " pod="openshift-marketplace/community-operators-kl8hf" Oct 02 02:22:59 crc kubenswrapper[4885]: I1002 02:22:59.347946 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c38718b-d4ec-492a-a183-cee6b6a59bfe-utilities\") pod \"community-operators-kl8hf\" (UID: \"0c38718b-d4ec-492a-a183-cee6b6a59bfe\") " pod="openshift-marketplace/community-operators-kl8hf" Oct 02 02:22:59 crc kubenswrapper[4885]: I1002 02:22:59.348083 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c38718b-d4ec-492a-a183-cee6b6a59bfe-catalog-content\") pod \"community-operators-kl8hf\" (UID: \"0c38718b-d4ec-492a-a183-cee6b6a59bfe\") " pod="openshift-marketplace/community-operators-kl8hf" Oct 02 02:22:59 crc kubenswrapper[4885]: I1002 02:22:59.450248 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wplzh\" (UniqueName: \"kubernetes.io/projected/0c38718b-d4ec-492a-a183-cee6b6a59bfe-kube-api-access-wplzh\") pod \"community-operators-kl8hf\" (UID: \"0c38718b-d4ec-492a-a183-cee6b6a59bfe\") " pod="openshift-marketplace/community-operators-kl8hf" Oct 02 02:22:59 crc kubenswrapper[4885]: I1002 02:22:59.450352 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c38718b-d4ec-492a-a183-cee6b6a59bfe-utilities\") pod \"community-operators-kl8hf\" (UID: \"0c38718b-d4ec-492a-a183-cee6b6a59bfe\") " pod="openshift-marketplace/community-operators-kl8hf" Oct 02 02:22:59 crc kubenswrapper[4885]: I1002 02:22:59.450420 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c38718b-d4ec-492a-a183-cee6b6a59bfe-catalog-content\") pod \"community-operators-kl8hf\" (UID: \"0c38718b-d4ec-492a-a183-cee6b6a59bfe\") " pod="openshift-marketplace/community-operators-kl8hf" Oct 02 02:22:59 crc kubenswrapper[4885]: I1002 02:22:59.451045 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c38718b-d4ec-492a-a183-cee6b6a59bfe-utilities\") pod \"community-operators-kl8hf\" (UID: \"0c38718b-d4ec-492a-a183-cee6b6a59bfe\") " pod="openshift-marketplace/community-operators-kl8hf" Oct 02 02:22:59 crc kubenswrapper[4885]: I1002 02:22:59.451152 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c38718b-d4ec-492a-a183-cee6b6a59bfe-catalog-content\") pod \"community-operators-kl8hf\" (UID: \"0c38718b-d4ec-492a-a183-cee6b6a59bfe\") " pod="openshift-marketplace/community-operators-kl8hf" Oct 02 02:22:59 crc kubenswrapper[4885]: I1002 02:22:59.491760 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wplzh\" (UniqueName: \"kubernetes.io/projected/0c38718b-d4ec-492a-a183-cee6b6a59bfe-kube-api-access-wplzh\") pod \"community-operators-kl8hf\" (UID: \"0c38718b-d4ec-492a-a183-cee6b6a59bfe\") " pod="openshift-marketplace/community-operators-kl8hf" Oct 02 02:22:59 crc kubenswrapper[4885]: I1002 02:22:59.592095 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kl8hf" Oct 02 02:23:00 crc kubenswrapper[4885]: I1002 02:23:00.186653 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-kl8hf"] Oct 02 02:23:00 crc kubenswrapper[4885]: I1002 02:23:00.360775 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kl8hf" event={"ID":"0c38718b-d4ec-492a-a183-cee6b6a59bfe","Type":"ContainerStarted","Data":"83af80a614ce5cc30e9a62cd7ac9b2e7520b39f77d98897b7956bbb7b7121f33"} Oct 02 02:23:01 crc kubenswrapper[4885]: I1002 02:23:01.370521 4885 generic.go:334] "Generic (PLEG): container finished" podID="0c38718b-d4ec-492a-a183-cee6b6a59bfe" containerID="cd0c01b30fba66244b3eb5d1dacbf9265556037f274e42673d55aa7dfc1c1fac" exitCode=0 Oct 02 02:23:01 crc kubenswrapper[4885]: I1002 02:23:01.370613 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kl8hf" event={"ID":"0c38718b-d4ec-492a-a183-cee6b6a59bfe","Type":"ContainerDied","Data":"cd0c01b30fba66244b3eb5d1dacbf9265556037f274e42673d55aa7dfc1c1fac"} Oct 02 02:23:01 crc kubenswrapper[4885]: I1002 02:23:01.372637 4885 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 02:23:03 crc kubenswrapper[4885]: I1002 02:23:03.397552 4885 generic.go:334] "Generic (PLEG): container finished" podID="0c38718b-d4ec-492a-a183-cee6b6a59bfe" containerID="93fb0d7f421ea046adc4dcbdc38249936364a1e3cffa2f77e6653eab28639894" exitCode=0 Oct 02 02:23:03 crc kubenswrapper[4885]: I1002 02:23:03.397611 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kl8hf" event={"ID":"0c38718b-d4ec-492a-a183-cee6b6a59bfe","Type":"ContainerDied","Data":"93fb0d7f421ea046adc4dcbdc38249936364a1e3cffa2f77e6653eab28639894"} Oct 02 02:23:04 crc kubenswrapper[4885]: I1002 02:23:04.418236 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kl8hf" event={"ID":"0c38718b-d4ec-492a-a183-cee6b6a59bfe","Type":"ContainerStarted","Data":"71209c98e0e23105eceaa1c3493f2487e439b0d75b5e2dd33223be72174113bc"} Oct 02 02:23:04 crc kubenswrapper[4885]: I1002 02:23:04.445482 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-kl8hf" podStartSLOduration=2.9844331889999998 podStartE2EDuration="5.445454715s" podCreationTimestamp="2025-10-02 02:22:59 +0000 UTC" firstStartedPulling="2025-10-02 02:23:01.372189765 +0000 UTC m=+2170.183937174" lastFinishedPulling="2025-10-02 02:23:03.833211291 +0000 UTC m=+2172.644958700" observedRunningTime="2025-10-02 02:23:04.442790515 +0000 UTC m=+2173.254537954" watchObservedRunningTime="2025-10-02 02:23:04.445454715 +0000 UTC m=+2173.257202144" Oct 02 02:23:09 crc kubenswrapper[4885]: I1002 02:23:09.592405 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-kl8hf" Oct 02 02:23:09 crc kubenswrapper[4885]: I1002 02:23:09.593351 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-kl8hf" Oct 02 02:23:09 crc kubenswrapper[4885]: I1002 02:23:09.685850 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-kl8hf" Oct 02 02:23:10 crc kubenswrapper[4885]: I1002 02:23:10.565803 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-kl8hf" Oct 02 02:23:10 crc kubenswrapper[4885]: I1002 02:23:10.623644 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-kl8hf"] Oct 02 02:23:12 crc kubenswrapper[4885]: I1002 02:23:12.511115 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-kl8hf" podUID="0c38718b-d4ec-492a-a183-cee6b6a59bfe" containerName="registry-server" containerID="cri-o://71209c98e0e23105eceaa1c3493f2487e439b0d75b5e2dd33223be72174113bc" gracePeriod=2 Oct 02 02:23:13 crc kubenswrapper[4885]: I1002 02:23:13.036938 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kl8hf" Oct 02 02:23:13 crc kubenswrapper[4885]: I1002 02:23:13.170682 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c38718b-d4ec-492a-a183-cee6b6a59bfe-utilities\") pod \"0c38718b-d4ec-492a-a183-cee6b6a59bfe\" (UID: \"0c38718b-d4ec-492a-a183-cee6b6a59bfe\") " Oct 02 02:23:13 crc kubenswrapper[4885]: I1002 02:23:13.170968 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c38718b-d4ec-492a-a183-cee6b6a59bfe-catalog-content\") pod \"0c38718b-d4ec-492a-a183-cee6b6a59bfe\" (UID: \"0c38718b-d4ec-492a-a183-cee6b6a59bfe\") " Oct 02 02:23:13 crc kubenswrapper[4885]: I1002 02:23:13.171057 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wplzh\" (UniqueName: \"kubernetes.io/projected/0c38718b-d4ec-492a-a183-cee6b6a59bfe-kube-api-access-wplzh\") pod \"0c38718b-d4ec-492a-a183-cee6b6a59bfe\" (UID: \"0c38718b-d4ec-492a-a183-cee6b6a59bfe\") " Oct 02 02:23:13 crc kubenswrapper[4885]: I1002 02:23:13.172579 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c38718b-d4ec-492a-a183-cee6b6a59bfe-utilities" (OuterVolumeSpecName: "utilities") pod "0c38718b-d4ec-492a-a183-cee6b6a59bfe" (UID: "0c38718b-d4ec-492a-a183-cee6b6a59bfe"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:23:13 crc kubenswrapper[4885]: I1002 02:23:13.178773 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c38718b-d4ec-492a-a183-cee6b6a59bfe-kube-api-access-wplzh" (OuterVolumeSpecName: "kube-api-access-wplzh") pod "0c38718b-d4ec-492a-a183-cee6b6a59bfe" (UID: "0c38718b-d4ec-492a-a183-cee6b6a59bfe"). InnerVolumeSpecName "kube-api-access-wplzh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:23:13 crc kubenswrapper[4885]: I1002 02:23:13.265602 4885 patch_prober.go:28] interesting pod/machine-config-daemon-rttx8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 02:23:13 crc kubenswrapper[4885]: I1002 02:23:13.265664 4885 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 02:23:13 crc kubenswrapper[4885]: I1002 02:23:13.272793 4885 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c38718b-d4ec-492a-a183-cee6b6a59bfe-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 02:23:13 crc kubenswrapper[4885]: I1002 02:23:13.272821 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wplzh\" (UniqueName: \"kubernetes.io/projected/0c38718b-d4ec-492a-a183-cee6b6a59bfe-kube-api-access-wplzh\") on node \"crc\" DevicePath \"\"" Oct 02 02:23:13 crc kubenswrapper[4885]: I1002 02:23:13.523641 4885 generic.go:334] "Generic (PLEG): container finished" podID="0c38718b-d4ec-492a-a183-cee6b6a59bfe" containerID="71209c98e0e23105eceaa1c3493f2487e439b0d75b5e2dd33223be72174113bc" exitCode=0 Oct 02 02:23:13 crc kubenswrapper[4885]: I1002 02:23:13.523690 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kl8hf" event={"ID":"0c38718b-d4ec-492a-a183-cee6b6a59bfe","Type":"ContainerDied","Data":"71209c98e0e23105eceaa1c3493f2487e439b0d75b5e2dd33223be72174113bc"} Oct 02 02:23:13 crc kubenswrapper[4885]: I1002 02:23:13.523720 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kl8hf" event={"ID":"0c38718b-d4ec-492a-a183-cee6b6a59bfe","Type":"ContainerDied","Data":"83af80a614ce5cc30e9a62cd7ac9b2e7520b39f77d98897b7956bbb7b7121f33"} Oct 02 02:23:13 crc kubenswrapper[4885]: I1002 02:23:13.523731 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kl8hf" Oct 02 02:23:13 crc kubenswrapper[4885]: I1002 02:23:13.523746 4885 scope.go:117] "RemoveContainer" containerID="71209c98e0e23105eceaa1c3493f2487e439b0d75b5e2dd33223be72174113bc" Oct 02 02:23:13 crc kubenswrapper[4885]: I1002 02:23:13.558846 4885 scope.go:117] "RemoveContainer" containerID="93fb0d7f421ea046adc4dcbdc38249936364a1e3cffa2f77e6653eab28639894" Oct 02 02:23:13 crc kubenswrapper[4885]: I1002 02:23:13.601228 4885 scope.go:117] "RemoveContainer" containerID="cd0c01b30fba66244b3eb5d1dacbf9265556037f274e42673d55aa7dfc1c1fac" Oct 02 02:23:13 crc kubenswrapper[4885]: I1002 02:23:13.654658 4885 scope.go:117] "RemoveContainer" containerID="71209c98e0e23105eceaa1c3493f2487e439b0d75b5e2dd33223be72174113bc" Oct 02 02:23:13 crc kubenswrapper[4885]: E1002 02:23:13.655293 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"71209c98e0e23105eceaa1c3493f2487e439b0d75b5e2dd33223be72174113bc\": container with ID starting with 71209c98e0e23105eceaa1c3493f2487e439b0d75b5e2dd33223be72174113bc not found: ID does not exist" containerID="71209c98e0e23105eceaa1c3493f2487e439b0d75b5e2dd33223be72174113bc" Oct 02 02:23:13 crc kubenswrapper[4885]: I1002 02:23:13.655373 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"71209c98e0e23105eceaa1c3493f2487e439b0d75b5e2dd33223be72174113bc"} err="failed to get container status \"71209c98e0e23105eceaa1c3493f2487e439b0d75b5e2dd33223be72174113bc\": rpc error: code = NotFound desc = could not find container \"71209c98e0e23105eceaa1c3493f2487e439b0d75b5e2dd33223be72174113bc\": container with ID starting with 71209c98e0e23105eceaa1c3493f2487e439b0d75b5e2dd33223be72174113bc not found: ID does not exist" Oct 02 02:23:13 crc kubenswrapper[4885]: I1002 02:23:13.655418 4885 scope.go:117] "RemoveContainer" containerID="93fb0d7f421ea046adc4dcbdc38249936364a1e3cffa2f77e6653eab28639894" Oct 02 02:23:13 crc kubenswrapper[4885]: E1002 02:23:13.656228 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"93fb0d7f421ea046adc4dcbdc38249936364a1e3cffa2f77e6653eab28639894\": container with ID starting with 93fb0d7f421ea046adc4dcbdc38249936364a1e3cffa2f77e6653eab28639894 not found: ID does not exist" containerID="93fb0d7f421ea046adc4dcbdc38249936364a1e3cffa2f77e6653eab28639894" Oct 02 02:23:13 crc kubenswrapper[4885]: I1002 02:23:13.656280 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93fb0d7f421ea046adc4dcbdc38249936364a1e3cffa2f77e6653eab28639894"} err="failed to get container status \"93fb0d7f421ea046adc4dcbdc38249936364a1e3cffa2f77e6653eab28639894\": rpc error: code = NotFound desc = could not find container \"93fb0d7f421ea046adc4dcbdc38249936364a1e3cffa2f77e6653eab28639894\": container with ID starting with 93fb0d7f421ea046adc4dcbdc38249936364a1e3cffa2f77e6653eab28639894 not found: ID does not exist" Oct 02 02:23:13 crc kubenswrapper[4885]: I1002 02:23:13.656307 4885 scope.go:117] "RemoveContainer" containerID="cd0c01b30fba66244b3eb5d1dacbf9265556037f274e42673d55aa7dfc1c1fac" Oct 02 02:23:13 crc kubenswrapper[4885]: E1002 02:23:13.656874 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cd0c01b30fba66244b3eb5d1dacbf9265556037f274e42673d55aa7dfc1c1fac\": container with ID starting with cd0c01b30fba66244b3eb5d1dacbf9265556037f274e42673d55aa7dfc1c1fac not found: ID does not exist" containerID="cd0c01b30fba66244b3eb5d1dacbf9265556037f274e42673d55aa7dfc1c1fac" Oct 02 02:23:13 crc kubenswrapper[4885]: I1002 02:23:13.656913 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd0c01b30fba66244b3eb5d1dacbf9265556037f274e42673d55aa7dfc1c1fac"} err="failed to get container status \"cd0c01b30fba66244b3eb5d1dacbf9265556037f274e42673d55aa7dfc1c1fac\": rpc error: code = NotFound desc = could not find container \"cd0c01b30fba66244b3eb5d1dacbf9265556037f274e42673d55aa7dfc1c1fac\": container with ID starting with cd0c01b30fba66244b3eb5d1dacbf9265556037f274e42673d55aa7dfc1c1fac not found: ID does not exist" Oct 02 02:23:13 crc kubenswrapper[4885]: I1002 02:23:13.960373 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c38718b-d4ec-492a-a183-cee6b6a59bfe-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0c38718b-d4ec-492a-a183-cee6b6a59bfe" (UID: "0c38718b-d4ec-492a-a183-cee6b6a59bfe"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:23:13 crc kubenswrapper[4885]: I1002 02:23:13.994822 4885 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c38718b-d4ec-492a-a183-cee6b6a59bfe-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 02:23:14 crc kubenswrapper[4885]: I1002 02:23:14.157591 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-kl8hf"] Oct 02 02:23:14 crc kubenswrapper[4885]: I1002 02:23:14.166292 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-kl8hf"] Oct 02 02:23:16 crc kubenswrapper[4885]: I1002 02:23:16.066683 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c38718b-d4ec-492a-a183-cee6b6a59bfe" path="/var/lib/kubelet/pods/0c38718b-d4ec-492a-a183-cee6b6a59bfe/volumes" Oct 02 02:23:43 crc kubenswrapper[4885]: I1002 02:23:43.266453 4885 patch_prober.go:28] interesting pod/machine-config-daemon-rttx8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 02:23:43 crc kubenswrapper[4885]: I1002 02:23:43.267292 4885 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 02:24:13 crc kubenswrapper[4885]: I1002 02:24:13.266911 4885 patch_prober.go:28] interesting pod/machine-config-daemon-rttx8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 02:24:13 crc kubenswrapper[4885]: I1002 02:24:13.267572 4885 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 02:24:13 crc kubenswrapper[4885]: I1002 02:24:13.267644 4885 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" Oct 02 02:24:13 crc kubenswrapper[4885]: I1002 02:24:13.268779 4885 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7d1be267228233dc9433f75e76a5a805aa23d98aa3b0edbd385a1a565a0d3231"} pod="openshift-machine-config-operator/machine-config-daemon-rttx8" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 02:24:13 crc kubenswrapper[4885]: I1002 02:24:13.268858 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" containerName="machine-config-daemon" containerID="cri-o://7d1be267228233dc9433f75e76a5a805aa23d98aa3b0edbd385a1a565a0d3231" gracePeriod=600 Oct 02 02:24:13 crc kubenswrapper[4885]: E1002 02:24:13.406078 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:24:13 crc kubenswrapper[4885]: E1002 02:24:13.556347 4885 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd169c8ba_000a_4218_8e3b_4ae53035b110.slice/crio-7d1be267228233dc9433f75e76a5a805aa23d98aa3b0edbd385a1a565a0d3231.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd169c8ba_000a_4218_8e3b_4ae53035b110.slice/crio-conmon-7d1be267228233dc9433f75e76a5a805aa23d98aa3b0edbd385a1a565a0d3231.scope\": RecentStats: unable to find data in memory cache]" Oct 02 02:24:14 crc kubenswrapper[4885]: I1002 02:24:14.237355 4885 generic.go:334] "Generic (PLEG): container finished" podID="d169c8ba-000a-4218-8e3b-4ae53035b110" containerID="7d1be267228233dc9433f75e76a5a805aa23d98aa3b0edbd385a1a565a0d3231" exitCode=0 Oct 02 02:24:14 crc kubenswrapper[4885]: I1002 02:24:14.237414 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" event={"ID":"d169c8ba-000a-4218-8e3b-4ae53035b110","Type":"ContainerDied","Data":"7d1be267228233dc9433f75e76a5a805aa23d98aa3b0edbd385a1a565a0d3231"} Oct 02 02:24:14 crc kubenswrapper[4885]: I1002 02:24:14.237797 4885 scope.go:117] "RemoveContainer" containerID="562009e16fbeefe72a46fc9d563e2da31ec7f209f7fd3da12e264d10773a1c88" Oct 02 02:24:14 crc kubenswrapper[4885]: I1002 02:24:14.238623 4885 scope.go:117] "RemoveContainer" containerID="7d1be267228233dc9433f75e76a5a805aa23d98aa3b0edbd385a1a565a0d3231" Oct 02 02:24:14 crc kubenswrapper[4885]: E1002 02:24:14.239045 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:24:26 crc kubenswrapper[4885]: I1002 02:24:26.047670 4885 scope.go:117] "RemoveContainer" containerID="7d1be267228233dc9433f75e76a5a805aa23d98aa3b0edbd385a1a565a0d3231" Oct 02 02:24:26 crc kubenswrapper[4885]: E1002 02:24:26.048945 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:24:38 crc kubenswrapper[4885]: I1002 02:24:38.048956 4885 scope.go:117] "RemoveContainer" containerID="7d1be267228233dc9433f75e76a5a805aa23d98aa3b0edbd385a1a565a0d3231" Oct 02 02:24:38 crc kubenswrapper[4885]: E1002 02:24:38.050312 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:24:51 crc kubenswrapper[4885]: I1002 02:24:51.046831 4885 scope.go:117] "RemoveContainer" containerID="7d1be267228233dc9433f75e76a5a805aa23d98aa3b0edbd385a1a565a0d3231" Oct 02 02:24:51 crc kubenswrapper[4885]: E1002 02:24:51.047942 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:25:03 crc kubenswrapper[4885]: I1002 02:25:03.047006 4885 scope.go:117] "RemoveContainer" containerID="7d1be267228233dc9433f75e76a5a805aa23d98aa3b0edbd385a1a565a0d3231" Oct 02 02:25:03 crc kubenswrapper[4885]: E1002 02:25:03.047934 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:25:16 crc kubenswrapper[4885]: I1002 02:25:16.048334 4885 scope.go:117] "RemoveContainer" containerID="7d1be267228233dc9433f75e76a5a805aa23d98aa3b0edbd385a1a565a0d3231" Oct 02 02:25:16 crc kubenswrapper[4885]: E1002 02:25:16.049056 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:25:29 crc kubenswrapper[4885]: I1002 02:25:29.047791 4885 scope.go:117] "RemoveContainer" containerID="7d1be267228233dc9433f75e76a5a805aa23d98aa3b0edbd385a1a565a0d3231" Oct 02 02:25:29 crc kubenswrapper[4885]: E1002 02:25:29.048937 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:25:42 crc kubenswrapper[4885]: I1002 02:25:42.054932 4885 scope.go:117] "RemoveContainer" containerID="7d1be267228233dc9433f75e76a5a805aa23d98aa3b0edbd385a1a565a0d3231" Oct 02 02:25:42 crc kubenswrapper[4885]: E1002 02:25:42.055982 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:25:49 crc kubenswrapper[4885]: I1002 02:25:49.372563 4885 generic.go:334] "Generic (PLEG): container finished" podID="54cf89d0-a777-4ac9-968b-49836d0b13b1" containerID="62a6c50591117ca89f8931562ac58e8042be832ab8a0c335ce598b0b7c4d4c2e" exitCode=0 Oct 02 02:25:49 crc kubenswrapper[4885]: I1002 02:25:49.372659 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-x47dm" event={"ID":"54cf89d0-a777-4ac9-968b-49836d0b13b1","Type":"ContainerDied","Data":"62a6c50591117ca89f8931562ac58e8042be832ab8a0c335ce598b0b7c4d4c2e"} Oct 02 02:25:50 crc kubenswrapper[4885]: I1002 02:25:50.797686 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-x47dm" Oct 02 02:25:50 crc kubenswrapper[4885]: I1002 02:25:50.928859 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/54cf89d0-a777-4ac9-968b-49836d0b13b1-libvirt-secret-0\") pod \"54cf89d0-a777-4ac9-968b-49836d0b13b1\" (UID: \"54cf89d0-a777-4ac9-968b-49836d0b13b1\") " Oct 02 02:25:50 crc kubenswrapper[4885]: I1002 02:25:50.929116 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/54cf89d0-a777-4ac9-968b-49836d0b13b1-inventory\") pod \"54cf89d0-a777-4ac9-968b-49836d0b13b1\" (UID: \"54cf89d0-a777-4ac9-968b-49836d0b13b1\") " Oct 02 02:25:50 crc kubenswrapper[4885]: I1002 02:25:50.929343 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54cf89d0-a777-4ac9-968b-49836d0b13b1-libvirt-combined-ca-bundle\") pod \"54cf89d0-a777-4ac9-968b-49836d0b13b1\" (UID: \"54cf89d0-a777-4ac9-968b-49836d0b13b1\") " Oct 02 02:25:50 crc kubenswrapper[4885]: I1002 02:25:50.929436 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/54cf89d0-a777-4ac9-968b-49836d0b13b1-ssh-key\") pod \"54cf89d0-a777-4ac9-968b-49836d0b13b1\" (UID: \"54cf89d0-a777-4ac9-968b-49836d0b13b1\") " Oct 02 02:25:50 crc kubenswrapper[4885]: I1002 02:25:50.929491 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ph8fc\" (UniqueName: \"kubernetes.io/projected/54cf89d0-a777-4ac9-968b-49836d0b13b1-kube-api-access-ph8fc\") pod \"54cf89d0-a777-4ac9-968b-49836d0b13b1\" (UID: \"54cf89d0-a777-4ac9-968b-49836d0b13b1\") " Oct 02 02:25:50 crc kubenswrapper[4885]: I1002 02:25:50.936240 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54cf89d0-a777-4ac9-968b-49836d0b13b1-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "54cf89d0-a777-4ac9-968b-49836d0b13b1" (UID: "54cf89d0-a777-4ac9-968b-49836d0b13b1"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:25:50 crc kubenswrapper[4885]: I1002 02:25:50.936492 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/54cf89d0-a777-4ac9-968b-49836d0b13b1-kube-api-access-ph8fc" (OuterVolumeSpecName: "kube-api-access-ph8fc") pod "54cf89d0-a777-4ac9-968b-49836d0b13b1" (UID: "54cf89d0-a777-4ac9-968b-49836d0b13b1"). InnerVolumeSpecName "kube-api-access-ph8fc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:25:50 crc kubenswrapper[4885]: I1002 02:25:50.963916 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54cf89d0-a777-4ac9-968b-49836d0b13b1-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "54cf89d0-a777-4ac9-968b-49836d0b13b1" (UID: "54cf89d0-a777-4ac9-968b-49836d0b13b1"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:25:50 crc kubenswrapper[4885]: I1002 02:25:50.965936 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54cf89d0-a777-4ac9-968b-49836d0b13b1-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "54cf89d0-a777-4ac9-968b-49836d0b13b1" (UID: "54cf89d0-a777-4ac9-968b-49836d0b13b1"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:25:50 crc kubenswrapper[4885]: I1002 02:25:50.972467 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54cf89d0-a777-4ac9-968b-49836d0b13b1-inventory" (OuterVolumeSpecName: "inventory") pod "54cf89d0-a777-4ac9-968b-49836d0b13b1" (UID: "54cf89d0-a777-4ac9-968b-49836d0b13b1"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:25:51 crc kubenswrapper[4885]: I1002 02:25:51.032830 4885 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54cf89d0-a777-4ac9-968b-49836d0b13b1-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:25:51 crc kubenswrapper[4885]: I1002 02:25:51.032874 4885 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/54cf89d0-a777-4ac9-968b-49836d0b13b1-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 02:25:51 crc kubenswrapper[4885]: I1002 02:25:51.032887 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ph8fc\" (UniqueName: \"kubernetes.io/projected/54cf89d0-a777-4ac9-968b-49836d0b13b1-kube-api-access-ph8fc\") on node \"crc\" DevicePath \"\"" Oct 02 02:25:51 crc kubenswrapper[4885]: I1002 02:25:51.032900 4885 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/54cf89d0-a777-4ac9-968b-49836d0b13b1-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Oct 02 02:25:51 crc kubenswrapper[4885]: I1002 02:25:51.032912 4885 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/54cf89d0-a777-4ac9-968b-49836d0b13b1-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 02:25:51 crc kubenswrapper[4885]: I1002 02:25:51.400076 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-x47dm" event={"ID":"54cf89d0-a777-4ac9-968b-49836d0b13b1","Type":"ContainerDied","Data":"978e2d4949a047641300bba9c41a77527e0736cefc5bb06ca536910885095780"} Oct 02 02:25:51 crc kubenswrapper[4885]: I1002 02:25:51.400120 4885 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="978e2d4949a047641300bba9c41a77527e0736cefc5bb06ca536910885095780" Oct 02 02:25:51 crc kubenswrapper[4885]: I1002 02:25:51.400173 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-x47dm" Oct 02 02:25:51 crc kubenswrapper[4885]: I1002 02:25:51.508621 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-c6gw4"] Oct 02 02:25:51 crc kubenswrapper[4885]: E1002 02:25:51.509045 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c38718b-d4ec-492a-a183-cee6b6a59bfe" containerName="extract-content" Oct 02 02:25:51 crc kubenswrapper[4885]: I1002 02:25:51.509066 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c38718b-d4ec-492a-a183-cee6b6a59bfe" containerName="extract-content" Oct 02 02:25:51 crc kubenswrapper[4885]: E1002 02:25:51.509082 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c38718b-d4ec-492a-a183-cee6b6a59bfe" containerName="registry-server" Oct 02 02:25:51 crc kubenswrapper[4885]: I1002 02:25:51.509089 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c38718b-d4ec-492a-a183-cee6b6a59bfe" containerName="registry-server" Oct 02 02:25:51 crc kubenswrapper[4885]: E1002 02:25:51.509105 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54cf89d0-a777-4ac9-968b-49836d0b13b1" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Oct 02 02:25:51 crc kubenswrapper[4885]: I1002 02:25:51.509114 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="54cf89d0-a777-4ac9-968b-49836d0b13b1" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Oct 02 02:25:51 crc kubenswrapper[4885]: E1002 02:25:51.509140 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c38718b-d4ec-492a-a183-cee6b6a59bfe" containerName="extract-utilities" Oct 02 02:25:51 crc kubenswrapper[4885]: I1002 02:25:51.509148 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c38718b-d4ec-492a-a183-cee6b6a59bfe" containerName="extract-utilities" Oct 02 02:25:51 crc kubenswrapper[4885]: I1002 02:25:51.509404 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="54cf89d0-a777-4ac9-968b-49836d0b13b1" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Oct 02 02:25:51 crc kubenswrapper[4885]: I1002 02:25:51.509436 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c38718b-d4ec-492a-a183-cee6b6a59bfe" containerName="registry-server" Oct 02 02:25:51 crc kubenswrapper[4885]: I1002 02:25:51.510163 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-c6gw4" Oct 02 02:25:51 crc kubenswrapper[4885]: I1002 02:25:51.513087 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 02:25:51 crc kubenswrapper[4885]: I1002 02:25:51.513234 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-f8tvb" Oct 02 02:25:51 crc kubenswrapper[4885]: I1002 02:25:51.513299 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Oct 02 02:25:51 crc kubenswrapper[4885]: I1002 02:25:51.513806 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 02:25:51 crc kubenswrapper[4885]: I1002 02:25:51.513916 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 02:25:51 crc kubenswrapper[4885]: I1002 02:25:51.513953 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Oct 02 02:25:51 crc kubenswrapper[4885]: I1002 02:25:51.515244 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Oct 02 02:25:51 crc kubenswrapper[4885]: I1002 02:25:51.530232 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-c6gw4"] Oct 02 02:25:51 crc kubenswrapper[4885]: I1002 02:25:51.645676 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v8q27\" (UniqueName: \"kubernetes.io/projected/14bdd01b-529b-45b4-9ac3-188fdf3d3855-kube-api-access-v8q27\") pod \"nova-edpm-deployment-openstack-edpm-ipam-c6gw4\" (UID: \"14bdd01b-529b-45b4-9ac3-188fdf3d3855\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-c6gw4" Oct 02 02:25:51 crc kubenswrapper[4885]: I1002 02:25:51.645903 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/14bdd01b-529b-45b4-9ac3-188fdf3d3855-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-c6gw4\" (UID: \"14bdd01b-529b-45b4-9ac3-188fdf3d3855\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-c6gw4" Oct 02 02:25:51 crc kubenswrapper[4885]: I1002 02:25:51.645977 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/14bdd01b-529b-45b4-9ac3-188fdf3d3855-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-c6gw4\" (UID: \"14bdd01b-529b-45b4-9ac3-188fdf3d3855\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-c6gw4" Oct 02 02:25:51 crc kubenswrapper[4885]: I1002 02:25:51.646190 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14bdd01b-529b-45b4-9ac3-188fdf3d3855-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-c6gw4\" (UID: \"14bdd01b-529b-45b4-9ac3-188fdf3d3855\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-c6gw4" Oct 02 02:25:51 crc kubenswrapper[4885]: I1002 02:25:51.646226 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/14bdd01b-529b-45b4-9ac3-188fdf3d3855-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-c6gw4\" (UID: \"14bdd01b-529b-45b4-9ac3-188fdf3d3855\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-c6gw4" Oct 02 02:25:51 crc kubenswrapper[4885]: I1002 02:25:51.646482 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/14bdd01b-529b-45b4-9ac3-188fdf3d3855-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-c6gw4\" (UID: \"14bdd01b-529b-45b4-9ac3-188fdf3d3855\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-c6gw4" Oct 02 02:25:51 crc kubenswrapper[4885]: I1002 02:25:51.646535 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/14bdd01b-529b-45b4-9ac3-188fdf3d3855-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-c6gw4\" (UID: \"14bdd01b-529b-45b4-9ac3-188fdf3d3855\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-c6gw4" Oct 02 02:25:51 crc kubenswrapper[4885]: I1002 02:25:51.646591 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/14bdd01b-529b-45b4-9ac3-188fdf3d3855-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-c6gw4\" (UID: \"14bdd01b-529b-45b4-9ac3-188fdf3d3855\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-c6gw4" Oct 02 02:25:51 crc kubenswrapper[4885]: I1002 02:25:51.646650 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/14bdd01b-529b-45b4-9ac3-188fdf3d3855-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-c6gw4\" (UID: \"14bdd01b-529b-45b4-9ac3-188fdf3d3855\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-c6gw4" Oct 02 02:25:51 crc kubenswrapper[4885]: I1002 02:25:51.747895 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v8q27\" (UniqueName: \"kubernetes.io/projected/14bdd01b-529b-45b4-9ac3-188fdf3d3855-kube-api-access-v8q27\") pod \"nova-edpm-deployment-openstack-edpm-ipam-c6gw4\" (UID: \"14bdd01b-529b-45b4-9ac3-188fdf3d3855\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-c6gw4" Oct 02 02:25:51 crc kubenswrapper[4885]: I1002 02:25:51.747973 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/14bdd01b-529b-45b4-9ac3-188fdf3d3855-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-c6gw4\" (UID: \"14bdd01b-529b-45b4-9ac3-188fdf3d3855\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-c6gw4" Oct 02 02:25:51 crc kubenswrapper[4885]: I1002 02:25:51.747998 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/14bdd01b-529b-45b4-9ac3-188fdf3d3855-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-c6gw4\" (UID: \"14bdd01b-529b-45b4-9ac3-188fdf3d3855\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-c6gw4" Oct 02 02:25:51 crc kubenswrapper[4885]: I1002 02:25:51.748047 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14bdd01b-529b-45b4-9ac3-188fdf3d3855-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-c6gw4\" (UID: \"14bdd01b-529b-45b4-9ac3-188fdf3d3855\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-c6gw4" Oct 02 02:25:51 crc kubenswrapper[4885]: I1002 02:25:51.748070 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/14bdd01b-529b-45b4-9ac3-188fdf3d3855-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-c6gw4\" (UID: \"14bdd01b-529b-45b4-9ac3-188fdf3d3855\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-c6gw4" Oct 02 02:25:51 crc kubenswrapper[4885]: I1002 02:25:51.748107 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/14bdd01b-529b-45b4-9ac3-188fdf3d3855-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-c6gw4\" (UID: \"14bdd01b-529b-45b4-9ac3-188fdf3d3855\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-c6gw4" Oct 02 02:25:51 crc kubenswrapper[4885]: I1002 02:25:51.748131 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/14bdd01b-529b-45b4-9ac3-188fdf3d3855-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-c6gw4\" (UID: \"14bdd01b-529b-45b4-9ac3-188fdf3d3855\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-c6gw4" Oct 02 02:25:51 crc kubenswrapper[4885]: I1002 02:25:51.748157 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/14bdd01b-529b-45b4-9ac3-188fdf3d3855-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-c6gw4\" (UID: \"14bdd01b-529b-45b4-9ac3-188fdf3d3855\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-c6gw4" Oct 02 02:25:51 crc kubenswrapper[4885]: I1002 02:25:51.748182 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/14bdd01b-529b-45b4-9ac3-188fdf3d3855-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-c6gw4\" (UID: \"14bdd01b-529b-45b4-9ac3-188fdf3d3855\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-c6gw4" Oct 02 02:25:51 crc kubenswrapper[4885]: I1002 02:25:51.750105 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/14bdd01b-529b-45b4-9ac3-188fdf3d3855-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-c6gw4\" (UID: \"14bdd01b-529b-45b4-9ac3-188fdf3d3855\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-c6gw4" Oct 02 02:25:51 crc kubenswrapper[4885]: I1002 02:25:51.754531 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/14bdd01b-529b-45b4-9ac3-188fdf3d3855-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-c6gw4\" (UID: \"14bdd01b-529b-45b4-9ac3-188fdf3d3855\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-c6gw4" Oct 02 02:25:51 crc kubenswrapper[4885]: I1002 02:25:51.754820 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/14bdd01b-529b-45b4-9ac3-188fdf3d3855-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-c6gw4\" (UID: \"14bdd01b-529b-45b4-9ac3-188fdf3d3855\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-c6gw4" Oct 02 02:25:51 crc kubenswrapper[4885]: I1002 02:25:51.754911 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/14bdd01b-529b-45b4-9ac3-188fdf3d3855-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-c6gw4\" (UID: \"14bdd01b-529b-45b4-9ac3-188fdf3d3855\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-c6gw4" Oct 02 02:25:51 crc kubenswrapper[4885]: I1002 02:25:51.755790 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14bdd01b-529b-45b4-9ac3-188fdf3d3855-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-c6gw4\" (UID: \"14bdd01b-529b-45b4-9ac3-188fdf3d3855\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-c6gw4" Oct 02 02:25:51 crc kubenswrapper[4885]: I1002 02:25:51.756543 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/14bdd01b-529b-45b4-9ac3-188fdf3d3855-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-c6gw4\" (UID: \"14bdd01b-529b-45b4-9ac3-188fdf3d3855\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-c6gw4" Oct 02 02:25:51 crc kubenswrapper[4885]: I1002 02:25:51.757653 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/14bdd01b-529b-45b4-9ac3-188fdf3d3855-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-c6gw4\" (UID: \"14bdd01b-529b-45b4-9ac3-188fdf3d3855\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-c6gw4" Oct 02 02:25:51 crc kubenswrapper[4885]: I1002 02:25:51.759083 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/14bdd01b-529b-45b4-9ac3-188fdf3d3855-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-c6gw4\" (UID: \"14bdd01b-529b-45b4-9ac3-188fdf3d3855\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-c6gw4" Oct 02 02:25:51 crc kubenswrapper[4885]: I1002 02:25:51.780722 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v8q27\" (UniqueName: \"kubernetes.io/projected/14bdd01b-529b-45b4-9ac3-188fdf3d3855-kube-api-access-v8q27\") pod \"nova-edpm-deployment-openstack-edpm-ipam-c6gw4\" (UID: \"14bdd01b-529b-45b4-9ac3-188fdf3d3855\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-c6gw4" Oct 02 02:25:51 crc kubenswrapper[4885]: I1002 02:25:51.828792 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-c6gw4" Oct 02 02:25:52 crc kubenswrapper[4885]: I1002 02:25:52.502834 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-c6gw4"] Oct 02 02:25:52 crc kubenswrapper[4885]: I1002 02:25:52.994724 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 02:25:53 crc kubenswrapper[4885]: I1002 02:25:53.047305 4885 scope.go:117] "RemoveContainer" containerID="7d1be267228233dc9433f75e76a5a805aa23d98aa3b0edbd385a1a565a0d3231" Oct 02 02:25:53 crc kubenswrapper[4885]: E1002 02:25:53.047726 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:25:53 crc kubenswrapper[4885]: I1002 02:25:53.425754 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-c6gw4" event={"ID":"14bdd01b-529b-45b4-9ac3-188fdf3d3855","Type":"ContainerStarted","Data":"07741622db80033b74c21df5824568f5d17a2b20d14e3c8c1d8df8c07b24347d"} Oct 02 02:25:53 crc kubenswrapper[4885]: I1002 02:25:53.426256 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-c6gw4" event={"ID":"14bdd01b-529b-45b4-9ac3-188fdf3d3855","Type":"ContainerStarted","Data":"628e9325eb6d394d488080264309a506e38c2a97b36e681bba0ffb18d52a9c51"} Oct 02 02:25:53 crc kubenswrapper[4885]: I1002 02:25:53.461784 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-c6gw4" podStartSLOduration=1.98341732 podStartE2EDuration="2.461727762s" podCreationTimestamp="2025-10-02 02:25:51 +0000 UTC" firstStartedPulling="2025-10-02 02:25:52.51276535 +0000 UTC m=+2341.324512789" lastFinishedPulling="2025-10-02 02:25:52.991075802 +0000 UTC m=+2341.802823231" observedRunningTime="2025-10-02 02:25:53.45085086 +0000 UTC m=+2342.262598269" watchObservedRunningTime="2025-10-02 02:25:53.461727762 +0000 UTC m=+2342.273475191" Oct 02 02:26:06 crc kubenswrapper[4885]: I1002 02:26:06.047618 4885 scope.go:117] "RemoveContainer" containerID="7d1be267228233dc9433f75e76a5a805aa23d98aa3b0edbd385a1a565a0d3231" Oct 02 02:26:06 crc kubenswrapper[4885]: E1002 02:26:06.048969 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:26:19 crc kubenswrapper[4885]: I1002 02:26:19.046161 4885 scope.go:117] "RemoveContainer" containerID="7d1be267228233dc9433f75e76a5a805aa23d98aa3b0edbd385a1a565a0d3231" Oct 02 02:26:19 crc kubenswrapper[4885]: E1002 02:26:19.047148 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:26:31 crc kubenswrapper[4885]: I1002 02:26:31.048045 4885 scope.go:117] "RemoveContainer" containerID="7d1be267228233dc9433f75e76a5a805aa23d98aa3b0edbd385a1a565a0d3231" Oct 02 02:26:31 crc kubenswrapper[4885]: E1002 02:26:31.048816 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:26:43 crc kubenswrapper[4885]: I1002 02:26:43.045904 4885 scope.go:117] "RemoveContainer" containerID="7d1be267228233dc9433f75e76a5a805aa23d98aa3b0edbd385a1a565a0d3231" Oct 02 02:26:43 crc kubenswrapper[4885]: E1002 02:26:43.046703 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:26:55 crc kubenswrapper[4885]: I1002 02:26:55.047445 4885 scope.go:117] "RemoveContainer" containerID="7d1be267228233dc9433f75e76a5a805aa23d98aa3b0edbd385a1a565a0d3231" Oct 02 02:26:55 crc kubenswrapper[4885]: E1002 02:26:55.049606 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:27:06 crc kubenswrapper[4885]: I1002 02:27:06.047070 4885 scope.go:117] "RemoveContainer" containerID="7d1be267228233dc9433f75e76a5a805aa23d98aa3b0edbd385a1a565a0d3231" Oct 02 02:27:06 crc kubenswrapper[4885]: E1002 02:27:06.048177 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:27:10 crc kubenswrapper[4885]: I1002 02:27:10.163949 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-qpv6k"] Oct 02 02:27:10 crc kubenswrapper[4885]: I1002 02:27:10.169522 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qpv6k" Oct 02 02:27:10 crc kubenswrapper[4885]: I1002 02:27:10.172830 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qpv6k"] Oct 02 02:27:10 crc kubenswrapper[4885]: I1002 02:27:10.281201 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08919919-e9d1-4980-8c89-2f3891f3c576-utilities\") pod \"redhat-marketplace-qpv6k\" (UID: \"08919919-e9d1-4980-8c89-2f3891f3c576\") " pod="openshift-marketplace/redhat-marketplace-qpv6k" Oct 02 02:27:10 crc kubenswrapper[4885]: I1002 02:27:10.281296 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08919919-e9d1-4980-8c89-2f3891f3c576-catalog-content\") pod \"redhat-marketplace-qpv6k\" (UID: \"08919919-e9d1-4980-8c89-2f3891f3c576\") " pod="openshift-marketplace/redhat-marketplace-qpv6k" Oct 02 02:27:10 crc kubenswrapper[4885]: I1002 02:27:10.281385 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7rf7b\" (UniqueName: \"kubernetes.io/projected/08919919-e9d1-4980-8c89-2f3891f3c576-kube-api-access-7rf7b\") pod \"redhat-marketplace-qpv6k\" (UID: \"08919919-e9d1-4980-8c89-2f3891f3c576\") " pod="openshift-marketplace/redhat-marketplace-qpv6k" Oct 02 02:27:10 crc kubenswrapper[4885]: I1002 02:27:10.383387 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08919919-e9d1-4980-8c89-2f3891f3c576-utilities\") pod \"redhat-marketplace-qpv6k\" (UID: \"08919919-e9d1-4980-8c89-2f3891f3c576\") " pod="openshift-marketplace/redhat-marketplace-qpv6k" Oct 02 02:27:10 crc kubenswrapper[4885]: I1002 02:27:10.383443 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08919919-e9d1-4980-8c89-2f3891f3c576-catalog-content\") pod \"redhat-marketplace-qpv6k\" (UID: \"08919919-e9d1-4980-8c89-2f3891f3c576\") " pod="openshift-marketplace/redhat-marketplace-qpv6k" Oct 02 02:27:10 crc kubenswrapper[4885]: I1002 02:27:10.383496 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7rf7b\" (UniqueName: \"kubernetes.io/projected/08919919-e9d1-4980-8c89-2f3891f3c576-kube-api-access-7rf7b\") pod \"redhat-marketplace-qpv6k\" (UID: \"08919919-e9d1-4980-8c89-2f3891f3c576\") " pod="openshift-marketplace/redhat-marketplace-qpv6k" Oct 02 02:27:10 crc kubenswrapper[4885]: I1002 02:27:10.384075 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08919919-e9d1-4980-8c89-2f3891f3c576-utilities\") pod \"redhat-marketplace-qpv6k\" (UID: \"08919919-e9d1-4980-8c89-2f3891f3c576\") " pod="openshift-marketplace/redhat-marketplace-qpv6k" Oct 02 02:27:10 crc kubenswrapper[4885]: I1002 02:27:10.384158 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08919919-e9d1-4980-8c89-2f3891f3c576-catalog-content\") pod \"redhat-marketplace-qpv6k\" (UID: \"08919919-e9d1-4980-8c89-2f3891f3c576\") " pod="openshift-marketplace/redhat-marketplace-qpv6k" Oct 02 02:27:10 crc kubenswrapper[4885]: I1002 02:27:10.405869 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7rf7b\" (UniqueName: \"kubernetes.io/projected/08919919-e9d1-4980-8c89-2f3891f3c576-kube-api-access-7rf7b\") pod \"redhat-marketplace-qpv6k\" (UID: \"08919919-e9d1-4980-8c89-2f3891f3c576\") " pod="openshift-marketplace/redhat-marketplace-qpv6k" Oct 02 02:27:10 crc kubenswrapper[4885]: I1002 02:27:10.499976 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qpv6k" Oct 02 02:27:10 crc kubenswrapper[4885]: I1002 02:27:10.972021 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qpv6k"] Oct 02 02:27:11 crc kubenswrapper[4885]: I1002 02:27:11.325079 4885 generic.go:334] "Generic (PLEG): container finished" podID="08919919-e9d1-4980-8c89-2f3891f3c576" containerID="60fdbb5e85f14c6430ca41e200ba8e9ce88087674e1ac7a5122b6026917ad7ea" exitCode=0 Oct 02 02:27:11 crc kubenswrapper[4885]: I1002 02:27:11.325234 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qpv6k" event={"ID":"08919919-e9d1-4980-8c89-2f3891f3c576","Type":"ContainerDied","Data":"60fdbb5e85f14c6430ca41e200ba8e9ce88087674e1ac7a5122b6026917ad7ea"} Oct 02 02:27:11 crc kubenswrapper[4885]: I1002 02:27:11.325528 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qpv6k" event={"ID":"08919919-e9d1-4980-8c89-2f3891f3c576","Type":"ContainerStarted","Data":"1e5eb9006d927044d9c0a073828d118982b1e3925a72544a1e533237c24b4a68"} Oct 02 02:27:13 crc kubenswrapper[4885]: I1002 02:27:13.371474 4885 generic.go:334] "Generic (PLEG): container finished" podID="08919919-e9d1-4980-8c89-2f3891f3c576" containerID="e9991f15d20ed2999323588b2d7f478d025fe4ab4ef8763b85ba4892476b51a7" exitCode=0 Oct 02 02:27:13 crc kubenswrapper[4885]: I1002 02:27:13.372186 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qpv6k" event={"ID":"08919919-e9d1-4980-8c89-2f3891f3c576","Type":"ContainerDied","Data":"e9991f15d20ed2999323588b2d7f478d025fe4ab4ef8763b85ba4892476b51a7"} Oct 02 02:27:14 crc kubenswrapper[4885]: I1002 02:27:14.386012 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qpv6k" event={"ID":"08919919-e9d1-4980-8c89-2f3891f3c576","Type":"ContainerStarted","Data":"33ab2a1362ed209d7cafd7ceec0435bdb5318ee2ba63bea556d8a678d1cd33c9"} Oct 02 02:27:14 crc kubenswrapper[4885]: I1002 02:27:14.422319 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-qpv6k" podStartSLOduration=1.860301338 podStartE2EDuration="4.422300325s" podCreationTimestamp="2025-10-02 02:27:10 +0000 UTC" firstStartedPulling="2025-10-02 02:27:11.327305079 +0000 UTC m=+2420.139052478" lastFinishedPulling="2025-10-02 02:27:13.889304066 +0000 UTC m=+2422.701051465" observedRunningTime="2025-10-02 02:27:14.417075197 +0000 UTC m=+2423.228822636" watchObservedRunningTime="2025-10-02 02:27:14.422300325 +0000 UTC m=+2423.234047734" Oct 02 02:27:18 crc kubenswrapper[4885]: I1002 02:27:18.046985 4885 scope.go:117] "RemoveContainer" containerID="7d1be267228233dc9433f75e76a5a805aa23d98aa3b0edbd385a1a565a0d3231" Oct 02 02:27:18 crc kubenswrapper[4885]: E1002 02:27:18.047693 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:27:20 crc kubenswrapper[4885]: I1002 02:27:20.501031 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-qpv6k" Oct 02 02:27:20 crc kubenswrapper[4885]: I1002 02:27:20.501465 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-qpv6k" Oct 02 02:27:20 crc kubenswrapper[4885]: I1002 02:27:20.587997 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-qpv6k" Oct 02 02:27:21 crc kubenswrapper[4885]: I1002 02:27:21.557618 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-qpv6k" Oct 02 02:27:21 crc kubenswrapper[4885]: I1002 02:27:21.626916 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qpv6k"] Oct 02 02:27:23 crc kubenswrapper[4885]: I1002 02:27:23.496122 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-qpv6k" podUID="08919919-e9d1-4980-8c89-2f3891f3c576" containerName="registry-server" containerID="cri-o://33ab2a1362ed209d7cafd7ceec0435bdb5318ee2ba63bea556d8a678d1cd33c9" gracePeriod=2 Oct 02 02:27:24 crc kubenswrapper[4885]: I1002 02:27:24.050138 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qpv6k" Oct 02 02:27:24 crc kubenswrapper[4885]: I1002 02:27:24.199437 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08919919-e9d1-4980-8c89-2f3891f3c576-catalog-content\") pod \"08919919-e9d1-4980-8c89-2f3891f3c576\" (UID: \"08919919-e9d1-4980-8c89-2f3891f3c576\") " Oct 02 02:27:24 crc kubenswrapper[4885]: I1002 02:27:24.199523 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7rf7b\" (UniqueName: \"kubernetes.io/projected/08919919-e9d1-4980-8c89-2f3891f3c576-kube-api-access-7rf7b\") pod \"08919919-e9d1-4980-8c89-2f3891f3c576\" (UID: \"08919919-e9d1-4980-8c89-2f3891f3c576\") " Oct 02 02:27:24 crc kubenswrapper[4885]: I1002 02:27:24.199623 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08919919-e9d1-4980-8c89-2f3891f3c576-utilities\") pod \"08919919-e9d1-4980-8c89-2f3891f3c576\" (UID: \"08919919-e9d1-4980-8c89-2f3891f3c576\") " Oct 02 02:27:24 crc kubenswrapper[4885]: I1002 02:27:24.200836 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/08919919-e9d1-4980-8c89-2f3891f3c576-utilities" (OuterVolumeSpecName: "utilities") pod "08919919-e9d1-4980-8c89-2f3891f3c576" (UID: "08919919-e9d1-4980-8c89-2f3891f3c576"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:27:24 crc kubenswrapper[4885]: I1002 02:27:24.208001 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08919919-e9d1-4980-8c89-2f3891f3c576-kube-api-access-7rf7b" (OuterVolumeSpecName: "kube-api-access-7rf7b") pod "08919919-e9d1-4980-8c89-2f3891f3c576" (UID: "08919919-e9d1-4980-8c89-2f3891f3c576"). InnerVolumeSpecName "kube-api-access-7rf7b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:27:24 crc kubenswrapper[4885]: I1002 02:27:24.216880 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/08919919-e9d1-4980-8c89-2f3891f3c576-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "08919919-e9d1-4980-8c89-2f3891f3c576" (UID: "08919919-e9d1-4980-8c89-2f3891f3c576"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:27:24 crc kubenswrapper[4885]: I1002 02:27:24.301787 4885 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08919919-e9d1-4980-8c89-2f3891f3c576-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 02:27:24 crc kubenswrapper[4885]: I1002 02:27:24.301825 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7rf7b\" (UniqueName: \"kubernetes.io/projected/08919919-e9d1-4980-8c89-2f3891f3c576-kube-api-access-7rf7b\") on node \"crc\" DevicePath \"\"" Oct 02 02:27:24 crc kubenswrapper[4885]: I1002 02:27:24.301841 4885 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08919919-e9d1-4980-8c89-2f3891f3c576-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 02:27:24 crc kubenswrapper[4885]: I1002 02:27:24.515187 4885 generic.go:334] "Generic (PLEG): container finished" podID="08919919-e9d1-4980-8c89-2f3891f3c576" containerID="33ab2a1362ed209d7cafd7ceec0435bdb5318ee2ba63bea556d8a678d1cd33c9" exitCode=0 Oct 02 02:27:24 crc kubenswrapper[4885]: I1002 02:27:24.515224 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qpv6k" event={"ID":"08919919-e9d1-4980-8c89-2f3891f3c576","Type":"ContainerDied","Data":"33ab2a1362ed209d7cafd7ceec0435bdb5318ee2ba63bea556d8a678d1cd33c9"} Oct 02 02:27:24 crc kubenswrapper[4885]: I1002 02:27:24.515247 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qpv6k" event={"ID":"08919919-e9d1-4980-8c89-2f3891f3c576","Type":"ContainerDied","Data":"1e5eb9006d927044d9c0a073828d118982b1e3925a72544a1e533237c24b4a68"} Oct 02 02:27:24 crc kubenswrapper[4885]: I1002 02:27:24.515289 4885 scope.go:117] "RemoveContainer" containerID="33ab2a1362ed209d7cafd7ceec0435bdb5318ee2ba63bea556d8a678d1cd33c9" Oct 02 02:27:24 crc kubenswrapper[4885]: I1002 02:27:24.515405 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qpv6k" Oct 02 02:27:24 crc kubenswrapper[4885]: I1002 02:27:24.552614 4885 scope.go:117] "RemoveContainer" containerID="e9991f15d20ed2999323588b2d7f478d025fe4ab4ef8763b85ba4892476b51a7" Oct 02 02:27:24 crc kubenswrapper[4885]: I1002 02:27:24.562921 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qpv6k"] Oct 02 02:27:24 crc kubenswrapper[4885]: I1002 02:27:24.571367 4885 scope.go:117] "RemoveContainer" containerID="60fdbb5e85f14c6430ca41e200ba8e9ce88087674e1ac7a5122b6026917ad7ea" Oct 02 02:27:24 crc kubenswrapper[4885]: I1002 02:27:24.573921 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-qpv6k"] Oct 02 02:27:24 crc kubenswrapper[4885]: I1002 02:27:24.622635 4885 scope.go:117] "RemoveContainer" containerID="33ab2a1362ed209d7cafd7ceec0435bdb5318ee2ba63bea556d8a678d1cd33c9" Oct 02 02:27:24 crc kubenswrapper[4885]: E1002 02:27:24.623158 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"33ab2a1362ed209d7cafd7ceec0435bdb5318ee2ba63bea556d8a678d1cd33c9\": container with ID starting with 33ab2a1362ed209d7cafd7ceec0435bdb5318ee2ba63bea556d8a678d1cd33c9 not found: ID does not exist" containerID="33ab2a1362ed209d7cafd7ceec0435bdb5318ee2ba63bea556d8a678d1cd33c9" Oct 02 02:27:24 crc kubenswrapper[4885]: I1002 02:27:24.623236 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"33ab2a1362ed209d7cafd7ceec0435bdb5318ee2ba63bea556d8a678d1cd33c9"} err="failed to get container status \"33ab2a1362ed209d7cafd7ceec0435bdb5318ee2ba63bea556d8a678d1cd33c9\": rpc error: code = NotFound desc = could not find container \"33ab2a1362ed209d7cafd7ceec0435bdb5318ee2ba63bea556d8a678d1cd33c9\": container with ID starting with 33ab2a1362ed209d7cafd7ceec0435bdb5318ee2ba63bea556d8a678d1cd33c9 not found: ID does not exist" Oct 02 02:27:24 crc kubenswrapper[4885]: I1002 02:27:24.623346 4885 scope.go:117] "RemoveContainer" containerID="e9991f15d20ed2999323588b2d7f478d025fe4ab4ef8763b85ba4892476b51a7" Oct 02 02:27:24 crc kubenswrapper[4885]: E1002 02:27:24.624366 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e9991f15d20ed2999323588b2d7f478d025fe4ab4ef8763b85ba4892476b51a7\": container with ID starting with e9991f15d20ed2999323588b2d7f478d025fe4ab4ef8763b85ba4892476b51a7 not found: ID does not exist" containerID="e9991f15d20ed2999323588b2d7f478d025fe4ab4ef8763b85ba4892476b51a7" Oct 02 02:27:24 crc kubenswrapper[4885]: I1002 02:27:24.624407 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e9991f15d20ed2999323588b2d7f478d025fe4ab4ef8763b85ba4892476b51a7"} err="failed to get container status \"e9991f15d20ed2999323588b2d7f478d025fe4ab4ef8763b85ba4892476b51a7\": rpc error: code = NotFound desc = could not find container \"e9991f15d20ed2999323588b2d7f478d025fe4ab4ef8763b85ba4892476b51a7\": container with ID starting with e9991f15d20ed2999323588b2d7f478d025fe4ab4ef8763b85ba4892476b51a7 not found: ID does not exist" Oct 02 02:27:24 crc kubenswrapper[4885]: I1002 02:27:24.624433 4885 scope.go:117] "RemoveContainer" containerID="60fdbb5e85f14c6430ca41e200ba8e9ce88087674e1ac7a5122b6026917ad7ea" Oct 02 02:27:24 crc kubenswrapper[4885]: E1002 02:27:24.624844 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"60fdbb5e85f14c6430ca41e200ba8e9ce88087674e1ac7a5122b6026917ad7ea\": container with ID starting with 60fdbb5e85f14c6430ca41e200ba8e9ce88087674e1ac7a5122b6026917ad7ea not found: ID does not exist" containerID="60fdbb5e85f14c6430ca41e200ba8e9ce88087674e1ac7a5122b6026917ad7ea" Oct 02 02:27:24 crc kubenswrapper[4885]: I1002 02:27:24.624876 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"60fdbb5e85f14c6430ca41e200ba8e9ce88087674e1ac7a5122b6026917ad7ea"} err="failed to get container status \"60fdbb5e85f14c6430ca41e200ba8e9ce88087674e1ac7a5122b6026917ad7ea\": rpc error: code = NotFound desc = could not find container \"60fdbb5e85f14c6430ca41e200ba8e9ce88087674e1ac7a5122b6026917ad7ea\": container with ID starting with 60fdbb5e85f14c6430ca41e200ba8e9ce88087674e1ac7a5122b6026917ad7ea not found: ID does not exist" Oct 02 02:27:26 crc kubenswrapper[4885]: I1002 02:27:26.068427 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="08919919-e9d1-4980-8c89-2f3891f3c576" path="/var/lib/kubelet/pods/08919919-e9d1-4980-8c89-2f3891f3c576/volumes" Oct 02 02:27:32 crc kubenswrapper[4885]: I1002 02:27:32.066521 4885 scope.go:117] "RemoveContainer" containerID="7d1be267228233dc9433f75e76a5a805aa23d98aa3b0edbd385a1a565a0d3231" Oct 02 02:27:32 crc kubenswrapper[4885]: E1002 02:27:32.070913 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:27:47 crc kubenswrapper[4885]: I1002 02:27:47.048355 4885 scope.go:117] "RemoveContainer" containerID="7d1be267228233dc9433f75e76a5a805aa23d98aa3b0edbd385a1a565a0d3231" Oct 02 02:27:47 crc kubenswrapper[4885]: E1002 02:27:47.049540 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:27:59 crc kubenswrapper[4885]: I1002 02:27:59.047972 4885 scope.go:117] "RemoveContainer" containerID="7d1be267228233dc9433f75e76a5a805aa23d98aa3b0edbd385a1a565a0d3231" Oct 02 02:27:59 crc kubenswrapper[4885]: E1002 02:27:59.049304 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:28:10 crc kubenswrapper[4885]: I1002 02:28:10.050088 4885 scope.go:117] "RemoveContainer" containerID="7d1be267228233dc9433f75e76a5a805aa23d98aa3b0edbd385a1a565a0d3231" Oct 02 02:28:10 crc kubenswrapper[4885]: E1002 02:28:10.051038 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:28:24 crc kubenswrapper[4885]: I1002 02:28:24.046964 4885 scope.go:117] "RemoveContainer" containerID="7d1be267228233dc9433f75e76a5a805aa23d98aa3b0edbd385a1a565a0d3231" Oct 02 02:28:24 crc kubenswrapper[4885]: E1002 02:28:24.047787 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:28:37 crc kubenswrapper[4885]: I1002 02:28:37.047509 4885 scope.go:117] "RemoveContainer" containerID="7d1be267228233dc9433f75e76a5a805aa23d98aa3b0edbd385a1a565a0d3231" Oct 02 02:28:37 crc kubenswrapper[4885]: E1002 02:28:37.048664 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:28:50 crc kubenswrapper[4885]: I1002 02:28:50.047646 4885 scope.go:117] "RemoveContainer" containerID="7d1be267228233dc9433f75e76a5a805aa23d98aa3b0edbd385a1a565a0d3231" Oct 02 02:28:50 crc kubenswrapper[4885]: E1002 02:28:50.048870 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:29:05 crc kubenswrapper[4885]: I1002 02:29:05.047177 4885 scope.go:117] "RemoveContainer" containerID="7d1be267228233dc9433f75e76a5a805aa23d98aa3b0edbd385a1a565a0d3231" Oct 02 02:29:05 crc kubenswrapper[4885]: E1002 02:29:05.048209 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:29:16 crc kubenswrapper[4885]: I1002 02:29:16.047052 4885 scope.go:117] "RemoveContainer" containerID="7d1be267228233dc9433f75e76a5a805aa23d98aa3b0edbd385a1a565a0d3231" Oct 02 02:29:16 crc kubenswrapper[4885]: I1002 02:29:16.881028 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" event={"ID":"d169c8ba-000a-4218-8e3b-4ae53035b110","Type":"ContainerStarted","Data":"5bbcad619e10a916cd09b9cfd4ad5c9334559ab939f077b917dc9ae849fe425f"} Oct 02 02:29:25 crc kubenswrapper[4885]: I1002 02:29:25.988120 4885 generic.go:334] "Generic (PLEG): container finished" podID="14bdd01b-529b-45b4-9ac3-188fdf3d3855" containerID="07741622db80033b74c21df5824568f5d17a2b20d14e3c8c1d8df8c07b24347d" exitCode=0 Oct 02 02:29:25 crc kubenswrapper[4885]: I1002 02:29:25.988253 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-c6gw4" event={"ID":"14bdd01b-529b-45b4-9ac3-188fdf3d3855","Type":"ContainerDied","Data":"07741622db80033b74c21df5824568f5d17a2b20d14e3c8c1d8df8c07b24347d"} Oct 02 02:29:27 crc kubenswrapper[4885]: I1002 02:29:27.509565 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-c6gw4" Oct 02 02:29:27 crc kubenswrapper[4885]: I1002 02:29:27.641529 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/14bdd01b-529b-45b4-9ac3-188fdf3d3855-nova-migration-ssh-key-1\") pod \"14bdd01b-529b-45b4-9ac3-188fdf3d3855\" (UID: \"14bdd01b-529b-45b4-9ac3-188fdf3d3855\") " Oct 02 02:29:27 crc kubenswrapper[4885]: I1002 02:29:27.641799 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/14bdd01b-529b-45b4-9ac3-188fdf3d3855-nova-migration-ssh-key-0\") pod \"14bdd01b-529b-45b4-9ac3-188fdf3d3855\" (UID: \"14bdd01b-529b-45b4-9ac3-188fdf3d3855\") " Oct 02 02:29:27 crc kubenswrapper[4885]: I1002 02:29:27.641847 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14bdd01b-529b-45b4-9ac3-188fdf3d3855-nova-combined-ca-bundle\") pod \"14bdd01b-529b-45b4-9ac3-188fdf3d3855\" (UID: \"14bdd01b-529b-45b4-9ac3-188fdf3d3855\") " Oct 02 02:29:27 crc kubenswrapper[4885]: I1002 02:29:27.641920 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/14bdd01b-529b-45b4-9ac3-188fdf3d3855-nova-cell1-compute-config-0\") pod \"14bdd01b-529b-45b4-9ac3-188fdf3d3855\" (UID: \"14bdd01b-529b-45b4-9ac3-188fdf3d3855\") " Oct 02 02:29:27 crc kubenswrapper[4885]: I1002 02:29:27.641942 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/14bdd01b-529b-45b4-9ac3-188fdf3d3855-ssh-key\") pod \"14bdd01b-529b-45b4-9ac3-188fdf3d3855\" (UID: \"14bdd01b-529b-45b4-9ac3-188fdf3d3855\") " Oct 02 02:29:27 crc kubenswrapper[4885]: I1002 02:29:27.641982 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v8q27\" (UniqueName: \"kubernetes.io/projected/14bdd01b-529b-45b4-9ac3-188fdf3d3855-kube-api-access-v8q27\") pod \"14bdd01b-529b-45b4-9ac3-188fdf3d3855\" (UID: \"14bdd01b-529b-45b4-9ac3-188fdf3d3855\") " Oct 02 02:29:27 crc kubenswrapper[4885]: I1002 02:29:27.642045 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/14bdd01b-529b-45b4-9ac3-188fdf3d3855-nova-cell1-compute-config-1\") pod \"14bdd01b-529b-45b4-9ac3-188fdf3d3855\" (UID: \"14bdd01b-529b-45b4-9ac3-188fdf3d3855\") " Oct 02 02:29:27 crc kubenswrapper[4885]: I1002 02:29:27.642108 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/14bdd01b-529b-45b4-9ac3-188fdf3d3855-inventory\") pod \"14bdd01b-529b-45b4-9ac3-188fdf3d3855\" (UID: \"14bdd01b-529b-45b4-9ac3-188fdf3d3855\") " Oct 02 02:29:27 crc kubenswrapper[4885]: I1002 02:29:27.642128 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/14bdd01b-529b-45b4-9ac3-188fdf3d3855-nova-extra-config-0\") pod \"14bdd01b-529b-45b4-9ac3-188fdf3d3855\" (UID: \"14bdd01b-529b-45b4-9ac3-188fdf3d3855\") " Oct 02 02:29:27 crc kubenswrapper[4885]: I1002 02:29:27.671603 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14bdd01b-529b-45b4-9ac3-188fdf3d3855-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "14bdd01b-529b-45b4-9ac3-188fdf3d3855" (UID: "14bdd01b-529b-45b4-9ac3-188fdf3d3855"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:29:27 crc kubenswrapper[4885]: I1002 02:29:27.681086 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14bdd01b-529b-45b4-9ac3-188fdf3d3855-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "14bdd01b-529b-45b4-9ac3-188fdf3d3855" (UID: "14bdd01b-529b-45b4-9ac3-188fdf3d3855"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:29:27 crc kubenswrapper[4885]: I1002 02:29:27.686712 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/14bdd01b-529b-45b4-9ac3-188fdf3d3855-kube-api-access-v8q27" (OuterVolumeSpecName: "kube-api-access-v8q27") pod "14bdd01b-529b-45b4-9ac3-188fdf3d3855" (UID: "14bdd01b-529b-45b4-9ac3-188fdf3d3855"). InnerVolumeSpecName "kube-api-access-v8q27". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:29:27 crc kubenswrapper[4885]: I1002 02:29:27.688057 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14bdd01b-529b-45b4-9ac3-188fdf3d3855-inventory" (OuterVolumeSpecName: "inventory") pod "14bdd01b-529b-45b4-9ac3-188fdf3d3855" (UID: "14bdd01b-529b-45b4-9ac3-188fdf3d3855"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:29:27 crc kubenswrapper[4885]: I1002 02:29:27.688867 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14bdd01b-529b-45b4-9ac3-188fdf3d3855-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "14bdd01b-529b-45b4-9ac3-188fdf3d3855" (UID: "14bdd01b-529b-45b4-9ac3-188fdf3d3855"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:29:27 crc kubenswrapper[4885]: I1002 02:29:27.689654 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14bdd01b-529b-45b4-9ac3-188fdf3d3855-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "14bdd01b-529b-45b4-9ac3-188fdf3d3855" (UID: "14bdd01b-529b-45b4-9ac3-188fdf3d3855"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:29:27 crc kubenswrapper[4885]: I1002 02:29:27.696858 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14bdd01b-529b-45b4-9ac3-188fdf3d3855-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "14bdd01b-529b-45b4-9ac3-188fdf3d3855" (UID: "14bdd01b-529b-45b4-9ac3-188fdf3d3855"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:29:27 crc kubenswrapper[4885]: I1002 02:29:27.708110 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14bdd01b-529b-45b4-9ac3-188fdf3d3855-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "14bdd01b-529b-45b4-9ac3-188fdf3d3855" (UID: "14bdd01b-529b-45b4-9ac3-188fdf3d3855"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:29:27 crc kubenswrapper[4885]: I1002 02:29:27.718678 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14bdd01b-529b-45b4-9ac3-188fdf3d3855-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "14bdd01b-529b-45b4-9ac3-188fdf3d3855" (UID: "14bdd01b-529b-45b4-9ac3-188fdf3d3855"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:29:27 crc kubenswrapper[4885]: I1002 02:29:27.744828 4885 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/14bdd01b-529b-45b4-9ac3-188fdf3d3855-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 02:29:27 crc kubenswrapper[4885]: I1002 02:29:27.744887 4885 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/14bdd01b-529b-45b4-9ac3-188fdf3d3855-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Oct 02 02:29:27 crc kubenswrapper[4885]: I1002 02:29:27.744905 4885 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/14bdd01b-529b-45b4-9ac3-188fdf3d3855-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Oct 02 02:29:27 crc kubenswrapper[4885]: I1002 02:29:27.744917 4885 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/14bdd01b-529b-45b4-9ac3-188fdf3d3855-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Oct 02 02:29:27 crc kubenswrapper[4885]: I1002 02:29:27.744929 4885 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14bdd01b-529b-45b4-9ac3-188fdf3d3855-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:29:27 crc kubenswrapper[4885]: I1002 02:29:27.744941 4885 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/14bdd01b-529b-45b4-9ac3-188fdf3d3855-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Oct 02 02:29:27 crc kubenswrapper[4885]: I1002 02:29:27.744996 4885 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/14bdd01b-529b-45b4-9ac3-188fdf3d3855-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 02:29:27 crc kubenswrapper[4885]: I1002 02:29:27.745010 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v8q27\" (UniqueName: \"kubernetes.io/projected/14bdd01b-529b-45b4-9ac3-188fdf3d3855-kube-api-access-v8q27\") on node \"crc\" DevicePath \"\"" Oct 02 02:29:27 crc kubenswrapper[4885]: I1002 02:29:27.745023 4885 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/14bdd01b-529b-45b4-9ac3-188fdf3d3855-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Oct 02 02:29:28 crc kubenswrapper[4885]: I1002 02:29:28.015132 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-c6gw4" event={"ID":"14bdd01b-529b-45b4-9ac3-188fdf3d3855","Type":"ContainerDied","Data":"628e9325eb6d394d488080264309a506e38c2a97b36e681bba0ffb18d52a9c51"} Oct 02 02:29:28 crc kubenswrapper[4885]: I1002 02:29:28.015169 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-c6gw4" Oct 02 02:29:28 crc kubenswrapper[4885]: I1002 02:29:28.015175 4885 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="628e9325eb6d394d488080264309a506e38c2a97b36e681bba0ffb18d52a9c51" Oct 02 02:29:28 crc kubenswrapper[4885]: I1002 02:29:28.142967 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v9h2z"] Oct 02 02:29:28 crc kubenswrapper[4885]: E1002 02:29:28.143439 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08919919-e9d1-4980-8c89-2f3891f3c576" containerName="registry-server" Oct 02 02:29:28 crc kubenswrapper[4885]: I1002 02:29:28.143470 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="08919919-e9d1-4980-8c89-2f3891f3c576" containerName="registry-server" Oct 02 02:29:28 crc kubenswrapper[4885]: E1002 02:29:28.143491 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08919919-e9d1-4980-8c89-2f3891f3c576" containerName="extract-utilities" Oct 02 02:29:28 crc kubenswrapper[4885]: I1002 02:29:28.143499 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="08919919-e9d1-4980-8c89-2f3891f3c576" containerName="extract-utilities" Oct 02 02:29:28 crc kubenswrapper[4885]: E1002 02:29:28.143519 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08919919-e9d1-4980-8c89-2f3891f3c576" containerName="extract-content" Oct 02 02:29:28 crc kubenswrapper[4885]: I1002 02:29:28.143526 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="08919919-e9d1-4980-8c89-2f3891f3c576" containerName="extract-content" Oct 02 02:29:28 crc kubenswrapper[4885]: E1002 02:29:28.143556 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14bdd01b-529b-45b4-9ac3-188fdf3d3855" containerName="nova-edpm-deployment-openstack-edpm-ipam" Oct 02 02:29:28 crc kubenswrapper[4885]: I1002 02:29:28.143565 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="14bdd01b-529b-45b4-9ac3-188fdf3d3855" containerName="nova-edpm-deployment-openstack-edpm-ipam" Oct 02 02:29:28 crc kubenswrapper[4885]: I1002 02:29:28.143801 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="08919919-e9d1-4980-8c89-2f3891f3c576" containerName="registry-server" Oct 02 02:29:28 crc kubenswrapper[4885]: I1002 02:29:28.143827 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="14bdd01b-529b-45b4-9ac3-188fdf3d3855" containerName="nova-edpm-deployment-openstack-edpm-ipam" Oct 02 02:29:28 crc kubenswrapper[4885]: I1002 02:29:28.144606 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v9h2z" Oct 02 02:29:28 crc kubenswrapper[4885]: I1002 02:29:28.147152 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 02:29:28 crc kubenswrapper[4885]: I1002 02:29:28.147639 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Oct 02 02:29:28 crc kubenswrapper[4885]: I1002 02:29:28.147793 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 02:29:28 crc kubenswrapper[4885]: I1002 02:29:28.147877 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-f8tvb" Oct 02 02:29:28 crc kubenswrapper[4885]: I1002 02:29:28.148222 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 02:29:28 crc kubenswrapper[4885]: I1002 02:29:28.164455 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v9h2z"] Oct 02 02:29:28 crc kubenswrapper[4885]: I1002 02:29:28.254588 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/ce63c1c1-34cc-4cba-a553-33b66431bc3c-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v9h2z\" (UID: \"ce63c1c1-34cc-4cba-a553-33b66431bc3c\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v9h2z" Oct 02 02:29:28 crc kubenswrapper[4885]: I1002 02:29:28.254729 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/ce63c1c1-34cc-4cba-a553-33b66431bc3c-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v9h2z\" (UID: \"ce63c1c1-34cc-4cba-a553-33b66431bc3c\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v9h2z" Oct 02 02:29:28 crc kubenswrapper[4885]: I1002 02:29:28.254755 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gq98j\" (UniqueName: \"kubernetes.io/projected/ce63c1c1-34cc-4cba-a553-33b66431bc3c-kube-api-access-gq98j\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v9h2z\" (UID: \"ce63c1c1-34cc-4cba-a553-33b66431bc3c\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v9h2z" Oct 02 02:29:28 crc kubenswrapper[4885]: I1002 02:29:28.254786 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ce63c1c1-34cc-4cba-a553-33b66431bc3c-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v9h2z\" (UID: \"ce63c1c1-34cc-4cba-a553-33b66431bc3c\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v9h2z" Oct 02 02:29:28 crc kubenswrapper[4885]: I1002 02:29:28.254830 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce63c1c1-34cc-4cba-a553-33b66431bc3c-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v9h2z\" (UID: \"ce63c1c1-34cc-4cba-a553-33b66431bc3c\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v9h2z" Oct 02 02:29:28 crc kubenswrapper[4885]: I1002 02:29:28.254960 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/ce63c1c1-34cc-4cba-a553-33b66431bc3c-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v9h2z\" (UID: \"ce63c1c1-34cc-4cba-a553-33b66431bc3c\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v9h2z" Oct 02 02:29:28 crc kubenswrapper[4885]: I1002 02:29:28.255014 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ce63c1c1-34cc-4cba-a553-33b66431bc3c-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v9h2z\" (UID: \"ce63c1c1-34cc-4cba-a553-33b66431bc3c\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v9h2z" Oct 02 02:29:28 crc kubenswrapper[4885]: I1002 02:29:28.357589 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce63c1c1-34cc-4cba-a553-33b66431bc3c-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v9h2z\" (UID: \"ce63c1c1-34cc-4cba-a553-33b66431bc3c\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v9h2z" Oct 02 02:29:28 crc kubenswrapper[4885]: I1002 02:29:28.357716 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/ce63c1c1-34cc-4cba-a553-33b66431bc3c-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v9h2z\" (UID: \"ce63c1c1-34cc-4cba-a553-33b66431bc3c\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v9h2z" Oct 02 02:29:28 crc kubenswrapper[4885]: I1002 02:29:28.357785 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ce63c1c1-34cc-4cba-a553-33b66431bc3c-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v9h2z\" (UID: \"ce63c1c1-34cc-4cba-a553-33b66431bc3c\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v9h2z" Oct 02 02:29:28 crc kubenswrapper[4885]: I1002 02:29:28.357942 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/ce63c1c1-34cc-4cba-a553-33b66431bc3c-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v9h2z\" (UID: \"ce63c1c1-34cc-4cba-a553-33b66431bc3c\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v9h2z" Oct 02 02:29:28 crc kubenswrapper[4885]: I1002 02:29:28.358064 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/ce63c1c1-34cc-4cba-a553-33b66431bc3c-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v9h2z\" (UID: \"ce63c1c1-34cc-4cba-a553-33b66431bc3c\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v9h2z" Oct 02 02:29:28 crc kubenswrapper[4885]: I1002 02:29:28.358108 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gq98j\" (UniqueName: \"kubernetes.io/projected/ce63c1c1-34cc-4cba-a553-33b66431bc3c-kube-api-access-gq98j\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v9h2z\" (UID: \"ce63c1c1-34cc-4cba-a553-33b66431bc3c\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v9h2z" Oct 02 02:29:28 crc kubenswrapper[4885]: I1002 02:29:28.358149 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ce63c1c1-34cc-4cba-a553-33b66431bc3c-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v9h2z\" (UID: \"ce63c1c1-34cc-4cba-a553-33b66431bc3c\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v9h2z" Oct 02 02:29:28 crc kubenswrapper[4885]: I1002 02:29:28.364091 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/ce63c1c1-34cc-4cba-a553-33b66431bc3c-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v9h2z\" (UID: \"ce63c1c1-34cc-4cba-a553-33b66431bc3c\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v9h2z" Oct 02 02:29:28 crc kubenswrapper[4885]: I1002 02:29:28.364616 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ce63c1c1-34cc-4cba-a553-33b66431bc3c-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v9h2z\" (UID: \"ce63c1c1-34cc-4cba-a553-33b66431bc3c\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v9h2z" Oct 02 02:29:28 crc kubenswrapper[4885]: I1002 02:29:28.364873 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce63c1c1-34cc-4cba-a553-33b66431bc3c-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v9h2z\" (UID: \"ce63c1c1-34cc-4cba-a553-33b66431bc3c\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v9h2z" Oct 02 02:29:28 crc kubenswrapper[4885]: I1002 02:29:28.366715 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/ce63c1c1-34cc-4cba-a553-33b66431bc3c-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v9h2z\" (UID: \"ce63c1c1-34cc-4cba-a553-33b66431bc3c\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v9h2z" Oct 02 02:29:28 crc kubenswrapper[4885]: I1002 02:29:28.367099 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ce63c1c1-34cc-4cba-a553-33b66431bc3c-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v9h2z\" (UID: \"ce63c1c1-34cc-4cba-a553-33b66431bc3c\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v9h2z" Oct 02 02:29:28 crc kubenswrapper[4885]: I1002 02:29:28.369812 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/ce63c1c1-34cc-4cba-a553-33b66431bc3c-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v9h2z\" (UID: \"ce63c1c1-34cc-4cba-a553-33b66431bc3c\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v9h2z" Oct 02 02:29:28 crc kubenswrapper[4885]: I1002 02:29:28.381024 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gq98j\" (UniqueName: \"kubernetes.io/projected/ce63c1c1-34cc-4cba-a553-33b66431bc3c-kube-api-access-gq98j\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v9h2z\" (UID: \"ce63c1c1-34cc-4cba-a553-33b66431bc3c\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v9h2z" Oct 02 02:29:28 crc kubenswrapper[4885]: I1002 02:29:28.470551 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v9h2z" Oct 02 02:29:28 crc kubenswrapper[4885]: I1002 02:29:28.850248 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v9h2z"] Oct 02 02:29:28 crc kubenswrapper[4885]: I1002 02:29:28.860225 4885 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 02:29:29 crc kubenswrapper[4885]: I1002 02:29:29.045461 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v9h2z" event={"ID":"ce63c1c1-34cc-4cba-a553-33b66431bc3c","Type":"ContainerStarted","Data":"83f63a436cd81c26858faa49959192a7d875abca3cdb13cceb7d139a7ef8552c"} Oct 02 02:29:38 crc kubenswrapper[4885]: E1002 02:29:38.972523 4885 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://quay.io/openstack-k8s-operators/openstack-ansibleee-runner:latest: Requesting bearer token: invalid status code from registry 504 (Gateway Timeout)" image="quay.io/openstack-k8s-operators/openstack-ansibleee-runner:latest" Oct 02 02:29:38 crc kubenswrapper[4885]: E1002 02:29:38.973051 4885 kuberuntime_manager.go:1274] "Unhandled Error" err=< Oct 02 02:29:38 crc kubenswrapper[4885]: container &Container{Name:telemetry-edpm-deployment-openstack-edpm-ipam,Image:quay.io/openstack-k8s-operators/openstack-ansibleee-runner:latest,Command:[],Args:[ansible-runner run /runner -p osp.edpm.telemetry -i telemetry-edpm-deployment-openstack-edpm-ipam],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:ANSIBLE_VERBOSITY,Value:2,ValueFrom:nil,},EnvVar{Name:RUNNER_PLAYBOOK,Value: Oct 02 02:29:38 crc kubenswrapper[4885]: osp.edpm.telemetry Oct 02 02:29:38 crc kubenswrapper[4885]: Oct 02 02:29:38 crc kubenswrapper[4885]: ,ValueFrom:nil,},EnvVar{Name:RUNNER_EXTRA_VARS,Value: Oct 02 02:29:38 crc kubenswrapper[4885]: edpm_override_hosts: openstack-edpm-ipam Oct 02 02:29:38 crc kubenswrapper[4885]: edpm_service_type: telemetry Oct 02 02:29:38 crc kubenswrapper[4885]: Oct 02 02:29:38 crc kubenswrapper[4885]: Oct 02 02:29:38 crc kubenswrapper[4885]: ,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:ceilometer-compute-config-data-0,ReadOnly:false,MountPath:/var/lib/openstack/configs/telemetry/ceilometer.conf,SubPath:ceilometer.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ceilometer-compute-config-data-1,ReadOnly:false,MountPath:/var/lib/openstack/configs/telemetry/custom.conf,SubPath:custom.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ceilometer-compute-config-data-2,ReadOnly:false,MountPath:/var/lib/openstack/configs/telemetry/polling.yaml,SubPath:polling.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:telemetry-combined-ca-bundle,ReadOnly:false,MountPath:/var/lib/openstack/cacerts/telemetry,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/runner/env/ssh_key,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:inventory,ReadOnly:false,MountPath:/runner/inventory/hosts,SubPath:inventory,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gq98j,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:openstack-aee-default-env,},Optional:*true,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-edpm-deployment-openstack-edpm-ipam-v9h2z_openstack(ce63c1c1-34cc-4cba-a553-33b66431bc3c): ErrImagePull: initializing source docker://quay.io/openstack-k8s-operators/openstack-ansibleee-runner:latest: Requesting bearer token: invalid status code from registry 504 (Gateway Timeout) Oct 02 02:29:38 crc kubenswrapper[4885]: > logger="UnhandledError" Oct 02 02:29:38 crc kubenswrapper[4885]: E1002 02:29:38.974211 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"telemetry-edpm-deployment-openstack-edpm-ipam\" with ErrImagePull: \"initializing source docker://quay.io/openstack-k8s-operators/openstack-ansibleee-runner:latest: Requesting bearer token: invalid status code from registry 504 (Gateway Timeout)\"" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v9h2z" podUID="ce63c1c1-34cc-4cba-a553-33b66431bc3c" Oct 02 02:29:39 crc kubenswrapper[4885]: E1002 02:29:39.195077 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"telemetry-edpm-deployment-openstack-edpm-ipam\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-ansibleee-runner:latest\\\"\"" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v9h2z" podUID="ce63c1c1-34cc-4cba-a553-33b66431bc3c" Oct 02 02:29:47 crc kubenswrapper[4885]: I1002 02:29:47.698122 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-mrhcc"] Oct 02 02:29:47 crc kubenswrapper[4885]: I1002 02:29:47.702859 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mrhcc" Oct 02 02:29:47 crc kubenswrapper[4885]: I1002 02:29:47.712703 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mrhcc"] Oct 02 02:29:47 crc kubenswrapper[4885]: I1002 02:29:47.788794 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e45a508f-eb2f-45e6-875b-f877bb85101d-utilities\") pod \"certified-operators-mrhcc\" (UID: \"e45a508f-eb2f-45e6-875b-f877bb85101d\") " pod="openshift-marketplace/certified-operators-mrhcc" Oct 02 02:29:47 crc kubenswrapper[4885]: I1002 02:29:47.788855 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kpgm8\" (UniqueName: \"kubernetes.io/projected/e45a508f-eb2f-45e6-875b-f877bb85101d-kube-api-access-kpgm8\") pod \"certified-operators-mrhcc\" (UID: \"e45a508f-eb2f-45e6-875b-f877bb85101d\") " pod="openshift-marketplace/certified-operators-mrhcc" Oct 02 02:29:47 crc kubenswrapper[4885]: I1002 02:29:47.789208 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e45a508f-eb2f-45e6-875b-f877bb85101d-catalog-content\") pod \"certified-operators-mrhcc\" (UID: \"e45a508f-eb2f-45e6-875b-f877bb85101d\") " pod="openshift-marketplace/certified-operators-mrhcc" Oct 02 02:29:47 crc kubenswrapper[4885]: I1002 02:29:47.891009 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e45a508f-eb2f-45e6-875b-f877bb85101d-utilities\") pod \"certified-operators-mrhcc\" (UID: \"e45a508f-eb2f-45e6-875b-f877bb85101d\") " pod="openshift-marketplace/certified-operators-mrhcc" Oct 02 02:29:47 crc kubenswrapper[4885]: I1002 02:29:47.891075 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kpgm8\" (UniqueName: \"kubernetes.io/projected/e45a508f-eb2f-45e6-875b-f877bb85101d-kube-api-access-kpgm8\") pod \"certified-operators-mrhcc\" (UID: \"e45a508f-eb2f-45e6-875b-f877bb85101d\") " pod="openshift-marketplace/certified-operators-mrhcc" Oct 02 02:29:47 crc kubenswrapper[4885]: I1002 02:29:47.891188 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e45a508f-eb2f-45e6-875b-f877bb85101d-catalog-content\") pod \"certified-operators-mrhcc\" (UID: \"e45a508f-eb2f-45e6-875b-f877bb85101d\") " pod="openshift-marketplace/certified-operators-mrhcc" Oct 02 02:29:47 crc kubenswrapper[4885]: I1002 02:29:47.891530 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e45a508f-eb2f-45e6-875b-f877bb85101d-utilities\") pod \"certified-operators-mrhcc\" (UID: \"e45a508f-eb2f-45e6-875b-f877bb85101d\") " pod="openshift-marketplace/certified-operators-mrhcc" Oct 02 02:29:47 crc kubenswrapper[4885]: I1002 02:29:47.891696 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e45a508f-eb2f-45e6-875b-f877bb85101d-catalog-content\") pod \"certified-operators-mrhcc\" (UID: \"e45a508f-eb2f-45e6-875b-f877bb85101d\") " pod="openshift-marketplace/certified-operators-mrhcc" Oct 02 02:29:47 crc kubenswrapper[4885]: I1002 02:29:47.919634 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kpgm8\" (UniqueName: \"kubernetes.io/projected/e45a508f-eb2f-45e6-875b-f877bb85101d-kube-api-access-kpgm8\") pod \"certified-operators-mrhcc\" (UID: \"e45a508f-eb2f-45e6-875b-f877bb85101d\") " pod="openshift-marketplace/certified-operators-mrhcc" Oct 02 02:29:48 crc kubenswrapper[4885]: I1002 02:29:48.035820 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mrhcc" Oct 02 02:29:48 crc kubenswrapper[4885]: I1002 02:29:48.598289 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mrhcc"] Oct 02 02:29:48 crc kubenswrapper[4885]: W1002 02:29:48.601154 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode45a508f_eb2f_45e6_875b_f877bb85101d.slice/crio-5a182b798ea41b5d54c86231a97279edd09ccc8e7322c7d2cb33576632523de8 WatchSource:0}: Error finding container 5a182b798ea41b5d54c86231a97279edd09ccc8e7322c7d2cb33576632523de8: Status 404 returned error can't find the container with id 5a182b798ea41b5d54c86231a97279edd09ccc8e7322c7d2cb33576632523de8 Oct 02 02:29:49 crc kubenswrapper[4885]: I1002 02:29:49.314543 4885 generic.go:334] "Generic (PLEG): container finished" podID="e45a508f-eb2f-45e6-875b-f877bb85101d" containerID="5fda54bb5c79929ec4e8cf0f421a8d466ed253b2e4bd6a3bc103b032de90cab5" exitCode=0 Oct 02 02:29:49 crc kubenswrapper[4885]: I1002 02:29:49.314627 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mrhcc" event={"ID":"e45a508f-eb2f-45e6-875b-f877bb85101d","Type":"ContainerDied","Data":"5fda54bb5c79929ec4e8cf0f421a8d466ed253b2e4bd6a3bc103b032de90cab5"} Oct 02 02:29:49 crc kubenswrapper[4885]: I1002 02:29:49.314895 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mrhcc" event={"ID":"e45a508f-eb2f-45e6-875b-f877bb85101d","Type":"ContainerStarted","Data":"5a182b798ea41b5d54c86231a97279edd09ccc8e7322c7d2cb33576632523de8"} Oct 02 02:29:51 crc kubenswrapper[4885]: I1002 02:29:51.344587 4885 generic.go:334] "Generic (PLEG): container finished" podID="e45a508f-eb2f-45e6-875b-f877bb85101d" containerID="5041923797db7b24a57bf7d30e90a1eef25187cf83a34edfeef39345ab2c10d6" exitCode=0 Oct 02 02:29:51 crc kubenswrapper[4885]: I1002 02:29:51.344662 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mrhcc" event={"ID":"e45a508f-eb2f-45e6-875b-f877bb85101d","Type":"ContainerDied","Data":"5041923797db7b24a57bf7d30e90a1eef25187cf83a34edfeef39345ab2c10d6"} Oct 02 02:29:52 crc kubenswrapper[4885]: I1002 02:29:52.358010 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v9h2z" event={"ID":"ce63c1c1-34cc-4cba-a553-33b66431bc3c","Type":"ContainerStarted","Data":"9c92bcdaf1b3ed3e96e5851ff9f76d5409006ad36b9c86452bae73704fc96a9c"} Oct 02 02:29:52 crc kubenswrapper[4885]: I1002 02:29:52.380540 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v9h2z" podStartSLOduration=1.563269733 podStartE2EDuration="24.380521215s" podCreationTimestamp="2025-10-02 02:29:28 +0000 UTC" firstStartedPulling="2025-10-02 02:29:28.859737508 +0000 UTC m=+2557.671484927" lastFinishedPulling="2025-10-02 02:29:51.67698897 +0000 UTC m=+2580.488736409" observedRunningTime="2025-10-02 02:29:52.379942507 +0000 UTC m=+2581.191689946" watchObservedRunningTime="2025-10-02 02:29:52.380521215 +0000 UTC m=+2581.192268604" Oct 02 02:29:53 crc kubenswrapper[4885]: I1002 02:29:53.372576 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mrhcc" event={"ID":"e45a508f-eb2f-45e6-875b-f877bb85101d","Type":"ContainerStarted","Data":"b055bae7125b0ef7fca54e932015c937bb8ddba6c863be9f5a1e76ecbb06426b"} Oct 02 02:29:53 crc kubenswrapper[4885]: I1002 02:29:53.409153 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-mrhcc" podStartSLOduration=3.6990487979999997 podStartE2EDuration="6.409119824s" podCreationTimestamp="2025-10-02 02:29:47 +0000 UTC" firstStartedPulling="2025-10-02 02:29:49.316931507 +0000 UTC m=+2578.128678916" lastFinishedPulling="2025-10-02 02:29:52.027002543 +0000 UTC m=+2580.838749942" observedRunningTime="2025-10-02 02:29:53.406351449 +0000 UTC m=+2582.218098878" watchObservedRunningTime="2025-10-02 02:29:53.409119824 +0000 UTC m=+2582.220867263" Oct 02 02:29:58 crc kubenswrapper[4885]: I1002 02:29:58.036293 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-mrhcc" Oct 02 02:29:58 crc kubenswrapper[4885]: I1002 02:29:58.037240 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-mrhcc" Oct 02 02:29:58 crc kubenswrapper[4885]: I1002 02:29:58.125586 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-mrhcc" Oct 02 02:29:58 crc kubenswrapper[4885]: I1002 02:29:58.472713 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-mrhcc" Oct 02 02:29:59 crc kubenswrapper[4885]: I1002 02:29:59.309984 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mrhcc"] Oct 02 02:30:00 crc kubenswrapper[4885]: I1002 02:30:00.161012 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322870-cbkw6"] Oct 02 02:30:00 crc kubenswrapper[4885]: I1002 02:30:00.164196 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322870-cbkw6" Oct 02 02:30:00 crc kubenswrapper[4885]: I1002 02:30:00.166525 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 02 02:30:00 crc kubenswrapper[4885]: I1002 02:30:00.166808 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 02 02:30:00 crc kubenswrapper[4885]: I1002 02:30:00.169369 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322870-cbkw6"] Oct 02 02:30:00 crc kubenswrapper[4885]: I1002 02:30:00.317540 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/126b33ad-f456-4e9a-ad7a-f16ba2613e9b-secret-volume\") pod \"collect-profiles-29322870-cbkw6\" (UID: \"126b33ad-f456-4e9a-ad7a-f16ba2613e9b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322870-cbkw6" Oct 02 02:30:00 crc kubenswrapper[4885]: I1002 02:30:00.317597 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/126b33ad-f456-4e9a-ad7a-f16ba2613e9b-config-volume\") pod \"collect-profiles-29322870-cbkw6\" (UID: \"126b33ad-f456-4e9a-ad7a-f16ba2613e9b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322870-cbkw6" Oct 02 02:30:00 crc kubenswrapper[4885]: I1002 02:30:00.317661 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nslc2\" (UniqueName: \"kubernetes.io/projected/126b33ad-f456-4e9a-ad7a-f16ba2613e9b-kube-api-access-nslc2\") pod \"collect-profiles-29322870-cbkw6\" (UID: \"126b33ad-f456-4e9a-ad7a-f16ba2613e9b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322870-cbkw6" Oct 02 02:30:00 crc kubenswrapper[4885]: I1002 02:30:00.419756 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/126b33ad-f456-4e9a-ad7a-f16ba2613e9b-secret-volume\") pod \"collect-profiles-29322870-cbkw6\" (UID: \"126b33ad-f456-4e9a-ad7a-f16ba2613e9b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322870-cbkw6" Oct 02 02:30:00 crc kubenswrapper[4885]: I1002 02:30:00.419800 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/126b33ad-f456-4e9a-ad7a-f16ba2613e9b-config-volume\") pod \"collect-profiles-29322870-cbkw6\" (UID: \"126b33ad-f456-4e9a-ad7a-f16ba2613e9b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322870-cbkw6" Oct 02 02:30:00 crc kubenswrapper[4885]: I1002 02:30:00.419850 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nslc2\" (UniqueName: \"kubernetes.io/projected/126b33ad-f456-4e9a-ad7a-f16ba2613e9b-kube-api-access-nslc2\") pod \"collect-profiles-29322870-cbkw6\" (UID: \"126b33ad-f456-4e9a-ad7a-f16ba2613e9b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322870-cbkw6" Oct 02 02:30:00 crc kubenswrapper[4885]: I1002 02:30:00.421801 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/126b33ad-f456-4e9a-ad7a-f16ba2613e9b-config-volume\") pod \"collect-profiles-29322870-cbkw6\" (UID: \"126b33ad-f456-4e9a-ad7a-f16ba2613e9b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322870-cbkw6" Oct 02 02:30:00 crc kubenswrapper[4885]: I1002 02:30:00.429101 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/126b33ad-f456-4e9a-ad7a-f16ba2613e9b-secret-volume\") pod \"collect-profiles-29322870-cbkw6\" (UID: \"126b33ad-f456-4e9a-ad7a-f16ba2613e9b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322870-cbkw6" Oct 02 02:30:00 crc kubenswrapper[4885]: I1002 02:30:00.443521 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nslc2\" (UniqueName: \"kubernetes.io/projected/126b33ad-f456-4e9a-ad7a-f16ba2613e9b-kube-api-access-nslc2\") pod \"collect-profiles-29322870-cbkw6\" (UID: \"126b33ad-f456-4e9a-ad7a-f16ba2613e9b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322870-cbkw6" Oct 02 02:30:00 crc kubenswrapper[4885]: I1002 02:30:00.450053 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-mrhcc" podUID="e45a508f-eb2f-45e6-875b-f877bb85101d" containerName="registry-server" containerID="cri-o://b055bae7125b0ef7fca54e932015c937bb8ddba6c863be9f5a1e76ecbb06426b" gracePeriod=2 Oct 02 02:30:00 crc kubenswrapper[4885]: I1002 02:30:00.490237 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322870-cbkw6" Oct 02 02:30:00 crc kubenswrapper[4885]: I1002 02:30:00.975403 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322870-cbkw6"] Oct 02 02:30:01 crc kubenswrapper[4885]: I1002 02:30:01.461408 4885 generic.go:334] "Generic (PLEG): container finished" podID="e45a508f-eb2f-45e6-875b-f877bb85101d" containerID="b055bae7125b0ef7fca54e932015c937bb8ddba6c863be9f5a1e76ecbb06426b" exitCode=0 Oct 02 02:30:01 crc kubenswrapper[4885]: I1002 02:30:01.461505 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mrhcc" event={"ID":"e45a508f-eb2f-45e6-875b-f877bb85101d","Type":"ContainerDied","Data":"b055bae7125b0ef7fca54e932015c937bb8ddba6c863be9f5a1e76ecbb06426b"} Oct 02 02:30:01 crc kubenswrapper[4885]: I1002 02:30:01.461879 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mrhcc" event={"ID":"e45a508f-eb2f-45e6-875b-f877bb85101d","Type":"ContainerDied","Data":"5a182b798ea41b5d54c86231a97279edd09ccc8e7322c7d2cb33576632523de8"} Oct 02 02:30:01 crc kubenswrapper[4885]: I1002 02:30:01.461898 4885 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5a182b798ea41b5d54c86231a97279edd09ccc8e7322c7d2cb33576632523de8" Oct 02 02:30:01 crc kubenswrapper[4885]: I1002 02:30:01.463521 4885 generic.go:334] "Generic (PLEG): container finished" podID="126b33ad-f456-4e9a-ad7a-f16ba2613e9b" containerID="3fa24ddcd89632d856af118548ac3e8311eed61793c8598df3d07c0a87a2ed11" exitCode=0 Oct 02 02:30:01 crc kubenswrapper[4885]: I1002 02:30:01.463563 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29322870-cbkw6" event={"ID":"126b33ad-f456-4e9a-ad7a-f16ba2613e9b","Type":"ContainerDied","Data":"3fa24ddcd89632d856af118548ac3e8311eed61793c8598df3d07c0a87a2ed11"} Oct 02 02:30:01 crc kubenswrapper[4885]: I1002 02:30:01.463591 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29322870-cbkw6" event={"ID":"126b33ad-f456-4e9a-ad7a-f16ba2613e9b","Type":"ContainerStarted","Data":"ba07868150fd33f187bb9a1a63d50a806af81a1b6bd615d11c8173330bd4063d"} Oct 02 02:30:01 crc kubenswrapper[4885]: I1002 02:30:01.544594 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mrhcc" Oct 02 02:30:01 crc kubenswrapper[4885]: I1002 02:30:01.644430 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e45a508f-eb2f-45e6-875b-f877bb85101d-utilities\") pod \"e45a508f-eb2f-45e6-875b-f877bb85101d\" (UID: \"e45a508f-eb2f-45e6-875b-f877bb85101d\") " Oct 02 02:30:01 crc kubenswrapper[4885]: I1002 02:30:01.644530 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kpgm8\" (UniqueName: \"kubernetes.io/projected/e45a508f-eb2f-45e6-875b-f877bb85101d-kube-api-access-kpgm8\") pod \"e45a508f-eb2f-45e6-875b-f877bb85101d\" (UID: \"e45a508f-eb2f-45e6-875b-f877bb85101d\") " Oct 02 02:30:01 crc kubenswrapper[4885]: I1002 02:30:01.644800 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e45a508f-eb2f-45e6-875b-f877bb85101d-catalog-content\") pod \"e45a508f-eb2f-45e6-875b-f877bb85101d\" (UID: \"e45a508f-eb2f-45e6-875b-f877bb85101d\") " Oct 02 02:30:01 crc kubenswrapper[4885]: I1002 02:30:01.646181 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e45a508f-eb2f-45e6-875b-f877bb85101d-utilities" (OuterVolumeSpecName: "utilities") pod "e45a508f-eb2f-45e6-875b-f877bb85101d" (UID: "e45a508f-eb2f-45e6-875b-f877bb85101d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:30:01 crc kubenswrapper[4885]: I1002 02:30:01.651488 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e45a508f-eb2f-45e6-875b-f877bb85101d-kube-api-access-kpgm8" (OuterVolumeSpecName: "kube-api-access-kpgm8") pod "e45a508f-eb2f-45e6-875b-f877bb85101d" (UID: "e45a508f-eb2f-45e6-875b-f877bb85101d"). InnerVolumeSpecName "kube-api-access-kpgm8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:30:01 crc kubenswrapper[4885]: I1002 02:30:01.704152 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e45a508f-eb2f-45e6-875b-f877bb85101d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e45a508f-eb2f-45e6-875b-f877bb85101d" (UID: "e45a508f-eb2f-45e6-875b-f877bb85101d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:30:01 crc kubenswrapper[4885]: I1002 02:30:01.748036 4885 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e45a508f-eb2f-45e6-875b-f877bb85101d-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 02:30:01 crc kubenswrapper[4885]: I1002 02:30:01.748104 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kpgm8\" (UniqueName: \"kubernetes.io/projected/e45a508f-eb2f-45e6-875b-f877bb85101d-kube-api-access-kpgm8\") on node \"crc\" DevicePath \"\"" Oct 02 02:30:01 crc kubenswrapper[4885]: I1002 02:30:01.748134 4885 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e45a508f-eb2f-45e6-875b-f877bb85101d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 02:30:02 crc kubenswrapper[4885]: I1002 02:30:02.475353 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mrhcc" Oct 02 02:30:02 crc kubenswrapper[4885]: I1002 02:30:02.516818 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mrhcc"] Oct 02 02:30:02 crc kubenswrapper[4885]: I1002 02:30:02.533033 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-mrhcc"] Oct 02 02:30:02 crc kubenswrapper[4885]: I1002 02:30:02.840776 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322870-cbkw6" Oct 02 02:30:02 crc kubenswrapper[4885]: I1002 02:30:02.973942 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/126b33ad-f456-4e9a-ad7a-f16ba2613e9b-config-volume\") pod \"126b33ad-f456-4e9a-ad7a-f16ba2613e9b\" (UID: \"126b33ad-f456-4e9a-ad7a-f16ba2613e9b\") " Oct 02 02:30:02 crc kubenswrapper[4885]: I1002 02:30:02.974140 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nslc2\" (UniqueName: \"kubernetes.io/projected/126b33ad-f456-4e9a-ad7a-f16ba2613e9b-kube-api-access-nslc2\") pod \"126b33ad-f456-4e9a-ad7a-f16ba2613e9b\" (UID: \"126b33ad-f456-4e9a-ad7a-f16ba2613e9b\") " Oct 02 02:30:02 crc kubenswrapper[4885]: I1002 02:30:02.974239 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/126b33ad-f456-4e9a-ad7a-f16ba2613e9b-secret-volume\") pod \"126b33ad-f456-4e9a-ad7a-f16ba2613e9b\" (UID: \"126b33ad-f456-4e9a-ad7a-f16ba2613e9b\") " Oct 02 02:30:02 crc kubenswrapper[4885]: I1002 02:30:02.975062 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/126b33ad-f456-4e9a-ad7a-f16ba2613e9b-config-volume" (OuterVolumeSpecName: "config-volume") pod "126b33ad-f456-4e9a-ad7a-f16ba2613e9b" (UID: "126b33ad-f456-4e9a-ad7a-f16ba2613e9b"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:30:02 crc kubenswrapper[4885]: I1002 02:30:02.981711 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/126b33ad-f456-4e9a-ad7a-f16ba2613e9b-kube-api-access-nslc2" (OuterVolumeSpecName: "kube-api-access-nslc2") pod "126b33ad-f456-4e9a-ad7a-f16ba2613e9b" (UID: "126b33ad-f456-4e9a-ad7a-f16ba2613e9b"). InnerVolumeSpecName "kube-api-access-nslc2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:30:02 crc kubenswrapper[4885]: I1002 02:30:02.981843 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/126b33ad-f456-4e9a-ad7a-f16ba2613e9b-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "126b33ad-f456-4e9a-ad7a-f16ba2613e9b" (UID: "126b33ad-f456-4e9a-ad7a-f16ba2613e9b"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:30:03 crc kubenswrapper[4885]: I1002 02:30:03.076155 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nslc2\" (UniqueName: \"kubernetes.io/projected/126b33ad-f456-4e9a-ad7a-f16ba2613e9b-kube-api-access-nslc2\") on node \"crc\" DevicePath \"\"" Oct 02 02:30:03 crc kubenswrapper[4885]: I1002 02:30:03.076198 4885 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/126b33ad-f456-4e9a-ad7a-f16ba2613e9b-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 02 02:30:03 crc kubenswrapper[4885]: I1002 02:30:03.076210 4885 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/126b33ad-f456-4e9a-ad7a-f16ba2613e9b-config-volume\") on node \"crc\" DevicePath \"\"" Oct 02 02:30:03 crc kubenswrapper[4885]: I1002 02:30:03.490786 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29322870-cbkw6" event={"ID":"126b33ad-f456-4e9a-ad7a-f16ba2613e9b","Type":"ContainerDied","Data":"ba07868150fd33f187bb9a1a63d50a806af81a1b6bd615d11c8173330bd4063d"} Oct 02 02:30:03 crc kubenswrapper[4885]: I1002 02:30:03.491179 4885 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ba07868150fd33f187bb9a1a63d50a806af81a1b6bd615d11c8173330bd4063d" Oct 02 02:30:03 crc kubenswrapper[4885]: I1002 02:30:03.490926 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322870-cbkw6" Oct 02 02:30:03 crc kubenswrapper[4885]: I1002 02:30:03.933398 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322825-b7zrt"] Oct 02 02:30:03 crc kubenswrapper[4885]: I1002 02:30:03.946855 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322825-b7zrt"] Oct 02 02:30:04 crc kubenswrapper[4885]: I1002 02:30:04.068575 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bc961f3-ae85-4c19-8a57-be0af8e7bd10" path="/var/lib/kubelet/pods/4bc961f3-ae85-4c19-8a57-be0af8e7bd10/volumes" Oct 02 02:30:04 crc kubenswrapper[4885]: I1002 02:30:04.069828 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e45a508f-eb2f-45e6-875b-f877bb85101d" path="/var/lib/kubelet/pods/e45a508f-eb2f-45e6-875b-f877bb85101d/volumes" Oct 02 02:30:14 crc kubenswrapper[4885]: I1002 02:30:14.541605 4885 scope.go:117] "RemoveContainer" containerID="6930508575fab66820a624c8626849925a65067064512e41a2af37ab86368c07" Oct 02 02:31:43 crc kubenswrapper[4885]: I1002 02:31:43.266222 4885 patch_prober.go:28] interesting pod/machine-config-daemon-rttx8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 02:31:43 crc kubenswrapper[4885]: I1002 02:31:43.266912 4885 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 02:31:59 crc kubenswrapper[4885]: I1002 02:31:59.809356 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-z6wbt"] Oct 02 02:31:59 crc kubenswrapper[4885]: E1002 02:31:59.810360 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e45a508f-eb2f-45e6-875b-f877bb85101d" containerName="extract-content" Oct 02 02:31:59 crc kubenswrapper[4885]: I1002 02:31:59.810374 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="e45a508f-eb2f-45e6-875b-f877bb85101d" containerName="extract-content" Oct 02 02:31:59 crc kubenswrapper[4885]: E1002 02:31:59.810396 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e45a508f-eb2f-45e6-875b-f877bb85101d" containerName="registry-server" Oct 02 02:31:59 crc kubenswrapper[4885]: I1002 02:31:59.810402 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="e45a508f-eb2f-45e6-875b-f877bb85101d" containerName="registry-server" Oct 02 02:31:59 crc kubenswrapper[4885]: E1002 02:31:59.810411 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="126b33ad-f456-4e9a-ad7a-f16ba2613e9b" containerName="collect-profiles" Oct 02 02:31:59 crc kubenswrapper[4885]: I1002 02:31:59.810419 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="126b33ad-f456-4e9a-ad7a-f16ba2613e9b" containerName="collect-profiles" Oct 02 02:31:59 crc kubenswrapper[4885]: E1002 02:31:59.810435 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e45a508f-eb2f-45e6-875b-f877bb85101d" containerName="extract-utilities" Oct 02 02:31:59 crc kubenswrapper[4885]: I1002 02:31:59.810441 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="e45a508f-eb2f-45e6-875b-f877bb85101d" containerName="extract-utilities" Oct 02 02:31:59 crc kubenswrapper[4885]: I1002 02:31:59.810624 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="126b33ad-f456-4e9a-ad7a-f16ba2613e9b" containerName="collect-profiles" Oct 02 02:31:59 crc kubenswrapper[4885]: I1002 02:31:59.810652 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="e45a508f-eb2f-45e6-875b-f877bb85101d" containerName="registry-server" Oct 02 02:31:59 crc kubenswrapper[4885]: I1002 02:31:59.812008 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-z6wbt" Oct 02 02:31:59 crc kubenswrapper[4885]: I1002 02:31:59.875436 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-z6wbt"] Oct 02 02:31:59 crc kubenswrapper[4885]: I1002 02:31:59.898203 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e749fb60-7c21-40e1-9b9e-d9ceb52c25f0-catalog-content\") pod \"redhat-operators-z6wbt\" (UID: \"e749fb60-7c21-40e1-9b9e-d9ceb52c25f0\") " pod="openshift-marketplace/redhat-operators-z6wbt" Oct 02 02:31:59 crc kubenswrapper[4885]: I1002 02:31:59.898287 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p8fpg\" (UniqueName: \"kubernetes.io/projected/e749fb60-7c21-40e1-9b9e-d9ceb52c25f0-kube-api-access-p8fpg\") pod \"redhat-operators-z6wbt\" (UID: \"e749fb60-7c21-40e1-9b9e-d9ceb52c25f0\") " pod="openshift-marketplace/redhat-operators-z6wbt" Oct 02 02:31:59 crc kubenswrapper[4885]: I1002 02:31:59.898584 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e749fb60-7c21-40e1-9b9e-d9ceb52c25f0-utilities\") pod \"redhat-operators-z6wbt\" (UID: \"e749fb60-7c21-40e1-9b9e-d9ceb52c25f0\") " pod="openshift-marketplace/redhat-operators-z6wbt" Oct 02 02:32:00 crc kubenswrapper[4885]: I1002 02:32:00.000793 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e749fb60-7c21-40e1-9b9e-d9ceb52c25f0-catalog-content\") pod \"redhat-operators-z6wbt\" (UID: \"e749fb60-7c21-40e1-9b9e-d9ceb52c25f0\") " pod="openshift-marketplace/redhat-operators-z6wbt" Oct 02 02:32:00 crc kubenswrapper[4885]: I1002 02:32:00.000849 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p8fpg\" (UniqueName: \"kubernetes.io/projected/e749fb60-7c21-40e1-9b9e-d9ceb52c25f0-kube-api-access-p8fpg\") pod \"redhat-operators-z6wbt\" (UID: \"e749fb60-7c21-40e1-9b9e-d9ceb52c25f0\") " pod="openshift-marketplace/redhat-operators-z6wbt" Oct 02 02:32:00 crc kubenswrapper[4885]: I1002 02:32:00.000950 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e749fb60-7c21-40e1-9b9e-d9ceb52c25f0-utilities\") pod \"redhat-operators-z6wbt\" (UID: \"e749fb60-7c21-40e1-9b9e-d9ceb52c25f0\") " pod="openshift-marketplace/redhat-operators-z6wbt" Oct 02 02:32:00 crc kubenswrapper[4885]: I1002 02:32:00.001323 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e749fb60-7c21-40e1-9b9e-d9ceb52c25f0-catalog-content\") pod \"redhat-operators-z6wbt\" (UID: \"e749fb60-7c21-40e1-9b9e-d9ceb52c25f0\") " pod="openshift-marketplace/redhat-operators-z6wbt" Oct 02 02:32:00 crc kubenswrapper[4885]: I1002 02:32:00.001394 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e749fb60-7c21-40e1-9b9e-d9ceb52c25f0-utilities\") pod \"redhat-operators-z6wbt\" (UID: \"e749fb60-7c21-40e1-9b9e-d9ceb52c25f0\") " pod="openshift-marketplace/redhat-operators-z6wbt" Oct 02 02:32:00 crc kubenswrapper[4885]: I1002 02:32:00.024215 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p8fpg\" (UniqueName: \"kubernetes.io/projected/e749fb60-7c21-40e1-9b9e-d9ceb52c25f0-kube-api-access-p8fpg\") pod \"redhat-operators-z6wbt\" (UID: \"e749fb60-7c21-40e1-9b9e-d9ceb52c25f0\") " pod="openshift-marketplace/redhat-operators-z6wbt" Oct 02 02:32:00 crc kubenswrapper[4885]: I1002 02:32:00.142364 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-z6wbt" Oct 02 02:32:00 crc kubenswrapper[4885]: I1002 02:32:00.589730 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-z6wbt"] Oct 02 02:32:00 crc kubenswrapper[4885]: I1002 02:32:00.929615 4885 generic.go:334] "Generic (PLEG): container finished" podID="e749fb60-7c21-40e1-9b9e-d9ceb52c25f0" containerID="81664d45dbbd929a563022d8bc2b9e108c445dd0d48a0451ec6b235a16d884f9" exitCode=0 Oct 02 02:32:00 crc kubenswrapper[4885]: I1002 02:32:00.929725 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z6wbt" event={"ID":"e749fb60-7c21-40e1-9b9e-d9ceb52c25f0","Type":"ContainerDied","Data":"81664d45dbbd929a563022d8bc2b9e108c445dd0d48a0451ec6b235a16d884f9"} Oct 02 02:32:00 crc kubenswrapper[4885]: I1002 02:32:00.929849 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z6wbt" event={"ID":"e749fb60-7c21-40e1-9b9e-d9ceb52c25f0","Type":"ContainerStarted","Data":"9999ddaf51dbcd0e4a4faa34b1089c60147b595f1ed8605b20b83dce0fb2763c"} Oct 02 02:32:02 crc kubenswrapper[4885]: I1002 02:32:02.968750 4885 generic.go:334] "Generic (PLEG): container finished" podID="e749fb60-7c21-40e1-9b9e-d9ceb52c25f0" containerID="c12aae53162ce08e16e6a05b2e63b3045f76ad7ca7df6bd685fdd22187dfdd48" exitCode=0 Oct 02 02:32:02 crc kubenswrapper[4885]: I1002 02:32:02.968948 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z6wbt" event={"ID":"e749fb60-7c21-40e1-9b9e-d9ceb52c25f0","Type":"ContainerDied","Data":"c12aae53162ce08e16e6a05b2e63b3045f76ad7ca7df6bd685fdd22187dfdd48"} Oct 02 02:32:03 crc kubenswrapper[4885]: I1002 02:32:03.978761 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z6wbt" event={"ID":"e749fb60-7c21-40e1-9b9e-d9ceb52c25f0","Type":"ContainerStarted","Data":"e3348a6bea5153fa99ff9db300d9ad812698b108ba776df39bd70cd35c106618"} Oct 02 02:32:04 crc kubenswrapper[4885]: I1002 02:32:04.011673 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-z6wbt" podStartSLOduration=2.546168604 podStartE2EDuration="5.011654344s" podCreationTimestamp="2025-10-02 02:31:59 +0000 UTC" firstStartedPulling="2025-10-02 02:32:00.931156701 +0000 UTC m=+2709.742904100" lastFinishedPulling="2025-10-02 02:32:03.396642401 +0000 UTC m=+2712.208389840" observedRunningTime="2025-10-02 02:32:04.003571158 +0000 UTC m=+2712.815318597" watchObservedRunningTime="2025-10-02 02:32:04.011654344 +0000 UTC m=+2712.823401743" Oct 02 02:32:10 crc kubenswrapper[4885]: I1002 02:32:10.142505 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-z6wbt" Oct 02 02:32:10 crc kubenswrapper[4885]: I1002 02:32:10.143115 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-z6wbt" Oct 02 02:32:10 crc kubenswrapper[4885]: I1002 02:32:10.207564 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-z6wbt" Oct 02 02:32:11 crc kubenswrapper[4885]: I1002 02:32:11.116420 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-z6wbt" Oct 02 02:32:11 crc kubenswrapper[4885]: I1002 02:32:11.184124 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-z6wbt"] Oct 02 02:32:13 crc kubenswrapper[4885]: I1002 02:32:13.085702 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-z6wbt" podUID="e749fb60-7c21-40e1-9b9e-d9ceb52c25f0" containerName="registry-server" containerID="cri-o://e3348a6bea5153fa99ff9db300d9ad812698b108ba776df39bd70cd35c106618" gracePeriod=2 Oct 02 02:32:13 crc kubenswrapper[4885]: I1002 02:32:13.265770 4885 patch_prober.go:28] interesting pod/machine-config-daemon-rttx8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 02:32:13 crc kubenswrapper[4885]: I1002 02:32:13.265831 4885 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 02:32:13 crc kubenswrapper[4885]: I1002 02:32:13.570223 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-z6wbt" Oct 02 02:32:13 crc kubenswrapper[4885]: I1002 02:32:13.672982 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e749fb60-7c21-40e1-9b9e-d9ceb52c25f0-catalog-content\") pod \"e749fb60-7c21-40e1-9b9e-d9ceb52c25f0\" (UID: \"e749fb60-7c21-40e1-9b9e-d9ceb52c25f0\") " Oct 02 02:32:13 crc kubenswrapper[4885]: I1002 02:32:13.673037 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e749fb60-7c21-40e1-9b9e-d9ceb52c25f0-utilities\") pod \"e749fb60-7c21-40e1-9b9e-d9ceb52c25f0\" (UID: \"e749fb60-7c21-40e1-9b9e-d9ceb52c25f0\") " Oct 02 02:32:13 crc kubenswrapper[4885]: I1002 02:32:13.673142 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p8fpg\" (UniqueName: \"kubernetes.io/projected/e749fb60-7c21-40e1-9b9e-d9ceb52c25f0-kube-api-access-p8fpg\") pod \"e749fb60-7c21-40e1-9b9e-d9ceb52c25f0\" (UID: \"e749fb60-7c21-40e1-9b9e-d9ceb52c25f0\") " Oct 02 02:32:13 crc kubenswrapper[4885]: I1002 02:32:13.674765 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e749fb60-7c21-40e1-9b9e-d9ceb52c25f0-utilities" (OuterVolumeSpecName: "utilities") pod "e749fb60-7c21-40e1-9b9e-d9ceb52c25f0" (UID: "e749fb60-7c21-40e1-9b9e-d9ceb52c25f0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:32:13 crc kubenswrapper[4885]: I1002 02:32:13.678714 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e749fb60-7c21-40e1-9b9e-d9ceb52c25f0-kube-api-access-p8fpg" (OuterVolumeSpecName: "kube-api-access-p8fpg") pod "e749fb60-7c21-40e1-9b9e-d9ceb52c25f0" (UID: "e749fb60-7c21-40e1-9b9e-d9ceb52c25f0"). InnerVolumeSpecName "kube-api-access-p8fpg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:32:13 crc kubenswrapper[4885]: I1002 02:32:13.760034 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e749fb60-7c21-40e1-9b9e-d9ceb52c25f0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e749fb60-7c21-40e1-9b9e-d9ceb52c25f0" (UID: "e749fb60-7c21-40e1-9b9e-d9ceb52c25f0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:32:13 crc kubenswrapper[4885]: I1002 02:32:13.774591 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p8fpg\" (UniqueName: \"kubernetes.io/projected/e749fb60-7c21-40e1-9b9e-d9ceb52c25f0-kube-api-access-p8fpg\") on node \"crc\" DevicePath \"\"" Oct 02 02:32:13 crc kubenswrapper[4885]: I1002 02:32:13.774621 4885 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e749fb60-7c21-40e1-9b9e-d9ceb52c25f0-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 02:32:13 crc kubenswrapper[4885]: I1002 02:32:13.774630 4885 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e749fb60-7c21-40e1-9b9e-d9ceb52c25f0-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 02:32:14 crc kubenswrapper[4885]: I1002 02:32:14.102995 4885 generic.go:334] "Generic (PLEG): container finished" podID="e749fb60-7c21-40e1-9b9e-d9ceb52c25f0" containerID="e3348a6bea5153fa99ff9db300d9ad812698b108ba776df39bd70cd35c106618" exitCode=0 Oct 02 02:32:14 crc kubenswrapper[4885]: I1002 02:32:14.103072 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z6wbt" event={"ID":"e749fb60-7c21-40e1-9b9e-d9ceb52c25f0","Type":"ContainerDied","Data":"e3348a6bea5153fa99ff9db300d9ad812698b108ba776df39bd70cd35c106618"} Oct 02 02:32:14 crc kubenswrapper[4885]: I1002 02:32:14.103206 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z6wbt" event={"ID":"e749fb60-7c21-40e1-9b9e-d9ceb52c25f0","Type":"ContainerDied","Data":"9999ddaf51dbcd0e4a4faa34b1089c60147b595f1ed8605b20b83dce0fb2763c"} Oct 02 02:32:14 crc kubenswrapper[4885]: I1002 02:32:14.103380 4885 scope.go:117] "RemoveContainer" containerID="e3348a6bea5153fa99ff9db300d9ad812698b108ba776df39bd70cd35c106618" Oct 02 02:32:14 crc kubenswrapper[4885]: I1002 02:32:14.103397 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-z6wbt" Oct 02 02:32:14 crc kubenswrapper[4885]: I1002 02:32:14.144469 4885 scope.go:117] "RemoveContainer" containerID="c12aae53162ce08e16e6a05b2e63b3045f76ad7ca7df6bd685fdd22187dfdd48" Oct 02 02:32:14 crc kubenswrapper[4885]: I1002 02:32:14.148745 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-z6wbt"] Oct 02 02:32:14 crc kubenswrapper[4885]: I1002 02:32:14.158360 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-z6wbt"] Oct 02 02:32:14 crc kubenswrapper[4885]: I1002 02:32:14.183334 4885 scope.go:117] "RemoveContainer" containerID="81664d45dbbd929a563022d8bc2b9e108c445dd0d48a0451ec6b235a16d884f9" Oct 02 02:32:14 crc kubenswrapper[4885]: I1002 02:32:14.223252 4885 scope.go:117] "RemoveContainer" containerID="e3348a6bea5153fa99ff9db300d9ad812698b108ba776df39bd70cd35c106618" Oct 02 02:32:14 crc kubenswrapper[4885]: E1002 02:32:14.223733 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e3348a6bea5153fa99ff9db300d9ad812698b108ba776df39bd70cd35c106618\": container with ID starting with e3348a6bea5153fa99ff9db300d9ad812698b108ba776df39bd70cd35c106618 not found: ID does not exist" containerID="e3348a6bea5153fa99ff9db300d9ad812698b108ba776df39bd70cd35c106618" Oct 02 02:32:14 crc kubenswrapper[4885]: I1002 02:32:14.223816 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e3348a6bea5153fa99ff9db300d9ad812698b108ba776df39bd70cd35c106618"} err="failed to get container status \"e3348a6bea5153fa99ff9db300d9ad812698b108ba776df39bd70cd35c106618\": rpc error: code = NotFound desc = could not find container \"e3348a6bea5153fa99ff9db300d9ad812698b108ba776df39bd70cd35c106618\": container with ID starting with e3348a6bea5153fa99ff9db300d9ad812698b108ba776df39bd70cd35c106618 not found: ID does not exist" Oct 02 02:32:14 crc kubenswrapper[4885]: I1002 02:32:14.223852 4885 scope.go:117] "RemoveContainer" containerID="c12aae53162ce08e16e6a05b2e63b3045f76ad7ca7df6bd685fdd22187dfdd48" Oct 02 02:32:14 crc kubenswrapper[4885]: E1002 02:32:14.224399 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c12aae53162ce08e16e6a05b2e63b3045f76ad7ca7df6bd685fdd22187dfdd48\": container with ID starting with c12aae53162ce08e16e6a05b2e63b3045f76ad7ca7df6bd685fdd22187dfdd48 not found: ID does not exist" containerID="c12aae53162ce08e16e6a05b2e63b3045f76ad7ca7df6bd685fdd22187dfdd48" Oct 02 02:32:14 crc kubenswrapper[4885]: I1002 02:32:14.224455 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c12aae53162ce08e16e6a05b2e63b3045f76ad7ca7df6bd685fdd22187dfdd48"} err="failed to get container status \"c12aae53162ce08e16e6a05b2e63b3045f76ad7ca7df6bd685fdd22187dfdd48\": rpc error: code = NotFound desc = could not find container \"c12aae53162ce08e16e6a05b2e63b3045f76ad7ca7df6bd685fdd22187dfdd48\": container with ID starting with c12aae53162ce08e16e6a05b2e63b3045f76ad7ca7df6bd685fdd22187dfdd48 not found: ID does not exist" Oct 02 02:32:14 crc kubenswrapper[4885]: I1002 02:32:14.224490 4885 scope.go:117] "RemoveContainer" containerID="81664d45dbbd929a563022d8bc2b9e108c445dd0d48a0451ec6b235a16d884f9" Oct 02 02:32:14 crc kubenswrapper[4885]: E1002 02:32:14.224964 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"81664d45dbbd929a563022d8bc2b9e108c445dd0d48a0451ec6b235a16d884f9\": container with ID starting with 81664d45dbbd929a563022d8bc2b9e108c445dd0d48a0451ec6b235a16d884f9 not found: ID does not exist" containerID="81664d45dbbd929a563022d8bc2b9e108c445dd0d48a0451ec6b235a16d884f9" Oct 02 02:32:14 crc kubenswrapper[4885]: I1002 02:32:14.224994 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"81664d45dbbd929a563022d8bc2b9e108c445dd0d48a0451ec6b235a16d884f9"} err="failed to get container status \"81664d45dbbd929a563022d8bc2b9e108c445dd0d48a0451ec6b235a16d884f9\": rpc error: code = NotFound desc = could not find container \"81664d45dbbd929a563022d8bc2b9e108c445dd0d48a0451ec6b235a16d884f9\": container with ID starting with 81664d45dbbd929a563022d8bc2b9e108c445dd0d48a0451ec6b235a16d884f9 not found: ID does not exist" Oct 02 02:32:16 crc kubenswrapper[4885]: I1002 02:32:16.066414 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e749fb60-7c21-40e1-9b9e-d9ceb52c25f0" path="/var/lib/kubelet/pods/e749fb60-7c21-40e1-9b9e-d9ceb52c25f0/volumes" Oct 02 02:32:34 crc kubenswrapper[4885]: I1002 02:32:34.353879 4885 generic.go:334] "Generic (PLEG): container finished" podID="ce63c1c1-34cc-4cba-a553-33b66431bc3c" containerID="9c92bcdaf1b3ed3e96e5851ff9f76d5409006ad36b9c86452bae73704fc96a9c" exitCode=0 Oct 02 02:32:34 crc kubenswrapper[4885]: I1002 02:32:34.353988 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v9h2z" event={"ID":"ce63c1c1-34cc-4cba-a553-33b66431bc3c","Type":"ContainerDied","Data":"9c92bcdaf1b3ed3e96e5851ff9f76d5409006ad36b9c86452bae73704fc96a9c"} Oct 02 02:32:35 crc kubenswrapper[4885]: I1002 02:32:35.849994 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v9h2z" Oct 02 02:32:35 crc kubenswrapper[4885]: I1002 02:32:35.941723 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/ce63c1c1-34cc-4cba-a553-33b66431bc3c-ceilometer-compute-config-data-2\") pod \"ce63c1c1-34cc-4cba-a553-33b66431bc3c\" (UID: \"ce63c1c1-34cc-4cba-a553-33b66431bc3c\") " Oct 02 02:32:35 crc kubenswrapper[4885]: I1002 02:32:35.942083 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gq98j\" (UniqueName: \"kubernetes.io/projected/ce63c1c1-34cc-4cba-a553-33b66431bc3c-kube-api-access-gq98j\") pod \"ce63c1c1-34cc-4cba-a553-33b66431bc3c\" (UID: \"ce63c1c1-34cc-4cba-a553-33b66431bc3c\") " Oct 02 02:32:35 crc kubenswrapper[4885]: I1002 02:32:35.942395 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ce63c1c1-34cc-4cba-a553-33b66431bc3c-inventory\") pod \"ce63c1c1-34cc-4cba-a553-33b66431bc3c\" (UID: \"ce63c1c1-34cc-4cba-a553-33b66431bc3c\") " Oct 02 02:32:35 crc kubenswrapper[4885]: I1002 02:32:35.942442 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce63c1c1-34cc-4cba-a553-33b66431bc3c-telemetry-combined-ca-bundle\") pod \"ce63c1c1-34cc-4cba-a553-33b66431bc3c\" (UID: \"ce63c1c1-34cc-4cba-a553-33b66431bc3c\") " Oct 02 02:32:35 crc kubenswrapper[4885]: I1002 02:32:35.942492 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ce63c1c1-34cc-4cba-a553-33b66431bc3c-ssh-key\") pod \"ce63c1c1-34cc-4cba-a553-33b66431bc3c\" (UID: \"ce63c1c1-34cc-4cba-a553-33b66431bc3c\") " Oct 02 02:32:35 crc kubenswrapper[4885]: I1002 02:32:35.942528 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/ce63c1c1-34cc-4cba-a553-33b66431bc3c-ceilometer-compute-config-data-1\") pod \"ce63c1c1-34cc-4cba-a553-33b66431bc3c\" (UID: \"ce63c1c1-34cc-4cba-a553-33b66431bc3c\") " Oct 02 02:32:35 crc kubenswrapper[4885]: I1002 02:32:35.942562 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/ce63c1c1-34cc-4cba-a553-33b66431bc3c-ceilometer-compute-config-data-0\") pod \"ce63c1c1-34cc-4cba-a553-33b66431bc3c\" (UID: \"ce63c1c1-34cc-4cba-a553-33b66431bc3c\") " Oct 02 02:32:35 crc kubenswrapper[4885]: I1002 02:32:35.977122 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce63c1c1-34cc-4cba-a553-33b66431bc3c-kube-api-access-gq98j" (OuterVolumeSpecName: "kube-api-access-gq98j") pod "ce63c1c1-34cc-4cba-a553-33b66431bc3c" (UID: "ce63c1c1-34cc-4cba-a553-33b66431bc3c"). InnerVolumeSpecName "kube-api-access-gq98j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:32:35 crc kubenswrapper[4885]: I1002 02:32:35.977918 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce63c1c1-34cc-4cba-a553-33b66431bc3c-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "ce63c1c1-34cc-4cba-a553-33b66431bc3c" (UID: "ce63c1c1-34cc-4cba-a553-33b66431bc3c"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:32:35 crc kubenswrapper[4885]: I1002 02:32:35.981383 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce63c1c1-34cc-4cba-a553-33b66431bc3c-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "ce63c1c1-34cc-4cba-a553-33b66431bc3c" (UID: "ce63c1c1-34cc-4cba-a553-33b66431bc3c"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:32:35 crc kubenswrapper[4885]: I1002 02:32:35.982513 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce63c1c1-34cc-4cba-a553-33b66431bc3c-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "ce63c1c1-34cc-4cba-a553-33b66431bc3c" (UID: "ce63c1c1-34cc-4cba-a553-33b66431bc3c"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:32:35 crc kubenswrapper[4885]: I1002 02:32:35.985530 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce63c1c1-34cc-4cba-a553-33b66431bc3c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ce63c1c1-34cc-4cba-a553-33b66431bc3c" (UID: "ce63c1c1-34cc-4cba-a553-33b66431bc3c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:32:35 crc kubenswrapper[4885]: I1002 02:32:35.989122 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce63c1c1-34cc-4cba-a553-33b66431bc3c-inventory" (OuterVolumeSpecName: "inventory") pod "ce63c1c1-34cc-4cba-a553-33b66431bc3c" (UID: "ce63c1c1-34cc-4cba-a553-33b66431bc3c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:32:36 crc kubenswrapper[4885]: I1002 02:32:36.003603 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce63c1c1-34cc-4cba-a553-33b66431bc3c-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "ce63c1c1-34cc-4cba-a553-33b66431bc3c" (UID: "ce63c1c1-34cc-4cba-a553-33b66431bc3c"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:32:36 crc kubenswrapper[4885]: I1002 02:32:36.045050 4885 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ce63c1c1-34cc-4cba-a553-33b66431bc3c-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 02:32:36 crc kubenswrapper[4885]: I1002 02:32:36.045088 4885 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce63c1c1-34cc-4cba-a553-33b66431bc3c-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:32:36 crc kubenswrapper[4885]: I1002 02:32:36.045108 4885 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ce63c1c1-34cc-4cba-a553-33b66431bc3c-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 02:32:36 crc kubenswrapper[4885]: I1002 02:32:36.045124 4885 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/ce63c1c1-34cc-4cba-a553-33b66431bc3c-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Oct 02 02:32:36 crc kubenswrapper[4885]: I1002 02:32:36.045175 4885 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/ce63c1c1-34cc-4cba-a553-33b66431bc3c-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Oct 02 02:32:36 crc kubenswrapper[4885]: I1002 02:32:36.045192 4885 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/ce63c1c1-34cc-4cba-a553-33b66431bc3c-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Oct 02 02:32:36 crc kubenswrapper[4885]: I1002 02:32:36.045209 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gq98j\" (UniqueName: \"kubernetes.io/projected/ce63c1c1-34cc-4cba-a553-33b66431bc3c-kube-api-access-gq98j\") on node \"crc\" DevicePath \"\"" Oct 02 02:32:36 crc kubenswrapper[4885]: I1002 02:32:36.379342 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v9h2z" event={"ID":"ce63c1c1-34cc-4cba-a553-33b66431bc3c","Type":"ContainerDied","Data":"83f63a436cd81c26858faa49959192a7d875abca3cdb13cceb7d139a7ef8552c"} Oct 02 02:32:36 crc kubenswrapper[4885]: I1002 02:32:36.379418 4885 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="83f63a436cd81c26858faa49959192a7d875abca3cdb13cceb7d139a7ef8552c" Oct 02 02:32:36 crc kubenswrapper[4885]: I1002 02:32:36.379425 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v9h2z" Oct 02 02:32:43 crc kubenswrapper[4885]: I1002 02:32:43.265928 4885 patch_prober.go:28] interesting pod/machine-config-daemon-rttx8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 02:32:43 crc kubenswrapper[4885]: I1002 02:32:43.266815 4885 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 02:32:43 crc kubenswrapper[4885]: I1002 02:32:43.266901 4885 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" Oct 02 02:32:43 crc kubenswrapper[4885]: I1002 02:32:43.268223 4885 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5bbcad619e10a916cd09b9cfd4ad5c9334559ab939f077b917dc9ae849fe425f"} pod="openshift-machine-config-operator/machine-config-daemon-rttx8" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 02:32:43 crc kubenswrapper[4885]: I1002 02:32:43.268394 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" containerName="machine-config-daemon" containerID="cri-o://5bbcad619e10a916cd09b9cfd4ad5c9334559ab939f077b917dc9ae849fe425f" gracePeriod=600 Oct 02 02:32:43 crc kubenswrapper[4885]: I1002 02:32:43.462831 4885 generic.go:334] "Generic (PLEG): container finished" podID="d169c8ba-000a-4218-8e3b-4ae53035b110" containerID="5bbcad619e10a916cd09b9cfd4ad5c9334559ab939f077b917dc9ae849fe425f" exitCode=0 Oct 02 02:32:43 crc kubenswrapper[4885]: I1002 02:32:43.462879 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" event={"ID":"d169c8ba-000a-4218-8e3b-4ae53035b110","Type":"ContainerDied","Data":"5bbcad619e10a916cd09b9cfd4ad5c9334559ab939f077b917dc9ae849fe425f"} Oct 02 02:32:43 crc kubenswrapper[4885]: I1002 02:32:43.462914 4885 scope.go:117] "RemoveContainer" containerID="7d1be267228233dc9433f75e76a5a805aa23d98aa3b0edbd385a1a565a0d3231" Oct 02 02:32:44 crc kubenswrapper[4885]: I1002 02:32:44.482650 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" event={"ID":"d169c8ba-000a-4218-8e3b-4ae53035b110","Type":"ContainerStarted","Data":"02916a559cc5f6ab8f5bc76ee73c22c5f71f201c585eeb74e3c14d51fc63cccb"} Oct 02 02:33:33 crc kubenswrapper[4885]: I1002 02:33:33.424598 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Oct 02 02:33:33 crc kubenswrapper[4885]: E1002 02:33:33.426112 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e749fb60-7c21-40e1-9b9e-d9ceb52c25f0" containerName="extract-utilities" Oct 02 02:33:33 crc kubenswrapper[4885]: I1002 02:33:33.426149 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="e749fb60-7c21-40e1-9b9e-d9ceb52c25f0" containerName="extract-utilities" Oct 02 02:33:33 crc kubenswrapper[4885]: E1002 02:33:33.426189 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e749fb60-7c21-40e1-9b9e-d9ceb52c25f0" containerName="extract-content" Oct 02 02:33:33 crc kubenswrapper[4885]: I1002 02:33:33.426206 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="e749fb60-7c21-40e1-9b9e-d9ceb52c25f0" containerName="extract-content" Oct 02 02:33:33 crc kubenswrapper[4885]: E1002 02:33:33.426232 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e749fb60-7c21-40e1-9b9e-d9ceb52c25f0" containerName="registry-server" Oct 02 02:33:33 crc kubenswrapper[4885]: I1002 02:33:33.426250 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="e749fb60-7c21-40e1-9b9e-d9ceb52c25f0" containerName="registry-server" Oct 02 02:33:33 crc kubenswrapper[4885]: E1002 02:33:33.426328 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce63c1c1-34cc-4cba-a553-33b66431bc3c" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Oct 02 02:33:33 crc kubenswrapper[4885]: I1002 02:33:33.426347 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce63c1c1-34cc-4cba-a553-33b66431bc3c" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Oct 02 02:33:33 crc kubenswrapper[4885]: I1002 02:33:33.426785 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="e749fb60-7c21-40e1-9b9e-d9ceb52c25f0" containerName="registry-server" Oct 02 02:33:33 crc kubenswrapper[4885]: I1002 02:33:33.426827 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce63c1c1-34cc-4cba-a553-33b66431bc3c" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Oct 02 02:33:33 crc kubenswrapper[4885]: I1002 02:33:33.428228 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 02 02:33:33 crc kubenswrapper[4885]: I1002 02:33:33.433463 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Oct 02 02:33:33 crc kubenswrapper[4885]: I1002 02:33:33.434088 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-mg457" Oct 02 02:33:33 crc kubenswrapper[4885]: I1002 02:33:33.433763 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Oct 02 02:33:33 crc kubenswrapper[4885]: I1002 02:33:33.434197 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Oct 02 02:33:33 crc kubenswrapper[4885]: I1002 02:33:33.464431 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Oct 02 02:33:33 crc kubenswrapper[4885]: I1002 02:33:33.528024 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/7a715636-06d0-4508-a9d0-fe83f4c05d01-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"7a715636-06d0-4508-a9d0-fe83f4c05d01\") " pod="openstack/tempest-tests-tempest" Oct 02 02:33:33 crc kubenswrapper[4885]: I1002 02:33:33.528092 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/7a715636-06d0-4508-a9d0-fe83f4c05d01-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"7a715636-06d0-4508-a9d0-fe83f4c05d01\") " pod="openstack/tempest-tests-tempest" Oct 02 02:33:33 crc kubenswrapper[4885]: I1002 02:33:33.528136 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/7a715636-06d0-4508-a9d0-fe83f4c05d01-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"7a715636-06d0-4508-a9d0-fe83f4c05d01\") " pod="openstack/tempest-tests-tempest" Oct 02 02:33:33 crc kubenswrapper[4885]: I1002 02:33:33.528174 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/7a715636-06d0-4508-a9d0-fe83f4c05d01-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"7a715636-06d0-4508-a9d0-fe83f4c05d01\") " pod="openstack/tempest-tests-tempest" Oct 02 02:33:33 crc kubenswrapper[4885]: I1002 02:33:33.528225 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7a715636-06d0-4508-a9d0-fe83f4c05d01-config-data\") pod \"tempest-tests-tempest\" (UID: \"7a715636-06d0-4508-a9d0-fe83f4c05d01\") " pod="openstack/tempest-tests-tempest" Oct 02 02:33:33 crc kubenswrapper[4885]: I1002 02:33:33.528328 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7a715636-06d0-4508-a9d0-fe83f4c05d01-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"7a715636-06d0-4508-a9d0-fe83f4c05d01\") " pod="openstack/tempest-tests-tempest" Oct 02 02:33:33 crc kubenswrapper[4885]: I1002 02:33:33.528455 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/7a715636-06d0-4508-a9d0-fe83f4c05d01-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"7a715636-06d0-4508-a9d0-fe83f4c05d01\") " pod="openstack/tempest-tests-tempest" Oct 02 02:33:33 crc kubenswrapper[4885]: I1002 02:33:33.528541 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"tempest-tests-tempest\" (UID: \"7a715636-06d0-4508-a9d0-fe83f4c05d01\") " pod="openstack/tempest-tests-tempest" Oct 02 02:33:33 crc kubenswrapper[4885]: I1002 02:33:33.529246 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vjxlh\" (UniqueName: \"kubernetes.io/projected/7a715636-06d0-4508-a9d0-fe83f4c05d01-kube-api-access-vjxlh\") pod \"tempest-tests-tempest\" (UID: \"7a715636-06d0-4508-a9d0-fe83f4c05d01\") " pod="openstack/tempest-tests-tempest" Oct 02 02:33:33 crc kubenswrapper[4885]: I1002 02:33:33.632565 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/7a715636-06d0-4508-a9d0-fe83f4c05d01-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"7a715636-06d0-4508-a9d0-fe83f4c05d01\") " pod="openstack/tempest-tests-tempest" Oct 02 02:33:33 crc kubenswrapper[4885]: I1002 02:33:33.632683 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"tempest-tests-tempest\" (UID: \"7a715636-06d0-4508-a9d0-fe83f4c05d01\") " pod="openstack/tempest-tests-tempest" Oct 02 02:33:33 crc kubenswrapper[4885]: I1002 02:33:33.632882 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vjxlh\" (UniqueName: \"kubernetes.io/projected/7a715636-06d0-4508-a9d0-fe83f4c05d01-kube-api-access-vjxlh\") pod \"tempest-tests-tempest\" (UID: \"7a715636-06d0-4508-a9d0-fe83f4c05d01\") " pod="openstack/tempest-tests-tempest" Oct 02 02:33:33 crc kubenswrapper[4885]: I1002 02:33:33.632951 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/7a715636-06d0-4508-a9d0-fe83f4c05d01-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"7a715636-06d0-4508-a9d0-fe83f4c05d01\") " pod="openstack/tempest-tests-tempest" Oct 02 02:33:33 crc kubenswrapper[4885]: I1002 02:33:33.632987 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/7a715636-06d0-4508-a9d0-fe83f4c05d01-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"7a715636-06d0-4508-a9d0-fe83f4c05d01\") " pod="openstack/tempest-tests-tempest" Oct 02 02:33:33 crc kubenswrapper[4885]: I1002 02:33:33.633031 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/7a715636-06d0-4508-a9d0-fe83f4c05d01-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"7a715636-06d0-4508-a9d0-fe83f4c05d01\") " pod="openstack/tempest-tests-tempest" Oct 02 02:33:33 crc kubenswrapper[4885]: I1002 02:33:33.633080 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/7a715636-06d0-4508-a9d0-fe83f4c05d01-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"7a715636-06d0-4508-a9d0-fe83f4c05d01\") " pod="openstack/tempest-tests-tempest" Oct 02 02:33:33 crc kubenswrapper[4885]: I1002 02:33:33.633130 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7a715636-06d0-4508-a9d0-fe83f4c05d01-config-data\") pod \"tempest-tests-tempest\" (UID: \"7a715636-06d0-4508-a9d0-fe83f4c05d01\") " pod="openstack/tempest-tests-tempest" Oct 02 02:33:33 crc kubenswrapper[4885]: I1002 02:33:33.633196 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7a715636-06d0-4508-a9d0-fe83f4c05d01-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"7a715636-06d0-4508-a9d0-fe83f4c05d01\") " pod="openstack/tempest-tests-tempest" Oct 02 02:33:33 crc kubenswrapper[4885]: I1002 02:33:33.633299 4885 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"tempest-tests-tempest\" (UID: \"7a715636-06d0-4508-a9d0-fe83f4c05d01\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/tempest-tests-tempest" Oct 02 02:33:33 crc kubenswrapper[4885]: I1002 02:33:33.634748 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/7a715636-06d0-4508-a9d0-fe83f4c05d01-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"7a715636-06d0-4508-a9d0-fe83f4c05d01\") " pod="openstack/tempest-tests-tempest" Oct 02 02:33:33 crc kubenswrapper[4885]: I1002 02:33:33.635232 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/7a715636-06d0-4508-a9d0-fe83f4c05d01-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"7a715636-06d0-4508-a9d0-fe83f4c05d01\") " pod="openstack/tempest-tests-tempest" Oct 02 02:33:33 crc kubenswrapper[4885]: I1002 02:33:33.635742 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/7a715636-06d0-4508-a9d0-fe83f4c05d01-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"7a715636-06d0-4508-a9d0-fe83f4c05d01\") " pod="openstack/tempest-tests-tempest" Oct 02 02:33:33 crc kubenswrapper[4885]: I1002 02:33:33.637901 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7a715636-06d0-4508-a9d0-fe83f4c05d01-config-data\") pod \"tempest-tests-tempest\" (UID: \"7a715636-06d0-4508-a9d0-fe83f4c05d01\") " pod="openstack/tempest-tests-tempest" Oct 02 02:33:33 crc kubenswrapper[4885]: I1002 02:33:33.640726 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7a715636-06d0-4508-a9d0-fe83f4c05d01-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"7a715636-06d0-4508-a9d0-fe83f4c05d01\") " pod="openstack/tempest-tests-tempest" Oct 02 02:33:33 crc kubenswrapper[4885]: I1002 02:33:33.641031 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/7a715636-06d0-4508-a9d0-fe83f4c05d01-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"7a715636-06d0-4508-a9d0-fe83f4c05d01\") " pod="openstack/tempest-tests-tempest" Oct 02 02:33:33 crc kubenswrapper[4885]: I1002 02:33:33.649213 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/7a715636-06d0-4508-a9d0-fe83f4c05d01-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"7a715636-06d0-4508-a9d0-fe83f4c05d01\") " pod="openstack/tempest-tests-tempest" Oct 02 02:33:33 crc kubenswrapper[4885]: I1002 02:33:33.655099 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vjxlh\" (UniqueName: \"kubernetes.io/projected/7a715636-06d0-4508-a9d0-fe83f4c05d01-kube-api-access-vjxlh\") pod \"tempest-tests-tempest\" (UID: \"7a715636-06d0-4508-a9d0-fe83f4c05d01\") " pod="openstack/tempest-tests-tempest" Oct 02 02:33:33 crc kubenswrapper[4885]: I1002 02:33:33.666037 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"tempest-tests-tempest\" (UID: \"7a715636-06d0-4508-a9d0-fe83f4c05d01\") " pod="openstack/tempest-tests-tempest" Oct 02 02:33:33 crc kubenswrapper[4885]: I1002 02:33:33.765913 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 02 02:33:34 crc kubenswrapper[4885]: I1002 02:33:34.259988 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Oct 02 02:33:35 crc kubenswrapper[4885]: I1002 02:33:35.046136 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"7a715636-06d0-4508-a9d0-fe83f4c05d01","Type":"ContainerStarted","Data":"fe8572d94ee5941888c10edcfa3a7c885868c94bf5ad32a55ddb21de0cfacc4a"} Oct 02 02:33:38 crc kubenswrapper[4885]: I1002 02:33:38.197490 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-9vhn2"] Oct 02 02:33:38 crc kubenswrapper[4885]: I1002 02:33:38.201150 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9vhn2" Oct 02 02:33:38 crc kubenswrapper[4885]: I1002 02:33:38.221033 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9vhn2"] Oct 02 02:33:38 crc kubenswrapper[4885]: I1002 02:33:38.328543 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8ba1223-c598-48cd-a282-d903090a4cda-catalog-content\") pod \"community-operators-9vhn2\" (UID: \"e8ba1223-c598-48cd-a282-d903090a4cda\") " pod="openshift-marketplace/community-operators-9vhn2" Oct 02 02:33:38 crc kubenswrapper[4885]: I1002 02:33:38.328675 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x6zws\" (UniqueName: \"kubernetes.io/projected/e8ba1223-c598-48cd-a282-d903090a4cda-kube-api-access-x6zws\") pod \"community-operators-9vhn2\" (UID: \"e8ba1223-c598-48cd-a282-d903090a4cda\") " pod="openshift-marketplace/community-operators-9vhn2" Oct 02 02:33:38 crc kubenswrapper[4885]: I1002 02:33:38.328893 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8ba1223-c598-48cd-a282-d903090a4cda-utilities\") pod \"community-operators-9vhn2\" (UID: \"e8ba1223-c598-48cd-a282-d903090a4cda\") " pod="openshift-marketplace/community-operators-9vhn2" Oct 02 02:33:38 crc kubenswrapper[4885]: I1002 02:33:38.430577 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8ba1223-c598-48cd-a282-d903090a4cda-utilities\") pod \"community-operators-9vhn2\" (UID: \"e8ba1223-c598-48cd-a282-d903090a4cda\") " pod="openshift-marketplace/community-operators-9vhn2" Oct 02 02:33:38 crc kubenswrapper[4885]: I1002 02:33:38.430707 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8ba1223-c598-48cd-a282-d903090a4cda-catalog-content\") pod \"community-operators-9vhn2\" (UID: \"e8ba1223-c598-48cd-a282-d903090a4cda\") " pod="openshift-marketplace/community-operators-9vhn2" Oct 02 02:33:38 crc kubenswrapper[4885]: I1002 02:33:38.430780 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x6zws\" (UniqueName: \"kubernetes.io/projected/e8ba1223-c598-48cd-a282-d903090a4cda-kube-api-access-x6zws\") pod \"community-operators-9vhn2\" (UID: \"e8ba1223-c598-48cd-a282-d903090a4cda\") " pod="openshift-marketplace/community-operators-9vhn2" Oct 02 02:33:38 crc kubenswrapper[4885]: I1002 02:33:38.431144 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8ba1223-c598-48cd-a282-d903090a4cda-utilities\") pod \"community-operators-9vhn2\" (UID: \"e8ba1223-c598-48cd-a282-d903090a4cda\") " pod="openshift-marketplace/community-operators-9vhn2" Oct 02 02:33:38 crc kubenswrapper[4885]: I1002 02:33:38.431399 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8ba1223-c598-48cd-a282-d903090a4cda-catalog-content\") pod \"community-operators-9vhn2\" (UID: \"e8ba1223-c598-48cd-a282-d903090a4cda\") " pod="openshift-marketplace/community-operators-9vhn2" Oct 02 02:33:38 crc kubenswrapper[4885]: I1002 02:33:38.451494 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x6zws\" (UniqueName: \"kubernetes.io/projected/e8ba1223-c598-48cd-a282-d903090a4cda-kube-api-access-x6zws\") pod \"community-operators-9vhn2\" (UID: \"e8ba1223-c598-48cd-a282-d903090a4cda\") " pod="openshift-marketplace/community-operators-9vhn2" Oct 02 02:33:38 crc kubenswrapper[4885]: I1002 02:33:38.527505 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9vhn2" Oct 02 02:33:39 crc kubenswrapper[4885]: I1002 02:33:39.632905 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9vhn2"] Oct 02 02:33:40 crc kubenswrapper[4885]: I1002 02:33:40.113931 4885 generic.go:334] "Generic (PLEG): container finished" podID="e8ba1223-c598-48cd-a282-d903090a4cda" containerID="6e5a4c50fea770bc5e8d116f89bf8bbe66060cc48f4d92be3e04adaeb9adb9b0" exitCode=0 Oct 02 02:33:40 crc kubenswrapper[4885]: I1002 02:33:40.113975 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9vhn2" event={"ID":"e8ba1223-c598-48cd-a282-d903090a4cda","Type":"ContainerDied","Data":"6e5a4c50fea770bc5e8d116f89bf8bbe66060cc48f4d92be3e04adaeb9adb9b0"} Oct 02 02:33:40 crc kubenswrapper[4885]: I1002 02:33:40.114354 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9vhn2" event={"ID":"e8ba1223-c598-48cd-a282-d903090a4cda","Type":"ContainerStarted","Data":"2e4f5cd8b3a5cef139f887c203509f37aea1b9f32f82b37917a8d8d9e06257d1"} Oct 02 02:33:41 crc kubenswrapper[4885]: I1002 02:33:41.126998 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9vhn2" event={"ID":"e8ba1223-c598-48cd-a282-d903090a4cda","Type":"ContainerStarted","Data":"9a034eed69c7dbc8c4a023d1dd10fa4af7629511c88415ad5e61fba1cc33f1ec"} Oct 02 02:33:42 crc kubenswrapper[4885]: I1002 02:33:42.138992 4885 generic.go:334] "Generic (PLEG): container finished" podID="e8ba1223-c598-48cd-a282-d903090a4cda" containerID="9a034eed69c7dbc8c4a023d1dd10fa4af7629511c88415ad5e61fba1cc33f1ec" exitCode=0 Oct 02 02:33:42 crc kubenswrapper[4885]: I1002 02:33:42.139290 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9vhn2" event={"ID":"e8ba1223-c598-48cd-a282-d903090a4cda","Type":"ContainerDied","Data":"9a034eed69c7dbc8c4a023d1dd10fa4af7629511c88415ad5e61fba1cc33f1ec"} Oct 02 02:33:43 crc kubenswrapper[4885]: I1002 02:33:43.148893 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9vhn2" event={"ID":"e8ba1223-c598-48cd-a282-d903090a4cda","Type":"ContainerStarted","Data":"cdb30fb183380f09db21ad8692f07c8d28d9f859a9214164c8a2e556c6765b61"} Oct 02 02:33:43 crc kubenswrapper[4885]: I1002 02:33:43.171703 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-9vhn2" podStartSLOduration=2.694532252 podStartE2EDuration="5.171661268s" podCreationTimestamp="2025-10-02 02:33:38 +0000 UTC" firstStartedPulling="2025-10-02 02:33:40.118105707 +0000 UTC m=+2808.929853106" lastFinishedPulling="2025-10-02 02:33:42.595234723 +0000 UTC m=+2811.406982122" observedRunningTime="2025-10-02 02:33:43.16679266 +0000 UTC m=+2811.978540079" watchObservedRunningTime="2025-10-02 02:33:43.171661268 +0000 UTC m=+2811.983408667" Oct 02 02:33:48 crc kubenswrapper[4885]: I1002 02:33:48.527792 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-9vhn2" Oct 02 02:33:48 crc kubenswrapper[4885]: I1002 02:33:48.528242 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-9vhn2" Oct 02 02:33:48 crc kubenswrapper[4885]: I1002 02:33:48.579433 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-9vhn2" Oct 02 02:33:49 crc kubenswrapper[4885]: I1002 02:33:49.267193 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-9vhn2" Oct 02 02:33:51 crc kubenswrapper[4885]: I1002 02:33:51.158444 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9vhn2"] Oct 02 02:33:51 crc kubenswrapper[4885]: I1002 02:33:51.246839 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-9vhn2" podUID="e8ba1223-c598-48cd-a282-d903090a4cda" containerName="registry-server" containerID="cri-o://cdb30fb183380f09db21ad8692f07c8d28d9f859a9214164c8a2e556c6765b61" gracePeriod=2 Oct 02 02:33:52 crc kubenswrapper[4885]: I1002 02:33:52.272147 4885 generic.go:334] "Generic (PLEG): container finished" podID="e8ba1223-c598-48cd-a282-d903090a4cda" containerID="cdb30fb183380f09db21ad8692f07c8d28d9f859a9214164c8a2e556c6765b61" exitCode=0 Oct 02 02:33:52 crc kubenswrapper[4885]: I1002 02:33:52.272177 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9vhn2" event={"ID":"e8ba1223-c598-48cd-a282-d903090a4cda","Type":"ContainerDied","Data":"cdb30fb183380f09db21ad8692f07c8d28d9f859a9214164c8a2e556c6765b61"} Oct 02 02:33:58 crc kubenswrapper[4885]: E1002 02:33:58.529166 4885 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of cdb30fb183380f09db21ad8692f07c8d28d9f859a9214164c8a2e556c6765b61 is running failed: container process not found" containerID="cdb30fb183380f09db21ad8692f07c8d28d9f859a9214164c8a2e556c6765b61" cmd=["grpc_health_probe","-addr=:50051"] Oct 02 02:33:58 crc kubenswrapper[4885]: E1002 02:33:58.530142 4885 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of cdb30fb183380f09db21ad8692f07c8d28d9f859a9214164c8a2e556c6765b61 is running failed: container process not found" containerID="cdb30fb183380f09db21ad8692f07c8d28d9f859a9214164c8a2e556c6765b61" cmd=["grpc_health_probe","-addr=:50051"] Oct 02 02:33:58 crc kubenswrapper[4885]: E1002 02:33:58.537122 4885 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of cdb30fb183380f09db21ad8692f07c8d28d9f859a9214164c8a2e556c6765b61 is running failed: container process not found" containerID="cdb30fb183380f09db21ad8692f07c8d28d9f859a9214164c8a2e556c6765b61" cmd=["grpc_health_probe","-addr=:50051"] Oct 02 02:33:58 crc kubenswrapper[4885]: E1002 02:33:58.537241 4885 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of cdb30fb183380f09db21ad8692f07c8d28d9f859a9214164c8a2e556c6765b61 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/community-operators-9vhn2" podUID="e8ba1223-c598-48cd-a282-d903090a4cda" containerName="registry-server" Oct 02 02:34:04 crc kubenswrapper[4885]: E1002 02:34:04.500050 4885 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Oct 02 02:34:04 crc kubenswrapper[4885]: E1002 02:34:04.501150 4885 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vjxlh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(7a715636-06d0-4508-a9d0-fe83f4c05d01): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 02:34:04 crc kubenswrapper[4885]: E1002 02:34:04.502575 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="7a715636-06d0-4508-a9d0-fe83f4c05d01" Oct 02 02:34:04 crc kubenswrapper[4885]: I1002 02:34:04.798949 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9vhn2" Oct 02 02:34:04 crc kubenswrapper[4885]: I1002 02:34:04.908096 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8ba1223-c598-48cd-a282-d903090a4cda-catalog-content\") pod \"e8ba1223-c598-48cd-a282-d903090a4cda\" (UID: \"e8ba1223-c598-48cd-a282-d903090a4cda\") " Oct 02 02:34:04 crc kubenswrapper[4885]: I1002 02:34:04.908187 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8ba1223-c598-48cd-a282-d903090a4cda-utilities\") pod \"e8ba1223-c598-48cd-a282-d903090a4cda\" (UID: \"e8ba1223-c598-48cd-a282-d903090a4cda\") " Oct 02 02:34:04 crc kubenswrapper[4885]: I1002 02:34:04.908456 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x6zws\" (UniqueName: \"kubernetes.io/projected/e8ba1223-c598-48cd-a282-d903090a4cda-kube-api-access-x6zws\") pod \"e8ba1223-c598-48cd-a282-d903090a4cda\" (UID: \"e8ba1223-c598-48cd-a282-d903090a4cda\") " Oct 02 02:34:04 crc kubenswrapper[4885]: I1002 02:34:04.909989 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e8ba1223-c598-48cd-a282-d903090a4cda-utilities" (OuterVolumeSpecName: "utilities") pod "e8ba1223-c598-48cd-a282-d903090a4cda" (UID: "e8ba1223-c598-48cd-a282-d903090a4cda"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:34:04 crc kubenswrapper[4885]: I1002 02:34:04.920134 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8ba1223-c598-48cd-a282-d903090a4cda-kube-api-access-x6zws" (OuterVolumeSpecName: "kube-api-access-x6zws") pod "e8ba1223-c598-48cd-a282-d903090a4cda" (UID: "e8ba1223-c598-48cd-a282-d903090a4cda"). InnerVolumeSpecName "kube-api-access-x6zws". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:34:04 crc kubenswrapper[4885]: I1002 02:34:04.969352 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e8ba1223-c598-48cd-a282-d903090a4cda-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e8ba1223-c598-48cd-a282-d903090a4cda" (UID: "e8ba1223-c598-48cd-a282-d903090a4cda"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:34:05 crc kubenswrapper[4885]: I1002 02:34:05.010857 4885 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8ba1223-c598-48cd-a282-d903090a4cda-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 02:34:05 crc kubenswrapper[4885]: I1002 02:34:05.010897 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x6zws\" (UniqueName: \"kubernetes.io/projected/e8ba1223-c598-48cd-a282-d903090a4cda-kube-api-access-x6zws\") on node \"crc\" DevicePath \"\"" Oct 02 02:34:05 crc kubenswrapper[4885]: I1002 02:34:05.010912 4885 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8ba1223-c598-48cd-a282-d903090a4cda-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 02:34:05 crc kubenswrapper[4885]: I1002 02:34:05.412222 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9vhn2" Oct 02 02:34:05 crc kubenswrapper[4885]: I1002 02:34:05.412219 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9vhn2" event={"ID":"e8ba1223-c598-48cd-a282-d903090a4cda","Type":"ContainerDied","Data":"2e4f5cd8b3a5cef139f887c203509f37aea1b9f32f82b37917a8d8d9e06257d1"} Oct 02 02:34:05 crc kubenswrapper[4885]: I1002 02:34:05.412643 4885 scope.go:117] "RemoveContainer" containerID="cdb30fb183380f09db21ad8692f07c8d28d9f859a9214164c8a2e556c6765b61" Oct 02 02:34:05 crc kubenswrapper[4885]: E1002 02:34:05.415627 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="7a715636-06d0-4508-a9d0-fe83f4c05d01" Oct 02 02:34:05 crc kubenswrapper[4885]: I1002 02:34:05.452305 4885 scope.go:117] "RemoveContainer" containerID="9a034eed69c7dbc8c4a023d1dd10fa4af7629511c88415ad5e61fba1cc33f1ec" Oct 02 02:34:05 crc kubenswrapper[4885]: I1002 02:34:05.479492 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9vhn2"] Oct 02 02:34:05 crc kubenswrapper[4885]: I1002 02:34:05.492888 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-9vhn2"] Oct 02 02:34:05 crc kubenswrapper[4885]: I1002 02:34:05.499820 4885 scope.go:117] "RemoveContainer" containerID="6e5a4c50fea770bc5e8d116f89bf8bbe66060cc48f4d92be3e04adaeb9adb9b0" Oct 02 02:34:06 crc kubenswrapper[4885]: I1002 02:34:06.065327 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e8ba1223-c598-48cd-a282-d903090a4cda" path="/var/lib/kubelet/pods/e8ba1223-c598-48cd-a282-d903090a4cda/volumes" Oct 02 02:34:17 crc kubenswrapper[4885]: I1002 02:34:17.464523 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Oct 02 02:34:18 crc kubenswrapper[4885]: I1002 02:34:18.545648 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"7a715636-06d0-4508-a9d0-fe83f4c05d01","Type":"ContainerStarted","Data":"226fd3dedfa1769e7bf88d810b0326fa5df173290057a082a896836458a0ef54"} Oct 02 02:34:18 crc kubenswrapper[4885]: I1002 02:34:18.575101 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=3.387816782 podStartE2EDuration="46.575077274s" podCreationTimestamp="2025-10-02 02:33:32 +0000 UTC" firstStartedPulling="2025-10-02 02:33:34.274713664 +0000 UTC m=+2803.086461093" lastFinishedPulling="2025-10-02 02:34:17.461974186 +0000 UTC m=+2846.273721585" observedRunningTime="2025-10-02 02:34:18.568480153 +0000 UTC m=+2847.380227562" watchObservedRunningTime="2025-10-02 02:34:18.575077274 +0000 UTC m=+2847.386824683" Oct 02 02:34:43 crc kubenswrapper[4885]: I1002 02:34:43.265709 4885 patch_prober.go:28] interesting pod/machine-config-daemon-rttx8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 02:34:43 crc kubenswrapper[4885]: I1002 02:34:43.268082 4885 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 02:35:13 crc kubenswrapper[4885]: I1002 02:35:13.265825 4885 patch_prober.go:28] interesting pod/machine-config-daemon-rttx8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 02:35:13 crc kubenswrapper[4885]: I1002 02:35:13.266623 4885 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 02:35:43 crc kubenswrapper[4885]: I1002 02:35:43.266042 4885 patch_prober.go:28] interesting pod/machine-config-daemon-rttx8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 02:35:43 crc kubenswrapper[4885]: I1002 02:35:43.266732 4885 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 02:35:43 crc kubenswrapper[4885]: I1002 02:35:43.266800 4885 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" Oct 02 02:35:43 crc kubenswrapper[4885]: I1002 02:35:43.267855 4885 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"02916a559cc5f6ab8f5bc76ee73c22c5f71f201c585eeb74e3c14d51fc63cccb"} pod="openshift-machine-config-operator/machine-config-daemon-rttx8" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 02:35:43 crc kubenswrapper[4885]: I1002 02:35:43.267938 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" containerName="machine-config-daemon" containerID="cri-o://02916a559cc5f6ab8f5bc76ee73c22c5f71f201c585eeb74e3c14d51fc63cccb" gracePeriod=600 Oct 02 02:35:43 crc kubenswrapper[4885]: E1002 02:35:43.404187 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:35:43 crc kubenswrapper[4885]: I1002 02:35:43.500432 4885 generic.go:334] "Generic (PLEG): container finished" podID="d169c8ba-000a-4218-8e3b-4ae53035b110" containerID="02916a559cc5f6ab8f5bc76ee73c22c5f71f201c585eeb74e3c14d51fc63cccb" exitCode=0 Oct 02 02:35:43 crc kubenswrapper[4885]: I1002 02:35:43.500503 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" event={"ID":"d169c8ba-000a-4218-8e3b-4ae53035b110","Type":"ContainerDied","Data":"02916a559cc5f6ab8f5bc76ee73c22c5f71f201c585eeb74e3c14d51fc63cccb"} Oct 02 02:35:43 crc kubenswrapper[4885]: I1002 02:35:43.500552 4885 scope.go:117] "RemoveContainer" containerID="5bbcad619e10a916cd09b9cfd4ad5c9334559ab939f077b917dc9ae849fe425f" Oct 02 02:35:43 crc kubenswrapper[4885]: I1002 02:35:43.501324 4885 scope.go:117] "RemoveContainer" containerID="02916a559cc5f6ab8f5bc76ee73c22c5f71f201c585eeb74e3c14d51fc63cccb" Oct 02 02:35:43 crc kubenswrapper[4885]: E1002 02:35:43.501722 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:35:54 crc kubenswrapper[4885]: I1002 02:35:54.046681 4885 scope.go:117] "RemoveContainer" containerID="02916a559cc5f6ab8f5bc76ee73c22c5f71f201c585eeb74e3c14d51fc63cccb" Oct 02 02:35:54 crc kubenswrapper[4885]: E1002 02:35:54.048201 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:36:08 crc kubenswrapper[4885]: I1002 02:36:08.047467 4885 scope.go:117] "RemoveContainer" containerID="02916a559cc5f6ab8f5bc76ee73c22c5f71f201c585eeb74e3c14d51fc63cccb" Oct 02 02:36:08 crc kubenswrapper[4885]: E1002 02:36:08.048580 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:36:14 crc kubenswrapper[4885]: I1002 02:36:14.773964 4885 scope.go:117] "RemoveContainer" containerID="5fda54bb5c79929ec4e8cf0f421a8d466ed253b2e4bd6a3bc103b032de90cab5" Oct 02 02:36:14 crc kubenswrapper[4885]: I1002 02:36:14.817006 4885 scope.go:117] "RemoveContainer" containerID="b055bae7125b0ef7fca54e932015c937bb8ddba6c863be9f5a1e76ecbb06426b" Oct 02 02:36:14 crc kubenswrapper[4885]: I1002 02:36:14.886401 4885 scope.go:117] "RemoveContainer" containerID="5041923797db7b24a57bf7d30e90a1eef25187cf83a34edfeef39345ab2c10d6" Oct 02 02:36:22 crc kubenswrapper[4885]: I1002 02:36:22.062244 4885 scope.go:117] "RemoveContainer" containerID="02916a559cc5f6ab8f5bc76ee73c22c5f71f201c585eeb74e3c14d51fc63cccb" Oct 02 02:36:22 crc kubenswrapper[4885]: E1002 02:36:22.063635 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:36:35 crc kubenswrapper[4885]: I1002 02:36:35.046954 4885 scope.go:117] "RemoveContainer" containerID="02916a559cc5f6ab8f5bc76ee73c22c5f71f201c585eeb74e3c14d51fc63cccb" Oct 02 02:36:35 crc kubenswrapper[4885]: E1002 02:36:35.047925 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:36:47 crc kubenswrapper[4885]: I1002 02:36:47.048589 4885 scope.go:117] "RemoveContainer" containerID="02916a559cc5f6ab8f5bc76ee73c22c5f71f201c585eeb74e3c14d51fc63cccb" Oct 02 02:36:47 crc kubenswrapper[4885]: E1002 02:36:47.049666 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:37:01 crc kubenswrapper[4885]: I1002 02:37:01.048597 4885 scope.go:117] "RemoveContainer" containerID="02916a559cc5f6ab8f5bc76ee73c22c5f71f201c585eeb74e3c14d51fc63cccb" Oct 02 02:37:01 crc kubenswrapper[4885]: E1002 02:37:01.049697 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:37:14 crc kubenswrapper[4885]: I1002 02:37:14.047127 4885 scope.go:117] "RemoveContainer" containerID="02916a559cc5f6ab8f5bc76ee73c22c5f71f201c585eeb74e3c14d51fc63cccb" Oct 02 02:37:14 crc kubenswrapper[4885]: E1002 02:37:14.048019 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:37:27 crc kubenswrapper[4885]: I1002 02:37:27.048480 4885 scope.go:117] "RemoveContainer" containerID="02916a559cc5f6ab8f5bc76ee73c22c5f71f201c585eeb74e3c14d51fc63cccb" Oct 02 02:37:27 crc kubenswrapper[4885]: E1002 02:37:27.049534 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:37:39 crc kubenswrapper[4885]: I1002 02:37:39.046765 4885 scope.go:117] "RemoveContainer" containerID="02916a559cc5f6ab8f5bc76ee73c22c5f71f201c585eeb74e3c14d51fc63cccb" Oct 02 02:37:39 crc kubenswrapper[4885]: E1002 02:37:39.047812 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:37:54 crc kubenswrapper[4885]: I1002 02:37:54.046721 4885 scope.go:117] "RemoveContainer" containerID="02916a559cc5f6ab8f5bc76ee73c22c5f71f201c585eeb74e3c14d51fc63cccb" Oct 02 02:37:54 crc kubenswrapper[4885]: E1002 02:37:54.047629 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:38:08 crc kubenswrapper[4885]: I1002 02:38:08.047198 4885 scope.go:117] "RemoveContainer" containerID="02916a559cc5f6ab8f5bc76ee73c22c5f71f201c585eeb74e3c14d51fc63cccb" Oct 02 02:38:08 crc kubenswrapper[4885]: E1002 02:38:08.047911 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:38:20 crc kubenswrapper[4885]: I1002 02:38:20.046990 4885 scope.go:117] "RemoveContainer" containerID="02916a559cc5f6ab8f5bc76ee73c22c5f71f201c585eeb74e3c14d51fc63cccb" Oct 02 02:38:20 crc kubenswrapper[4885]: E1002 02:38:20.047828 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:38:31 crc kubenswrapper[4885]: I1002 02:38:31.046863 4885 scope.go:117] "RemoveContainer" containerID="02916a559cc5f6ab8f5bc76ee73c22c5f71f201c585eeb74e3c14d51fc63cccb" Oct 02 02:38:31 crc kubenswrapper[4885]: E1002 02:38:31.047633 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:38:45 crc kubenswrapper[4885]: I1002 02:38:45.047577 4885 scope.go:117] "RemoveContainer" containerID="02916a559cc5f6ab8f5bc76ee73c22c5f71f201c585eeb74e3c14d51fc63cccb" Oct 02 02:38:45 crc kubenswrapper[4885]: E1002 02:38:45.048633 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:38:58 crc kubenswrapper[4885]: I1002 02:38:58.047150 4885 scope.go:117] "RemoveContainer" containerID="02916a559cc5f6ab8f5bc76ee73c22c5f71f201c585eeb74e3c14d51fc63cccb" Oct 02 02:38:58 crc kubenswrapper[4885]: E1002 02:38:58.047907 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:39:12 crc kubenswrapper[4885]: I1002 02:39:12.059857 4885 scope.go:117] "RemoveContainer" containerID="02916a559cc5f6ab8f5bc76ee73c22c5f71f201c585eeb74e3c14d51fc63cccb" Oct 02 02:39:12 crc kubenswrapper[4885]: E1002 02:39:12.060758 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:39:27 crc kubenswrapper[4885]: I1002 02:39:27.046830 4885 scope.go:117] "RemoveContainer" containerID="02916a559cc5f6ab8f5bc76ee73c22c5f71f201c585eeb74e3c14d51fc63cccb" Oct 02 02:39:27 crc kubenswrapper[4885]: E1002 02:39:27.047847 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:39:40 crc kubenswrapper[4885]: I1002 02:39:40.046790 4885 scope.go:117] "RemoveContainer" containerID="02916a559cc5f6ab8f5bc76ee73c22c5f71f201c585eeb74e3c14d51fc63cccb" Oct 02 02:39:40 crc kubenswrapper[4885]: E1002 02:39:40.047689 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:39:51 crc kubenswrapper[4885]: I1002 02:39:51.046758 4885 scope.go:117] "RemoveContainer" containerID="02916a559cc5f6ab8f5bc76ee73c22c5f71f201c585eeb74e3c14d51fc63cccb" Oct 02 02:39:51 crc kubenswrapper[4885]: E1002 02:39:51.049354 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:40:02 crc kubenswrapper[4885]: I1002 02:40:02.065648 4885 scope.go:117] "RemoveContainer" containerID="02916a559cc5f6ab8f5bc76ee73c22c5f71f201c585eeb74e3c14d51fc63cccb" Oct 02 02:40:02 crc kubenswrapper[4885]: E1002 02:40:02.069053 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:40:15 crc kubenswrapper[4885]: I1002 02:40:15.047109 4885 scope.go:117] "RemoveContainer" containerID="02916a559cc5f6ab8f5bc76ee73c22c5f71f201c585eeb74e3c14d51fc63cccb" Oct 02 02:40:15 crc kubenswrapper[4885]: E1002 02:40:15.048018 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:40:26 crc kubenswrapper[4885]: I1002 02:40:26.047058 4885 scope.go:117] "RemoveContainer" containerID="02916a559cc5f6ab8f5bc76ee73c22c5f71f201c585eeb74e3c14d51fc63cccb" Oct 02 02:40:26 crc kubenswrapper[4885]: E1002 02:40:26.048308 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:40:38 crc kubenswrapper[4885]: I1002 02:40:38.047598 4885 scope.go:117] "RemoveContainer" containerID="02916a559cc5f6ab8f5bc76ee73c22c5f71f201c585eeb74e3c14d51fc63cccb" Oct 02 02:40:38 crc kubenswrapper[4885]: E1002 02:40:38.049036 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:40:51 crc kubenswrapper[4885]: I1002 02:40:51.046958 4885 scope.go:117] "RemoveContainer" containerID="02916a559cc5f6ab8f5bc76ee73c22c5f71f201c585eeb74e3c14d51fc63cccb" Oct 02 02:40:51 crc kubenswrapper[4885]: I1002 02:40:51.992200 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" event={"ID":"d169c8ba-000a-4218-8e3b-4ae53035b110","Type":"ContainerStarted","Data":"263b36668b739af6c8b8dd19175897e6595f755c49d0d124cba769111393cb93"} Oct 02 02:42:47 crc kubenswrapper[4885]: I1002 02:42:47.116565 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-rpf4b"] Oct 02 02:42:47 crc kubenswrapper[4885]: E1002 02:42:47.117786 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8ba1223-c598-48cd-a282-d903090a4cda" containerName="registry-server" Oct 02 02:42:47 crc kubenswrapper[4885]: I1002 02:42:47.117806 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8ba1223-c598-48cd-a282-d903090a4cda" containerName="registry-server" Oct 02 02:42:47 crc kubenswrapper[4885]: E1002 02:42:47.117830 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8ba1223-c598-48cd-a282-d903090a4cda" containerName="extract-utilities" Oct 02 02:42:47 crc kubenswrapper[4885]: I1002 02:42:47.117841 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8ba1223-c598-48cd-a282-d903090a4cda" containerName="extract-utilities" Oct 02 02:42:47 crc kubenswrapper[4885]: E1002 02:42:47.117858 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8ba1223-c598-48cd-a282-d903090a4cda" containerName="extract-content" Oct 02 02:42:47 crc kubenswrapper[4885]: I1002 02:42:47.117867 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8ba1223-c598-48cd-a282-d903090a4cda" containerName="extract-content" Oct 02 02:42:47 crc kubenswrapper[4885]: I1002 02:42:47.118110 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8ba1223-c598-48cd-a282-d903090a4cda" containerName="registry-server" Oct 02 02:42:47 crc kubenswrapper[4885]: I1002 02:42:47.120183 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rpf4b" Oct 02 02:42:47 crc kubenswrapper[4885]: I1002 02:42:47.148489 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rpf4b"] Oct 02 02:42:47 crc kubenswrapper[4885]: I1002 02:42:47.185742 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ngwfw\" (UniqueName: \"kubernetes.io/projected/3ad269d3-74cd-4a34-8ef6-e3d25cbaca5d-kube-api-access-ngwfw\") pod \"redhat-operators-rpf4b\" (UID: \"3ad269d3-74cd-4a34-8ef6-e3d25cbaca5d\") " pod="openshift-marketplace/redhat-operators-rpf4b" Oct 02 02:42:47 crc kubenswrapper[4885]: I1002 02:42:47.187205 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3ad269d3-74cd-4a34-8ef6-e3d25cbaca5d-utilities\") pod \"redhat-operators-rpf4b\" (UID: \"3ad269d3-74cd-4a34-8ef6-e3d25cbaca5d\") " pod="openshift-marketplace/redhat-operators-rpf4b" Oct 02 02:42:47 crc kubenswrapper[4885]: I1002 02:42:47.187584 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3ad269d3-74cd-4a34-8ef6-e3d25cbaca5d-catalog-content\") pod \"redhat-operators-rpf4b\" (UID: \"3ad269d3-74cd-4a34-8ef6-e3d25cbaca5d\") " pod="openshift-marketplace/redhat-operators-rpf4b" Oct 02 02:42:47 crc kubenswrapper[4885]: I1002 02:42:47.289333 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3ad269d3-74cd-4a34-8ef6-e3d25cbaca5d-catalog-content\") pod \"redhat-operators-rpf4b\" (UID: \"3ad269d3-74cd-4a34-8ef6-e3d25cbaca5d\") " pod="openshift-marketplace/redhat-operators-rpf4b" Oct 02 02:42:47 crc kubenswrapper[4885]: I1002 02:42:47.289421 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ngwfw\" (UniqueName: \"kubernetes.io/projected/3ad269d3-74cd-4a34-8ef6-e3d25cbaca5d-kube-api-access-ngwfw\") pod \"redhat-operators-rpf4b\" (UID: \"3ad269d3-74cd-4a34-8ef6-e3d25cbaca5d\") " pod="openshift-marketplace/redhat-operators-rpf4b" Oct 02 02:42:47 crc kubenswrapper[4885]: I1002 02:42:47.289548 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3ad269d3-74cd-4a34-8ef6-e3d25cbaca5d-utilities\") pod \"redhat-operators-rpf4b\" (UID: \"3ad269d3-74cd-4a34-8ef6-e3d25cbaca5d\") " pod="openshift-marketplace/redhat-operators-rpf4b" Oct 02 02:42:47 crc kubenswrapper[4885]: I1002 02:42:47.290121 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3ad269d3-74cd-4a34-8ef6-e3d25cbaca5d-utilities\") pod \"redhat-operators-rpf4b\" (UID: \"3ad269d3-74cd-4a34-8ef6-e3d25cbaca5d\") " pod="openshift-marketplace/redhat-operators-rpf4b" Oct 02 02:42:47 crc kubenswrapper[4885]: I1002 02:42:47.290567 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3ad269d3-74cd-4a34-8ef6-e3d25cbaca5d-catalog-content\") pod \"redhat-operators-rpf4b\" (UID: \"3ad269d3-74cd-4a34-8ef6-e3d25cbaca5d\") " pod="openshift-marketplace/redhat-operators-rpf4b" Oct 02 02:42:47 crc kubenswrapper[4885]: I1002 02:42:47.316248 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ngwfw\" (UniqueName: \"kubernetes.io/projected/3ad269d3-74cd-4a34-8ef6-e3d25cbaca5d-kube-api-access-ngwfw\") pod \"redhat-operators-rpf4b\" (UID: \"3ad269d3-74cd-4a34-8ef6-e3d25cbaca5d\") " pod="openshift-marketplace/redhat-operators-rpf4b" Oct 02 02:42:47 crc kubenswrapper[4885]: I1002 02:42:47.473033 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rpf4b" Oct 02 02:42:47 crc kubenswrapper[4885]: I1002 02:42:47.982384 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rpf4b"] Oct 02 02:42:48 crc kubenswrapper[4885]: I1002 02:42:48.386349 4885 generic.go:334] "Generic (PLEG): container finished" podID="3ad269d3-74cd-4a34-8ef6-e3d25cbaca5d" containerID="2893ad13a5715ab344d73ccd6d4ac6b2471edace707c62caac1e3b553e5b8e79" exitCode=0 Oct 02 02:42:48 crc kubenswrapper[4885]: I1002 02:42:48.386544 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rpf4b" event={"ID":"3ad269d3-74cd-4a34-8ef6-e3d25cbaca5d","Type":"ContainerDied","Data":"2893ad13a5715ab344d73ccd6d4ac6b2471edace707c62caac1e3b553e5b8e79"} Oct 02 02:42:48 crc kubenswrapper[4885]: I1002 02:42:48.386672 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rpf4b" event={"ID":"3ad269d3-74cd-4a34-8ef6-e3d25cbaca5d","Type":"ContainerStarted","Data":"ec2a9fe1e1f3817475154e43a02ac14dabf2a5f5055662a3cb07d9e782623eea"} Oct 02 02:42:48 crc kubenswrapper[4885]: I1002 02:42:48.388527 4885 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 02:42:50 crc kubenswrapper[4885]: I1002 02:42:50.408299 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rpf4b" event={"ID":"3ad269d3-74cd-4a34-8ef6-e3d25cbaca5d","Type":"ContainerStarted","Data":"76566178bfd7d0dde88ff50a9b815b2b2d8f636c8b948862a0f581334a379e71"} Oct 02 02:42:51 crc kubenswrapper[4885]: I1002 02:42:51.422517 4885 generic.go:334] "Generic (PLEG): container finished" podID="3ad269d3-74cd-4a34-8ef6-e3d25cbaca5d" containerID="76566178bfd7d0dde88ff50a9b815b2b2d8f636c8b948862a0f581334a379e71" exitCode=0 Oct 02 02:42:51 crc kubenswrapper[4885]: I1002 02:42:51.422754 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rpf4b" event={"ID":"3ad269d3-74cd-4a34-8ef6-e3d25cbaca5d","Type":"ContainerDied","Data":"76566178bfd7d0dde88ff50a9b815b2b2d8f636c8b948862a0f581334a379e71"} Oct 02 02:42:52 crc kubenswrapper[4885]: I1002 02:42:52.435879 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rpf4b" event={"ID":"3ad269d3-74cd-4a34-8ef6-e3d25cbaca5d","Type":"ContainerStarted","Data":"39e5b9b3ad363e24e3a752572fb5e1c0f16b6065f3c6e67982111e5984685935"} Oct 02 02:42:52 crc kubenswrapper[4885]: I1002 02:42:52.459877 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-rpf4b" podStartSLOduration=1.988740574 podStartE2EDuration="5.459852808s" podCreationTimestamp="2025-10-02 02:42:47 +0000 UTC" firstStartedPulling="2025-10-02 02:42:48.3883186 +0000 UTC m=+3357.200065999" lastFinishedPulling="2025-10-02 02:42:51.859430824 +0000 UTC m=+3360.671178233" observedRunningTime="2025-10-02 02:42:52.454852245 +0000 UTC m=+3361.266599674" watchObservedRunningTime="2025-10-02 02:42:52.459852808 +0000 UTC m=+3361.271600247" Oct 02 02:42:57 crc kubenswrapper[4885]: I1002 02:42:57.473796 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-rpf4b" Oct 02 02:42:57 crc kubenswrapper[4885]: I1002 02:42:57.474420 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-rpf4b" Oct 02 02:42:57 crc kubenswrapper[4885]: I1002 02:42:57.561180 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-rpf4b" Oct 02 02:42:58 crc kubenswrapper[4885]: I1002 02:42:58.579962 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-rpf4b" Oct 02 02:42:58 crc kubenswrapper[4885]: I1002 02:42:58.650250 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rpf4b"] Oct 02 02:43:00 crc kubenswrapper[4885]: I1002 02:43:00.516613 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-rpf4b" podUID="3ad269d3-74cd-4a34-8ef6-e3d25cbaca5d" containerName="registry-server" containerID="cri-o://39e5b9b3ad363e24e3a752572fb5e1c0f16b6065f3c6e67982111e5984685935" gracePeriod=2 Oct 02 02:43:01 crc kubenswrapper[4885]: I1002 02:43:01.068996 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rpf4b" Oct 02 02:43:01 crc kubenswrapper[4885]: I1002 02:43:01.173006 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngwfw\" (UniqueName: \"kubernetes.io/projected/3ad269d3-74cd-4a34-8ef6-e3d25cbaca5d-kube-api-access-ngwfw\") pod \"3ad269d3-74cd-4a34-8ef6-e3d25cbaca5d\" (UID: \"3ad269d3-74cd-4a34-8ef6-e3d25cbaca5d\") " Oct 02 02:43:01 crc kubenswrapper[4885]: I1002 02:43:01.173150 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3ad269d3-74cd-4a34-8ef6-e3d25cbaca5d-catalog-content\") pod \"3ad269d3-74cd-4a34-8ef6-e3d25cbaca5d\" (UID: \"3ad269d3-74cd-4a34-8ef6-e3d25cbaca5d\") " Oct 02 02:43:01 crc kubenswrapper[4885]: I1002 02:43:01.173307 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3ad269d3-74cd-4a34-8ef6-e3d25cbaca5d-utilities\") pod \"3ad269d3-74cd-4a34-8ef6-e3d25cbaca5d\" (UID: \"3ad269d3-74cd-4a34-8ef6-e3d25cbaca5d\") " Oct 02 02:43:01 crc kubenswrapper[4885]: I1002 02:43:01.174435 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3ad269d3-74cd-4a34-8ef6-e3d25cbaca5d-utilities" (OuterVolumeSpecName: "utilities") pod "3ad269d3-74cd-4a34-8ef6-e3d25cbaca5d" (UID: "3ad269d3-74cd-4a34-8ef6-e3d25cbaca5d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:43:01 crc kubenswrapper[4885]: I1002 02:43:01.174832 4885 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3ad269d3-74cd-4a34-8ef6-e3d25cbaca5d-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 02:43:01 crc kubenswrapper[4885]: I1002 02:43:01.181775 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ad269d3-74cd-4a34-8ef6-e3d25cbaca5d-kube-api-access-ngwfw" (OuterVolumeSpecName: "kube-api-access-ngwfw") pod "3ad269d3-74cd-4a34-8ef6-e3d25cbaca5d" (UID: "3ad269d3-74cd-4a34-8ef6-e3d25cbaca5d"). InnerVolumeSpecName "kube-api-access-ngwfw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:43:01 crc kubenswrapper[4885]: I1002 02:43:01.272649 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3ad269d3-74cd-4a34-8ef6-e3d25cbaca5d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3ad269d3-74cd-4a34-8ef6-e3d25cbaca5d" (UID: "3ad269d3-74cd-4a34-8ef6-e3d25cbaca5d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:43:01 crc kubenswrapper[4885]: I1002 02:43:01.276551 4885 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3ad269d3-74cd-4a34-8ef6-e3d25cbaca5d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 02:43:01 crc kubenswrapper[4885]: I1002 02:43:01.276673 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngwfw\" (UniqueName: \"kubernetes.io/projected/3ad269d3-74cd-4a34-8ef6-e3d25cbaca5d-kube-api-access-ngwfw\") on node \"crc\" DevicePath \"\"" Oct 02 02:43:01 crc kubenswrapper[4885]: I1002 02:43:01.527395 4885 generic.go:334] "Generic (PLEG): container finished" podID="3ad269d3-74cd-4a34-8ef6-e3d25cbaca5d" containerID="39e5b9b3ad363e24e3a752572fb5e1c0f16b6065f3c6e67982111e5984685935" exitCode=0 Oct 02 02:43:01 crc kubenswrapper[4885]: I1002 02:43:01.527447 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rpf4b" event={"ID":"3ad269d3-74cd-4a34-8ef6-e3d25cbaca5d","Type":"ContainerDied","Data":"39e5b9b3ad363e24e3a752572fb5e1c0f16b6065f3c6e67982111e5984685935"} Oct 02 02:43:01 crc kubenswrapper[4885]: I1002 02:43:01.527481 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rpf4b" event={"ID":"3ad269d3-74cd-4a34-8ef6-e3d25cbaca5d","Type":"ContainerDied","Data":"ec2a9fe1e1f3817475154e43a02ac14dabf2a5f5055662a3cb07d9e782623eea"} Oct 02 02:43:01 crc kubenswrapper[4885]: I1002 02:43:01.527505 4885 scope.go:117] "RemoveContainer" containerID="39e5b9b3ad363e24e3a752572fb5e1c0f16b6065f3c6e67982111e5984685935" Oct 02 02:43:01 crc kubenswrapper[4885]: I1002 02:43:01.527512 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rpf4b" Oct 02 02:43:01 crc kubenswrapper[4885]: I1002 02:43:01.552175 4885 scope.go:117] "RemoveContainer" containerID="76566178bfd7d0dde88ff50a9b815b2b2d8f636c8b948862a0f581334a379e71" Oct 02 02:43:01 crc kubenswrapper[4885]: I1002 02:43:01.594657 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rpf4b"] Oct 02 02:43:01 crc kubenswrapper[4885]: I1002 02:43:01.594718 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-rpf4b"] Oct 02 02:43:01 crc kubenswrapper[4885]: I1002 02:43:01.621731 4885 scope.go:117] "RemoveContainer" containerID="2893ad13a5715ab344d73ccd6d4ac6b2471edace707c62caac1e3b553e5b8e79" Oct 02 02:43:01 crc kubenswrapper[4885]: I1002 02:43:01.654726 4885 scope.go:117] "RemoveContainer" containerID="39e5b9b3ad363e24e3a752572fb5e1c0f16b6065f3c6e67982111e5984685935" Oct 02 02:43:01 crc kubenswrapper[4885]: E1002 02:43:01.655225 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"39e5b9b3ad363e24e3a752572fb5e1c0f16b6065f3c6e67982111e5984685935\": container with ID starting with 39e5b9b3ad363e24e3a752572fb5e1c0f16b6065f3c6e67982111e5984685935 not found: ID does not exist" containerID="39e5b9b3ad363e24e3a752572fb5e1c0f16b6065f3c6e67982111e5984685935" Oct 02 02:43:01 crc kubenswrapper[4885]: I1002 02:43:01.655284 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39e5b9b3ad363e24e3a752572fb5e1c0f16b6065f3c6e67982111e5984685935"} err="failed to get container status \"39e5b9b3ad363e24e3a752572fb5e1c0f16b6065f3c6e67982111e5984685935\": rpc error: code = NotFound desc = could not find container \"39e5b9b3ad363e24e3a752572fb5e1c0f16b6065f3c6e67982111e5984685935\": container with ID starting with 39e5b9b3ad363e24e3a752572fb5e1c0f16b6065f3c6e67982111e5984685935 not found: ID does not exist" Oct 02 02:43:01 crc kubenswrapper[4885]: I1002 02:43:01.655310 4885 scope.go:117] "RemoveContainer" containerID="76566178bfd7d0dde88ff50a9b815b2b2d8f636c8b948862a0f581334a379e71" Oct 02 02:43:01 crc kubenswrapper[4885]: E1002 02:43:01.655919 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"76566178bfd7d0dde88ff50a9b815b2b2d8f636c8b948862a0f581334a379e71\": container with ID starting with 76566178bfd7d0dde88ff50a9b815b2b2d8f636c8b948862a0f581334a379e71 not found: ID does not exist" containerID="76566178bfd7d0dde88ff50a9b815b2b2d8f636c8b948862a0f581334a379e71" Oct 02 02:43:01 crc kubenswrapper[4885]: I1002 02:43:01.655979 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"76566178bfd7d0dde88ff50a9b815b2b2d8f636c8b948862a0f581334a379e71"} err="failed to get container status \"76566178bfd7d0dde88ff50a9b815b2b2d8f636c8b948862a0f581334a379e71\": rpc error: code = NotFound desc = could not find container \"76566178bfd7d0dde88ff50a9b815b2b2d8f636c8b948862a0f581334a379e71\": container with ID starting with 76566178bfd7d0dde88ff50a9b815b2b2d8f636c8b948862a0f581334a379e71 not found: ID does not exist" Oct 02 02:43:01 crc kubenswrapper[4885]: I1002 02:43:01.656028 4885 scope.go:117] "RemoveContainer" containerID="2893ad13a5715ab344d73ccd6d4ac6b2471edace707c62caac1e3b553e5b8e79" Oct 02 02:43:01 crc kubenswrapper[4885]: E1002 02:43:01.656575 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2893ad13a5715ab344d73ccd6d4ac6b2471edace707c62caac1e3b553e5b8e79\": container with ID starting with 2893ad13a5715ab344d73ccd6d4ac6b2471edace707c62caac1e3b553e5b8e79 not found: ID does not exist" containerID="2893ad13a5715ab344d73ccd6d4ac6b2471edace707c62caac1e3b553e5b8e79" Oct 02 02:43:01 crc kubenswrapper[4885]: I1002 02:43:01.656606 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2893ad13a5715ab344d73ccd6d4ac6b2471edace707c62caac1e3b553e5b8e79"} err="failed to get container status \"2893ad13a5715ab344d73ccd6d4ac6b2471edace707c62caac1e3b553e5b8e79\": rpc error: code = NotFound desc = could not find container \"2893ad13a5715ab344d73ccd6d4ac6b2471edace707c62caac1e3b553e5b8e79\": container with ID starting with 2893ad13a5715ab344d73ccd6d4ac6b2471edace707c62caac1e3b553e5b8e79 not found: ID does not exist" Oct 02 02:43:02 crc kubenswrapper[4885]: I1002 02:43:02.058238 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ad269d3-74cd-4a34-8ef6-e3d25cbaca5d" path="/var/lib/kubelet/pods/3ad269d3-74cd-4a34-8ef6-e3d25cbaca5d/volumes" Oct 02 02:43:13 crc kubenswrapper[4885]: I1002 02:43:13.266422 4885 patch_prober.go:28] interesting pod/machine-config-daemon-rttx8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 02:43:13 crc kubenswrapper[4885]: I1002 02:43:13.267334 4885 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 02:43:14 crc kubenswrapper[4885]: I1002 02:43:14.285463 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-69qp4"] Oct 02 02:43:14 crc kubenswrapper[4885]: E1002 02:43:14.286029 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ad269d3-74cd-4a34-8ef6-e3d25cbaca5d" containerName="registry-server" Oct 02 02:43:14 crc kubenswrapper[4885]: I1002 02:43:14.286041 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ad269d3-74cd-4a34-8ef6-e3d25cbaca5d" containerName="registry-server" Oct 02 02:43:14 crc kubenswrapper[4885]: E1002 02:43:14.286064 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ad269d3-74cd-4a34-8ef6-e3d25cbaca5d" containerName="extract-content" Oct 02 02:43:14 crc kubenswrapper[4885]: I1002 02:43:14.286070 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ad269d3-74cd-4a34-8ef6-e3d25cbaca5d" containerName="extract-content" Oct 02 02:43:14 crc kubenswrapper[4885]: E1002 02:43:14.286090 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ad269d3-74cd-4a34-8ef6-e3d25cbaca5d" containerName="extract-utilities" Oct 02 02:43:14 crc kubenswrapper[4885]: I1002 02:43:14.286097 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ad269d3-74cd-4a34-8ef6-e3d25cbaca5d" containerName="extract-utilities" Oct 02 02:43:14 crc kubenswrapper[4885]: I1002 02:43:14.286325 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ad269d3-74cd-4a34-8ef6-e3d25cbaca5d" containerName="registry-server" Oct 02 02:43:14 crc kubenswrapper[4885]: I1002 02:43:14.288656 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-69qp4" Oct 02 02:43:14 crc kubenswrapper[4885]: I1002 02:43:14.309462 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-69qp4"] Oct 02 02:43:14 crc kubenswrapper[4885]: I1002 02:43:14.469242 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5bd95bf8-235b-40e1-bacf-5cc861daf8b6-catalog-content\") pod \"certified-operators-69qp4\" (UID: \"5bd95bf8-235b-40e1-bacf-5cc861daf8b6\") " pod="openshift-marketplace/certified-operators-69qp4" Oct 02 02:43:14 crc kubenswrapper[4885]: I1002 02:43:14.469633 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pdhbk\" (UniqueName: \"kubernetes.io/projected/5bd95bf8-235b-40e1-bacf-5cc861daf8b6-kube-api-access-pdhbk\") pod \"certified-operators-69qp4\" (UID: \"5bd95bf8-235b-40e1-bacf-5cc861daf8b6\") " pod="openshift-marketplace/certified-operators-69qp4" Oct 02 02:43:14 crc kubenswrapper[4885]: I1002 02:43:14.469847 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5bd95bf8-235b-40e1-bacf-5cc861daf8b6-utilities\") pod \"certified-operators-69qp4\" (UID: \"5bd95bf8-235b-40e1-bacf-5cc861daf8b6\") " pod="openshift-marketplace/certified-operators-69qp4" Oct 02 02:43:14 crc kubenswrapper[4885]: I1002 02:43:14.571832 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5bd95bf8-235b-40e1-bacf-5cc861daf8b6-utilities\") pod \"certified-operators-69qp4\" (UID: \"5bd95bf8-235b-40e1-bacf-5cc861daf8b6\") " pod="openshift-marketplace/certified-operators-69qp4" Oct 02 02:43:14 crc kubenswrapper[4885]: I1002 02:43:14.571994 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5bd95bf8-235b-40e1-bacf-5cc861daf8b6-catalog-content\") pod \"certified-operators-69qp4\" (UID: \"5bd95bf8-235b-40e1-bacf-5cc861daf8b6\") " pod="openshift-marketplace/certified-operators-69qp4" Oct 02 02:43:14 crc kubenswrapper[4885]: I1002 02:43:14.572152 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pdhbk\" (UniqueName: \"kubernetes.io/projected/5bd95bf8-235b-40e1-bacf-5cc861daf8b6-kube-api-access-pdhbk\") pod \"certified-operators-69qp4\" (UID: \"5bd95bf8-235b-40e1-bacf-5cc861daf8b6\") " pod="openshift-marketplace/certified-operators-69qp4" Oct 02 02:43:14 crc kubenswrapper[4885]: I1002 02:43:14.573722 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5bd95bf8-235b-40e1-bacf-5cc861daf8b6-utilities\") pod \"certified-operators-69qp4\" (UID: \"5bd95bf8-235b-40e1-bacf-5cc861daf8b6\") " pod="openshift-marketplace/certified-operators-69qp4" Oct 02 02:43:14 crc kubenswrapper[4885]: I1002 02:43:14.574084 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5bd95bf8-235b-40e1-bacf-5cc861daf8b6-catalog-content\") pod \"certified-operators-69qp4\" (UID: \"5bd95bf8-235b-40e1-bacf-5cc861daf8b6\") " pod="openshift-marketplace/certified-operators-69qp4" Oct 02 02:43:14 crc kubenswrapper[4885]: I1002 02:43:14.621991 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pdhbk\" (UniqueName: \"kubernetes.io/projected/5bd95bf8-235b-40e1-bacf-5cc861daf8b6-kube-api-access-pdhbk\") pod \"certified-operators-69qp4\" (UID: \"5bd95bf8-235b-40e1-bacf-5cc861daf8b6\") " pod="openshift-marketplace/certified-operators-69qp4" Oct 02 02:43:14 crc kubenswrapper[4885]: I1002 02:43:14.626515 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-69qp4" Oct 02 02:43:15 crc kubenswrapper[4885]: I1002 02:43:15.197122 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-69qp4"] Oct 02 02:43:15 crc kubenswrapper[4885]: I1002 02:43:15.293528 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-nk7xq"] Oct 02 02:43:15 crc kubenswrapper[4885]: I1002 02:43:15.297193 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nk7xq" Oct 02 02:43:15 crc kubenswrapper[4885]: I1002 02:43:15.326552 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nk7xq"] Oct 02 02:43:15 crc kubenswrapper[4885]: I1002 02:43:15.388322 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ddb310b-5732-4b17-8662-5de244941282-catalog-content\") pod \"redhat-marketplace-nk7xq\" (UID: \"7ddb310b-5732-4b17-8662-5de244941282\") " pod="openshift-marketplace/redhat-marketplace-nk7xq" Oct 02 02:43:15 crc kubenswrapper[4885]: I1002 02:43:15.388609 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ddb310b-5732-4b17-8662-5de244941282-utilities\") pod \"redhat-marketplace-nk7xq\" (UID: \"7ddb310b-5732-4b17-8662-5de244941282\") " pod="openshift-marketplace/redhat-marketplace-nk7xq" Oct 02 02:43:15 crc kubenswrapper[4885]: I1002 02:43:15.388665 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wd6j4\" (UniqueName: \"kubernetes.io/projected/7ddb310b-5732-4b17-8662-5de244941282-kube-api-access-wd6j4\") pod \"redhat-marketplace-nk7xq\" (UID: \"7ddb310b-5732-4b17-8662-5de244941282\") " pod="openshift-marketplace/redhat-marketplace-nk7xq" Oct 02 02:43:15 crc kubenswrapper[4885]: I1002 02:43:15.489742 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ddb310b-5732-4b17-8662-5de244941282-utilities\") pod \"redhat-marketplace-nk7xq\" (UID: \"7ddb310b-5732-4b17-8662-5de244941282\") " pod="openshift-marketplace/redhat-marketplace-nk7xq" Oct 02 02:43:15 crc kubenswrapper[4885]: I1002 02:43:15.489823 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wd6j4\" (UniqueName: \"kubernetes.io/projected/7ddb310b-5732-4b17-8662-5de244941282-kube-api-access-wd6j4\") pod \"redhat-marketplace-nk7xq\" (UID: \"7ddb310b-5732-4b17-8662-5de244941282\") " pod="openshift-marketplace/redhat-marketplace-nk7xq" Oct 02 02:43:15 crc kubenswrapper[4885]: I1002 02:43:15.489889 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ddb310b-5732-4b17-8662-5de244941282-catalog-content\") pod \"redhat-marketplace-nk7xq\" (UID: \"7ddb310b-5732-4b17-8662-5de244941282\") " pod="openshift-marketplace/redhat-marketplace-nk7xq" Oct 02 02:43:15 crc kubenswrapper[4885]: I1002 02:43:15.490224 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ddb310b-5732-4b17-8662-5de244941282-utilities\") pod \"redhat-marketplace-nk7xq\" (UID: \"7ddb310b-5732-4b17-8662-5de244941282\") " pod="openshift-marketplace/redhat-marketplace-nk7xq" Oct 02 02:43:15 crc kubenswrapper[4885]: I1002 02:43:15.490607 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ddb310b-5732-4b17-8662-5de244941282-catalog-content\") pod \"redhat-marketplace-nk7xq\" (UID: \"7ddb310b-5732-4b17-8662-5de244941282\") " pod="openshift-marketplace/redhat-marketplace-nk7xq" Oct 02 02:43:15 crc kubenswrapper[4885]: I1002 02:43:15.512450 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wd6j4\" (UniqueName: \"kubernetes.io/projected/7ddb310b-5732-4b17-8662-5de244941282-kube-api-access-wd6j4\") pod \"redhat-marketplace-nk7xq\" (UID: \"7ddb310b-5732-4b17-8662-5de244941282\") " pod="openshift-marketplace/redhat-marketplace-nk7xq" Oct 02 02:43:15 crc kubenswrapper[4885]: I1002 02:43:15.648051 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nk7xq" Oct 02 02:43:15 crc kubenswrapper[4885]: I1002 02:43:15.707376 4885 generic.go:334] "Generic (PLEG): container finished" podID="5bd95bf8-235b-40e1-bacf-5cc861daf8b6" containerID="1f087cc98c63ea2b49f72c08586522bdd91cf91fdf01016af627fb0df0eec6bc" exitCode=0 Oct 02 02:43:15 crc kubenswrapper[4885]: I1002 02:43:15.707433 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-69qp4" event={"ID":"5bd95bf8-235b-40e1-bacf-5cc861daf8b6","Type":"ContainerDied","Data":"1f087cc98c63ea2b49f72c08586522bdd91cf91fdf01016af627fb0df0eec6bc"} Oct 02 02:43:15 crc kubenswrapper[4885]: I1002 02:43:15.707455 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-69qp4" event={"ID":"5bd95bf8-235b-40e1-bacf-5cc861daf8b6","Type":"ContainerStarted","Data":"b300de76c0bb8b9c9033845fd7e448895203af3c028de1a3048115743a6e7ae8"} Oct 02 02:43:16 crc kubenswrapper[4885]: I1002 02:43:16.163007 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nk7xq"] Oct 02 02:43:16 crc kubenswrapper[4885]: I1002 02:43:16.718191 4885 generic.go:334] "Generic (PLEG): container finished" podID="7ddb310b-5732-4b17-8662-5de244941282" containerID="a5dfde57ab7abeda15c39e65e0fcabc025b1b03275106f57b7828e72e7c49eab" exitCode=0 Oct 02 02:43:16 crc kubenswrapper[4885]: I1002 02:43:16.718305 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nk7xq" event={"ID":"7ddb310b-5732-4b17-8662-5de244941282","Type":"ContainerDied","Data":"a5dfde57ab7abeda15c39e65e0fcabc025b1b03275106f57b7828e72e7c49eab"} Oct 02 02:43:16 crc kubenswrapper[4885]: I1002 02:43:16.718557 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nk7xq" event={"ID":"7ddb310b-5732-4b17-8662-5de244941282","Type":"ContainerStarted","Data":"819e97a490ab7f2a75fc265d40f80e193e75d673f73ffd4efce1b2697b764b0a"} Oct 02 02:43:16 crc kubenswrapper[4885]: I1002 02:43:16.723393 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-69qp4" event={"ID":"5bd95bf8-235b-40e1-bacf-5cc861daf8b6","Type":"ContainerStarted","Data":"7313055aaa3900d8bc213c2b74c72049065806b7aa9fc36978c35b67f9bee262"} Oct 02 02:43:17 crc kubenswrapper[4885]: I1002 02:43:17.734420 4885 generic.go:334] "Generic (PLEG): container finished" podID="5bd95bf8-235b-40e1-bacf-5cc861daf8b6" containerID="7313055aaa3900d8bc213c2b74c72049065806b7aa9fc36978c35b67f9bee262" exitCode=0 Oct 02 02:43:17 crc kubenswrapper[4885]: I1002 02:43:17.734482 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-69qp4" event={"ID":"5bd95bf8-235b-40e1-bacf-5cc861daf8b6","Type":"ContainerDied","Data":"7313055aaa3900d8bc213c2b74c72049065806b7aa9fc36978c35b67f9bee262"} Oct 02 02:43:18 crc kubenswrapper[4885]: I1002 02:43:18.746123 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-69qp4" event={"ID":"5bd95bf8-235b-40e1-bacf-5cc861daf8b6","Type":"ContainerStarted","Data":"f623851a2b59f604016c819952a62c4ad165de4a8d9f02f2c50c315dd61fea07"} Oct 02 02:43:18 crc kubenswrapper[4885]: I1002 02:43:18.748645 4885 generic.go:334] "Generic (PLEG): container finished" podID="7ddb310b-5732-4b17-8662-5de244941282" containerID="519f7bed7ad2d047f16b248b7d21007b6b0b6076af131bd364caf04c43a9eaad" exitCode=0 Oct 02 02:43:18 crc kubenswrapper[4885]: I1002 02:43:18.748686 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nk7xq" event={"ID":"7ddb310b-5732-4b17-8662-5de244941282","Type":"ContainerDied","Data":"519f7bed7ad2d047f16b248b7d21007b6b0b6076af131bd364caf04c43a9eaad"} Oct 02 02:43:18 crc kubenswrapper[4885]: I1002 02:43:18.772611 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-69qp4" podStartSLOduration=2.218386093 podStartE2EDuration="4.772590361s" podCreationTimestamp="2025-10-02 02:43:14 +0000 UTC" firstStartedPulling="2025-10-02 02:43:15.709853182 +0000 UTC m=+3384.521600581" lastFinishedPulling="2025-10-02 02:43:18.26405741 +0000 UTC m=+3387.075804849" observedRunningTime="2025-10-02 02:43:18.768446764 +0000 UTC m=+3387.580194163" watchObservedRunningTime="2025-10-02 02:43:18.772590361 +0000 UTC m=+3387.584337760" Oct 02 02:43:19 crc kubenswrapper[4885]: I1002 02:43:19.761877 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nk7xq" event={"ID":"7ddb310b-5732-4b17-8662-5de244941282","Type":"ContainerStarted","Data":"f8c7fec8313fd7db988d1e39fb3265522f151b04e78fefc903cdb79b6b667386"} Oct 02 02:43:19 crc kubenswrapper[4885]: I1002 02:43:19.789493 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-nk7xq" podStartSLOduration=2.22972691 podStartE2EDuration="4.789443126s" podCreationTimestamp="2025-10-02 02:43:15 +0000 UTC" firstStartedPulling="2025-10-02 02:43:16.721470248 +0000 UTC m=+3385.533217657" lastFinishedPulling="2025-10-02 02:43:19.281186474 +0000 UTC m=+3388.092933873" observedRunningTime="2025-10-02 02:43:19.780911817 +0000 UTC m=+3388.592659216" watchObservedRunningTime="2025-10-02 02:43:19.789443126 +0000 UTC m=+3388.601190565" Oct 02 02:43:24 crc kubenswrapper[4885]: I1002 02:43:24.626682 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-69qp4" Oct 02 02:43:24 crc kubenswrapper[4885]: I1002 02:43:24.627921 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-69qp4" Oct 02 02:43:24 crc kubenswrapper[4885]: I1002 02:43:24.678986 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-69qp4" Oct 02 02:43:24 crc kubenswrapper[4885]: I1002 02:43:24.889078 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-69qp4" Oct 02 02:43:25 crc kubenswrapper[4885]: I1002 02:43:25.649181 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-nk7xq" Oct 02 02:43:25 crc kubenswrapper[4885]: I1002 02:43:25.649485 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-nk7xq" Oct 02 02:43:25 crc kubenswrapper[4885]: I1002 02:43:25.703516 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-nk7xq" Oct 02 02:43:25 crc kubenswrapper[4885]: I1002 02:43:25.916010 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-nk7xq" Oct 02 02:43:29 crc kubenswrapper[4885]: I1002 02:43:29.018847 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-69qp4"] Oct 02 02:43:29 crc kubenswrapper[4885]: I1002 02:43:29.019445 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-69qp4" podUID="5bd95bf8-235b-40e1-bacf-5cc861daf8b6" containerName="registry-server" containerID="cri-o://f623851a2b59f604016c819952a62c4ad165de4a8d9f02f2c50c315dd61fea07" gracePeriod=2 Oct 02 02:43:29 crc kubenswrapper[4885]: I1002 02:43:29.215468 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-nk7xq"] Oct 02 02:43:29 crc kubenswrapper[4885]: I1002 02:43:29.216773 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-nk7xq" podUID="7ddb310b-5732-4b17-8662-5de244941282" containerName="registry-server" containerID="cri-o://f8c7fec8313fd7db988d1e39fb3265522f151b04e78fefc903cdb79b6b667386" gracePeriod=2 Oct 02 02:43:29 crc kubenswrapper[4885]: I1002 02:43:29.556534 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-69qp4" Oct 02 02:43:29 crc kubenswrapper[4885]: I1002 02:43:29.688625 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nk7xq" Oct 02 02:43:29 crc kubenswrapper[4885]: I1002 02:43:29.725585 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5bd95bf8-235b-40e1-bacf-5cc861daf8b6-utilities\") pod \"5bd95bf8-235b-40e1-bacf-5cc861daf8b6\" (UID: \"5bd95bf8-235b-40e1-bacf-5cc861daf8b6\") " Oct 02 02:43:29 crc kubenswrapper[4885]: I1002 02:43:29.725851 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5bd95bf8-235b-40e1-bacf-5cc861daf8b6-catalog-content\") pod \"5bd95bf8-235b-40e1-bacf-5cc861daf8b6\" (UID: \"5bd95bf8-235b-40e1-bacf-5cc861daf8b6\") " Oct 02 02:43:29 crc kubenswrapper[4885]: I1002 02:43:29.725892 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pdhbk\" (UniqueName: \"kubernetes.io/projected/5bd95bf8-235b-40e1-bacf-5cc861daf8b6-kube-api-access-pdhbk\") pod \"5bd95bf8-235b-40e1-bacf-5cc861daf8b6\" (UID: \"5bd95bf8-235b-40e1-bacf-5cc861daf8b6\") " Oct 02 02:43:29 crc kubenswrapper[4885]: I1002 02:43:29.727815 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5bd95bf8-235b-40e1-bacf-5cc861daf8b6-utilities" (OuterVolumeSpecName: "utilities") pod "5bd95bf8-235b-40e1-bacf-5cc861daf8b6" (UID: "5bd95bf8-235b-40e1-bacf-5cc861daf8b6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:43:29 crc kubenswrapper[4885]: I1002 02:43:29.732685 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5bd95bf8-235b-40e1-bacf-5cc861daf8b6-kube-api-access-pdhbk" (OuterVolumeSpecName: "kube-api-access-pdhbk") pod "5bd95bf8-235b-40e1-bacf-5cc861daf8b6" (UID: "5bd95bf8-235b-40e1-bacf-5cc861daf8b6"). InnerVolumeSpecName "kube-api-access-pdhbk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:43:29 crc kubenswrapper[4885]: I1002 02:43:29.780743 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5bd95bf8-235b-40e1-bacf-5cc861daf8b6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5bd95bf8-235b-40e1-bacf-5cc861daf8b6" (UID: "5bd95bf8-235b-40e1-bacf-5cc861daf8b6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:43:29 crc kubenswrapper[4885]: I1002 02:43:29.828074 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ddb310b-5732-4b17-8662-5de244941282-utilities\") pod \"7ddb310b-5732-4b17-8662-5de244941282\" (UID: \"7ddb310b-5732-4b17-8662-5de244941282\") " Oct 02 02:43:29 crc kubenswrapper[4885]: I1002 02:43:29.828412 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wd6j4\" (UniqueName: \"kubernetes.io/projected/7ddb310b-5732-4b17-8662-5de244941282-kube-api-access-wd6j4\") pod \"7ddb310b-5732-4b17-8662-5de244941282\" (UID: \"7ddb310b-5732-4b17-8662-5de244941282\") " Oct 02 02:43:29 crc kubenswrapper[4885]: I1002 02:43:29.828619 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ddb310b-5732-4b17-8662-5de244941282-catalog-content\") pod \"7ddb310b-5732-4b17-8662-5de244941282\" (UID: \"7ddb310b-5732-4b17-8662-5de244941282\") " Oct 02 02:43:29 crc kubenswrapper[4885]: I1002 02:43:29.829649 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7ddb310b-5732-4b17-8662-5de244941282-utilities" (OuterVolumeSpecName: "utilities") pod "7ddb310b-5732-4b17-8662-5de244941282" (UID: "7ddb310b-5732-4b17-8662-5de244941282"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:43:29 crc kubenswrapper[4885]: I1002 02:43:29.835577 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ddb310b-5732-4b17-8662-5de244941282-kube-api-access-wd6j4" (OuterVolumeSpecName: "kube-api-access-wd6j4") pod "7ddb310b-5732-4b17-8662-5de244941282" (UID: "7ddb310b-5732-4b17-8662-5de244941282"). InnerVolumeSpecName "kube-api-access-wd6j4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:43:29 crc kubenswrapper[4885]: I1002 02:43:29.839744 4885 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5bd95bf8-235b-40e1-bacf-5cc861daf8b6-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 02:43:29 crc kubenswrapper[4885]: I1002 02:43:29.839889 4885 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ddb310b-5732-4b17-8662-5de244941282-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 02:43:29 crc kubenswrapper[4885]: I1002 02:43:29.839967 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wd6j4\" (UniqueName: \"kubernetes.io/projected/7ddb310b-5732-4b17-8662-5de244941282-kube-api-access-wd6j4\") on node \"crc\" DevicePath \"\"" Oct 02 02:43:29 crc kubenswrapper[4885]: I1002 02:43:29.840070 4885 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5bd95bf8-235b-40e1-bacf-5cc861daf8b6-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 02:43:29 crc kubenswrapper[4885]: I1002 02:43:29.840163 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pdhbk\" (UniqueName: \"kubernetes.io/projected/5bd95bf8-235b-40e1-bacf-5cc861daf8b6-kube-api-access-pdhbk\") on node \"crc\" DevicePath \"\"" Oct 02 02:43:29 crc kubenswrapper[4885]: I1002 02:43:29.844553 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7ddb310b-5732-4b17-8662-5de244941282-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7ddb310b-5732-4b17-8662-5de244941282" (UID: "7ddb310b-5732-4b17-8662-5de244941282"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:43:29 crc kubenswrapper[4885]: I1002 02:43:29.894625 4885 generic.go:334] "Generic (PLEG): container finished" podID="5bd95bf8-235b-40e1-bacf-5cc861daf8b6" containerID="f623851a2b59f604016c819952a62c4ad165de4a8d9f02f2c50c315dd61fea07" exitCode=0 Oct 02 02:43:29 crc kubenswrapper[4885]: I1002 02:43:29.894781 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-69qp4" event={"ID":"5bd95bf8-235b-40e1-bacf-5cc861daf8b6","Type":"ContainerDied","Data":"f623851a2b59f604016c819952a62c4ad165de4a8d9f02f2c50c315dd61fea07"} Oct 02 02:43:29 crc kubenswrapper[4885]: I1002 02:43:29.894902 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-69qp4" Oct 02 02:43:29 crc kubenswrapper[4885]: I1002 02:43:29.895070 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-69qp4" event={"ID":"5bd95bf8-235b-40e1-bacf-5cc861daf8b6","Type":"ContainerDied","Data":"b300de76c0bb8b9c9033845fd7e448895203af3c028de1a3048115743a6e7ae8"} Oct 02 02:43:29 crc kubenswrapper[4885]: I1002 02:43:29.895131 4885 scope.go:117] "RemoveContainer" containerID="f623851a2b59f604016c819952a62c4ad165de4a8d9f02f2c50c315dd61fea07" Oct 02 02:43:29 crc kubenswrapper[4885]: I1002 02:43:29.900525 4885 generic.go:334] "Generic (PLEG): container finished" podID="7ddb310b-5732-4b17-8662-5de244941282" containerID="f8c7fec8313fd7db988d1e39fb3265522f151b04e78fefc903cdb79b6b667386" exitCode=0 Oct 02 02:43:29 crc kubenswrapper[4885]: I1002 02:43:29.900565 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nk7xq" event={"ID":"7ddb310b-5732-4b17-8662-5de244941282","Type":"ContainerDied","Data":"f8c7fec8313fd7db988d1e39fb3265522f151b04e78fefc903cdb79b6b667386"} Oct 02 02:43:29 crc kubenswrapper[4885]: I1002 02:43:29.900625 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nk7xq" event={"ID":"7ddb310b-5732-4b17-8662-5de244941282","Type":"ContainerDied","Data":"819e97a490ab7f2a75fc265d40f80e193e75d673f73ffd4efce1b2697b764b0a"} Oct 02 02:43:29 crc kubenswrapper[4885]: I1002 02:43:29.900685 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nk7xq" Oct 02 02:43:29 crc kubenswrapper[4885]: I1002 02:43:29.927876 4885 scope.go:117] "RemoveContainer" containerID="7313055aaa3900d8bc213c2b74c72049065806b7aa9fc36978c35b67f9bee262" Oct 02 02:43:29 crc kubenswrapper[4885]: I1002 02:43:29.942021 4885 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ddb310b-5732-4b17-8662-5de244941282-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 02:43:29 crc kubenswrapper[4885]: I1002 02:43:29.945024 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-69qp4"] Oct 02 02:43:29 crc kubenswrapper[4885]: I1002 02:43:29.954110 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-69qp4"] Oct 02 02:43:29 crc kubenswrapper[4885]: I1002 02:43:29.963892 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-nk7xq"] Oct 02 02:43:29 crc kubenswrapper[4885]: I1002 02:43:29.967404 4885 scope.go:117] "RemoveContainer" containerID="1f087cc98c63ea2b49f72c08586522bdd91cf91fdf01016af627fb0df0eec6bc" Oct 02 02:43:29 crc kubenswrapper[4885]: I1002 02:43:29.975281 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-nk7xq"] Oct 02 02:43:29 crc kubenswrapper[4885]: I1002 02:43:29.990291 4885 scope.go:117] "RemoveContainer" containerID="f623851a2b59f604016c819952a62c4ad165de4a8d9f02f2c50c315dd61fea07" Oct 02 02:43:29 crc kubenswrapper[4885]: E1002 02:43:29.990922 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f623851a2b59f604016c819952a62c4ad165de4a8d9f02f2c50c315dd61fea07\": container with ID starting with f623851a2b59f604016c819952a62c4ad165de4a8d9f02f2c50c315dd61fea07 not found: ID does not exist" containerID="f623851a2b59f604016c819952a62c4ad165de4a8d9f02f2c50c315dd61fea07" Oct 02 02:43:29 crc kubenswrapper[4885]: I1002 02:43:29.990969 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f623851a2b59f604016c819952a62c4ad165de4a8d9f02f2c50c315dd61fea07"} err="failed to get container status \"f623851a2b59f604016c819952a62c4ad165de4a8d9f02f2c50c315dd61fea07\": rpc error: code = NotFound desc = could not find container \"f623851a2b59f604016c819952a62c4ad165de4a8d9f02f2c50c315dd61fea07\": container with ID starting with f623851a2b59f604016c819952a62c4ad165de4a8d9f02f2c50c315dd61fea07 not found: ID does not exist" Oct 02 02:43:29 crc kubenswrapper[4885]: I1002 02:43:29.990998 4885 scope.go:117] "RemoveContainer" containerID="7313055aaa3900d8bc213c2b74c72049065806b7aa9fc36978c35b67f9bee262" Oct 02 02:43:29 crc kubenswrapper[4885]: E1002 02:43:29.991435 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7313055aaa3900d8bc213c2b74c72049065806b7aa9fc36978c35b67f9bee262\": container with ID starting with 7313055aaa3900d8bc213c2b74c72049065806b7aa9fc36978c35b67f9bee262 not found: ID does not exist" containerID="7313055aaa3900d8bc213c2b74c72049065806b7aa9fc36978c35b67f9bee262" Oct 02 02:43:29 crc kubenswrapper[4885]: I1002 02:43:29.991483 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7313055aaa3900d8bc213c2b74c72049065806b7aa9fc36978c35b67f9bee262"} err="failed to get container status \"7313055aaa3900d8bc213c2b74c72049065806b7aa9fc36978c35b67f9bee262\": rpc error: code = NotFound desc = could not find container \"7313055aaa3900d8bc213c2b74c72049065806b7aa9fc36978c35b67f9bee262\": container with ID starting with 7313055aaa3900d8bc213c2b74c72049065806b7aa9fc36978c35b67f9bee262 not found: ID does not exist" Oct 02 02:43:29 crc kubenswrapper[4885]: I1002 02:43:29.991517 4885 scope.go:117] "RemoveContainer" containerID="1f087cc98c63ea2b49f72c08586522bdd91cf91fdf01016af627fb0df0eec6bc" Oct 02 02:43:29 crc kubenswrapper[4885]: E1002 02:43:29.992129 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1f087cc98c63ea2b49f72c08586522bdd91cf91fdf01016af627fb0df0eec6bc\": container with ID starting with 1f087cc98c63ea2b49f72c08586522bdd91cf91fdf01016af627fb0df0eec6bc not found: ID does not exist" containerID="1f087cc98c63ea2b49f72c08586522bdd91cf91fdf01016af627fb0df0eec6bc" Oct 02 02:43:29 crc kubenswrapper[4885]: I1002 02:43:29.992161 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f087cc98c63ea2b49f72c08586522bdd91cf91fdf01016af627fb0df0eec6bc"} err="failed to get container status \"1f087cc98c63ea2b49f72c08586522bdd91cf91fdf01016af627fb0df0eec6bc\": rpc error: code = NotFound desc = could not find container \"1f087cc98c63ea2b49f72c08586522bdd91cf91fdf01016af627fb0df0eec6bc\": container with ID starting with 1f087cc98c63ea2b49f72c08586522bdd91cf91fdf01016af627fb0df0eec6bc not found: ID does not exist" Oct 02 02:43:29 crc kubenswrapper[4885]: I1002 02:43:29.992179 4885 scope.go:117] "RemoveContainer" containerID="f8c7fec8313fd7db988d1e39fb3265522f151b04e78fefc903cdb79b6b667386" Oct 02 02:43:30 crc kubenswrapper[4885]: I1002 02:43:30.016686 4885 scope.go:117] "RemoveContainer" containerID="519f7bed7ad2d047f16b248b7d21007b6b0b6076af131bd364caf04c43a9eaad" Oct 02 02:43:30 crc kubenswrapper[4885]: I1002 02:43:30.043540 4885 scope.go:117] "RemoveContainer" containerID="a5dfde57ab7abeda15c39e65e0fcabc025b1b03275106f57b7828e72e7c49eab" Oct 02 02:43:30 crc kubenswrapper[4885]: I1002 02:43:30.058153 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5bd95bf8-235b-40e1-bacf-5cc861daf8b6" path="/var/lib/kubelet/pods/5bd95bf8-235b-40e1-bacf-5cc861daf8b6/volumes" Oct 02 02:43:30 crc kubenswrapper[4885]: I1002 02:43:30.058993 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ddb310b-5732-4b17-8662-5de244941282" path="/var/lib/kubelet/pods/7ddb310b-5732-4b17-8662-5de244941282/volumes" Oct 02 02:43:30 crc kubenswrapper[4885]: I1002 02:43:30.068248 4885 scope.go:117] "RemoveContainer" containerID="f8c7fec8313fd7db988d1e39fb3265522f151b04e78fefc903cdb79b6b667386" Oct 02 02:43:30 crc kubenswrapper[4885]: E1002 02:43:30.068776 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f8c7fec8313fd7db988d1e39fb3265522f151b04e78fefc903cdb79b6b667386\": container with ID starting with f8c7fec8313fd7db988d1e39fb3265522f151b04e78fefc903cdb79b6b667386 not found: ID does not exist" containerID="f8c7fec8313fd7db988d1e39fb3265522f151b04e78fefc903cdb79b6b667386" Oct 02 02:43:30 crc kubenswrapper[4885]: I1002 02:43:30.068842 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f8c7fec8313fd7db988d1e39fb3265522f151b04e78fefc903cdb79b6b667386"} err="failed to get container status \"f8c7fec8313fd7db988d1e39fb3265522f151b04e78fefc903cdb79b6b667386\": rpc error: code = NotFound desc = could not find container \"f8c7fec8313fd7db988d1e39fb3265522f151b04e78fefc903cdb79b6b667386\": container with ID starting with f8c7fec8313fd7db988d1e39fb3265522f151b04e78fefc903cdb79b6b667386 not found: ID does not exist" Oct 02 02:43:30 crc kubenswrapper[4885]: I1002 02:43:30.068883 4885 scope.go:117] "RemoveContainer" containerID="519f7bed7ad2d047f16b248b7d21007b6b0b6076af131bd364caf04c43a9eaad" Oct 02 02:43:30 crc kubenswrapper[4885]: E1002 02:43:30.069450 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"519f7bed7ad2d047f16b248b7d21007b6b0b6076af131bd364caf04c43a9eaad\": container with ID starting with 519f7bed7ad2d047f16b248b7d21007b6b0b6076af131bd364caf04c43a9eaad not found: ID does not exist" containerID="519f7bed7ad2d047f16b248b7d21007b6b0b6076af131bd364caf04c43a9eaad" Oct 02 02:43:30 crc kubenswrapper[4885]: I1002 02:43:30.069519 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"519f7bed7ad2d047f16b248b7d21007b6b0b6076af131bd364caf04c43a9eaad"} err="failed to get container status \"519f7bed7ad2d047f16b248b7d21007b6b0b6076af131bd364caf04c43a9eaad\": rpc error: code = NotFound desc = could not find container \"519f7bed7ad2d047f16b248b7d21007b6b0b6076af131bd364caf04c43a9eaad\": container with ID starting with 519f7bed7ad2d047f16b248b7d21007b6b0b6076af131bd364caf04c43a9eaad not found: ID does not exist" Oct 02 02:43:30 crc kubenswrapper[4885]: I1002 02:43:30.069568 4885 scope.go:117] "RemoveContainer" containerID="a5dfde57ab7abeda15c39e65e0fcabc025b1b03275106f57b7828e72e7c49eab" Oct 02 02:43:30 crc kubenswrapper[4885]: E1002 02:43:30.069851 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a5dfde57ab7abeda15c39e65e0fcabc025b1b03275106f57b7828e72e7c49eab\": container with ID starting with a5dfde57ab7abeda15c39e65e0fcabc025b1b03275106f57b7828e72e7c49eab not found: ID does not exist" containerID="a5dfde57ab7abeda15c39e65e0fcabc025b1b03275106f57b7828e72e7c49eab" Oct 02 02:43:30 crc kubenswrapper[4885]: I1002 02:43:30.069886 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a5dfde57ab7abeda15c39e65e0fcabc025b1b03275106f57b7828e72e7c49eab"} err="failed to get container status \"a5dfde57ab7abeda15c39e65e0fcabc025b1b03275106f57b7828e72e7c49eab\": rpc error: code = NotFound desc = could not find container \"a5dfde57ab7abeda15c39e65e0fcabc025b1b03275106f57b7828e72e7c49eab\": container with ID starting with a5dfde57ab7abeda15c39e65e0fcabc025b1b03275106f57b7828e72e7c49eab not found: ID does not exist" Oct 02 02:43:43 crc kubenswrapper[4885]: I1002 02:43:43.266409 4885 patch_prober.go:28] interesting pod/machine-config-daemon-rttx8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 02:43:43 crc kubenswrapper[4885]: I1002 02:43:43.267127 4885 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 02:44:10 crc kubenswrapper[4885]: I1002 02:44:10.906423 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-l8jsc"] Oct 02 02:44:10 crc kubenswrapper[4885]: E1002 02:44:10.907692 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5bd95bf8-235b-40e1-bacf-5cc861daf8b6" containerName="extract-utilities" Oct 02 02:44:10 crc kubenswrapper[4885]: I1002 02:44:10.907709 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="5bd95bf8-235b-40e1-bacf-5cc861daf8b6" containerName="extract-utilities" Oct 02 02:44:10 crc kubenswrapper[4885]: E1002 02:44:10.907727 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5bd95bf8-235b-40e1-bacf-5cc861daf8b6" containerName="extract-content" Oct 02 02:44:10 crc kubenswrapper[4885]: I1002 02:44:10.907736 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="5bd95bf8-235b-40e1-bacf-5cc861daf8b6" containerName="extract-content" Oct 02 02:44:10 crc kubenswrapper[4885]: E1002 02:44:10.907753 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5bd95bf8-235b-40e1-bacf-5cc861daf8b6" containerName="registry-server" Oct 02 02:44:10 crc kubenswrapper[4885]: I1002 02:44:10.907762 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="5bd95bf8-235b-40e1-bacf-5cc861daf8b6" containerName="registry-server" Oct 02 02:44:10 crc kubenswrapper[4885]: E1002 02:44:10.907792 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ddb310b-5732-4b17-8662-5de244941282" containerName="extract-utilities" Oct 02 02:44:10 crc kubenswrapper[4885]: I1002 02:44:10.907801 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ddb310b-5732-4b17-8662-5de244941282" containerName="extract-utilities" Oct 02 02:44:10 crc kubenswrapper[4885]: E1002 02:44:10.907816 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ddb310b-5732-4b17-8662-5de244941282" containerName="extract-content" Oct 02 02:44:10 crc kubenswrapper[4885]: I1002 02:44:10.907823 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ddb310b-5732-4b17-8662-5de244941282" containerName="extract-content" Oct 02 02:44:10 crc kubenswrapper[4885]: E1002 02:44:10.907841 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ddb310b-5732-4b17-8662-5de244941282" containerName="registry-server" Oct 02 02:44:10 crc kubenswrapper[4885]: I1002 02:44:10.907848 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ddb310b-5732-4b17-8662-5de244941282" containerName="registry-server" Oct 02 02:44:10 crc kubenswrapper[4885]: I1002 02:44:10.908078 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ddb310b-5732-4b17-8662-5de244941282" containerName="registry-server" Oct 02 02:44:10 crc kubenswrapper[4885]: I1002 02:44:10.908120 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="5bd95bf8-235b-40e1-bacf-5cc861daf8b6" containerName="registry-server" Oct 02 02:44:10 crc kubenswrapper[4885]: I1002 02:44:10.909792 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-l8jsc" Oct 02 02:44:10 crc kubenswrapper[4885]: I1002 02:44:10.951700 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-l8jsc"] Oct 02 02:44:11 crc kubenswrapper[4885]: I1002 02:44:11.041385 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-49qll\" (UniqueName: \"kubernetes.io/projected/bb50ccc8-3964-427b-95b7-fbe6d29bd7f7-kube-api-access-49qll\") pod \"community-operators-l8jsc\" (UID: \"bb50ccc8-3964-427b-95b7-fbe6d29bd7f7\") " pod="openshift-marketplace/community-operators-l8jsc" Oct 02 02:44:11 crc kubenswrapper[4885]: I1002 02:44:11.041531 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb50ccc8-3964-427b-95b7-fbe6d29bd7f7-utilities\") pod \"community-operators-l8jsc\" (UID: \"bb50ccc8-3964-427b-95b7-fbe6d29bd7f7\") " pod="openshift-marketplace/community-operators-l8jsc" Oct 02 02:44:11 crc kubenswrapper[4885]: I1002 02:44:11.041569 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb50ccc8-3964-427b-95b7-fbe6d29bd7f7-catalog-content\") pod \"community-operators-l8jsc\" (UID: \"bb50ccc8-3964-427b-95b7-fbe6d29bd7f7\") " pod="openshift-marketplace/community-operators-l8jsc" Oct 02 02:44:11 crc kubenswrapper[4885]: I1002 02:44:11.142805 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb50ccc8-3964-427b-95b7-fbe6d29bd7f7-catalog-content\") pod \"community-operators-l8jsc\" (UID: \"bb50ccc8-3964-427b-95b7-fbe6d29bd7f7\") " pod="openshift-marketplace/community-operators-l8jsc" Oct 02 02:44:11 crc kubenswrapper[4885]: I1002 02:44:11.142883 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-49qll\" (UniqueName: \"kubernetes.io/projected/bb50ccc8-3964-427b-95b7-fbe6d29bd7f7-kube-api-access-49qll\") pod \"community-operators-l8jsc\" (UID: \"bb50ccc8-3964-427b-95b7-fbe6d29bd7f7\") " pod="openshift-marketplace/community-operators-l8jsc" Oct 02 02:44:11 crc kubenswrapper[4885]: I1002 02:44:11.143001 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb50ccc8-3964-427b-95b7-fbe6d29bd7f7-utilities\") pod \"community-operators-l8jsc\" (UID: \"bb50ccc8-3964-427b-95b7-fbe6d29bd7f7\") " pod="openshift-marketplace/community-operators-l8jsc" Oct 02 02:44:11 crc kubenswrapper[4885]: I1002 02:44:11.143529 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb50ccc8-3964-427b-95b7-fbe6d29bd7f7-utilities\") pod \"community-operators-l8jsc\" (UID: \"bb50ccc8-3964-427b-95b7-fbe6d29bd7f7\") " pod="openshift-marketplace/community-operators-l8jsc" Oct 02 02:44:11 crc kubenswrapper[4885]: I1002 02:44:11.143643 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb50ccc8-3964-427b-95b7-fbe6d29bd7f7-catalog-content\") pod \"community-operators-l8jsc\" (UID: \"bb50ccc8-3964-427b-95b7-fbe6d29bd7f7\") " pod="openshift-marketplace/community-operators-l8jsc" Oct 02 02:44:11 crc kubenswrapper[4885]: I1002 02:44:11.161964 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-49qll\" (UniqueName: \"kubernetes.io/projected/bb50ccc8-3964-427b-95b7-fbe6d29bd7f7-kube-api-access-49qll\") pod \"community-operators-l8jsc\" (UID: \"bb50ccc8-3964-427b-95b7-fbe6d29bd7f7\") " pod="openshift-marketplace/community-operators-l8jsc" Oct 02 02:44:11 crc kubenswrapper[4885]: I1002 02:44:11.277498 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-l8jsc" Oct 02 02:44:11 crc kubenswrapper[4885]: I1002 02:44:11.785286 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-l8jsc"] Oct 02 02:44:11 crc kubenswrapper[4885]: W1002 02:44:11.794438 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbb50ccc8_3964_427b_95b7_fbe6d29bd7f7.slice/crio-70483fdb65322b49d30f420cef91464aa902fce032d42479c393f1b8a8922661 WatchSource:0}: Error finding container 70483fdb65322b49d30f420cef91464aa902fce032d42479c393f1b8a8922661: Status 404 returned error can't find the container with id 70483fdb65322b49d30f420cef91464aa902fce032d42479c393f1b8a8922661 Oct 02 02:44:12 crc kubenswrapper[4885]: I1002 02:44:12.378600 4885 generic.go:334] "Generic (PLEG): container finished" podID="bb50ccc8-3964-427b-95b7-fbe6d29bd7f7" containerID="cdd1988a65ab0ecc9eb58da66a2d1edddc41abd212e92af838e7ad57bf203a41" exitCode=0 Oct 02 02:44:12 crc kubenswrapper[4885]: I1002 02:44:12.378655 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l8jsc" event={"ID":"bb50ccc8-3964-427b-95b7-fbe6d29bd7f7","Type":"ContainerDied","Data":"cdd1988a65ab0ecc9eb58da66a2d1edddc41abd212e92af838e7ad57bf203a41"} Oct 02 02:44:12 crc kubenswrapper[4885]: I1002 02:44:12.378688 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l8jsc" event={"ID":"bb50ccc8-3964-427b-95b7-fbe6d29bd7f7","Type":"ContainerStarted","Data":"70483fdb65322b49d30f420cef91464aa902fce032d42479c393f1b8a8922661"} Oct 02 02:44:13 crc kubenswrapper[4885]: I1002 02:44:13.268474 4885 patch_prober.go:28] interesting pod/machine-config-daemon-rttx8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 02:44:13 crc kubenswrapper[4885]: I1002 02:44:13.268806 4885 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 02:44:13 crc kubenswrapper[4885]: I1002 02:44:13.268865 4885 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" Oct 02 02:44:13 crc kubenswrapper[4885]: I1002 02:44:13.269850 4885 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"263b36668b739af6c8b8dd19175897e6595f755c49d0d124cba769111393cb93"} pod="openshift-machine-config-operator/machine-config-daemon-rttx8" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 02:44:13 crc kubenswrapper[4885]: I1002 02:44:13.269947 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" containerName="machine-config-daemon" containerID="cri-o://263b36668b739af6c8b8dd19175897e6595f755c49d0d124cba769111393cb93" gracePeriod=600 Oct 02 02:44:14 crc kubenswrapper[4885]: I1002 02:44:14.413428 4885 generic.go:334] "Generic (PLEG): container finished" podID="bb50ccc8-3964-427b-95b7-fbe6d29bd7f7" containerID="dd3d6f108d96fd2c241bfa82683d3497a4bfe9a0fc30323bc0ed3325db018f90" exitCode=0 Oct 02 02:44:14 crc kubenswrapper[4885]: I1002 02:44:14.413513 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l8jsc" event={"ID":"bb50ccc8-3964-427b-95b7-fbe6d29bd7f7","Type":"ContainerDied","Data":"dd3d6f108d96fd2c241bfa82683d3497a4bfe9a0fc30323bc0ed3325db018f90"} Oct 02 02:44:14 crc kubenswrapper[4885]: I1002 02:44:14.419366 4885 generic.go:334] "Generic (PLEG): container finished" podID="d169c8ba-000a-4218-8e3b-4ae53035b110" containerID="263b36668b739af6c8b8dd19175897e6595f755c49d0d124cba769111393cb93" exitCode=0 Oct 02 02:44:14 crc kubenswrapper[4885]: I1002 02:44:14.419423 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" event={"ID":"d169c8ba-000a-4218-8e3b-4ae53035b110","Type":"ContainerDied","Data":"263b36668b739af6c8b8dd19175897e6595f755c49d0d124cba769111393cb93"} Oct 02 02:44:14 crc kubenswrapper[4885]: I1002 02:44:14.419471 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" event={"ID":"d169c8ba-000a-4218-8e3b-4ae53035b110","Type":"ContainerStarted","Data":"999b3e75477511af3aa11f1e9d932c1ffb739b6f215cc19f45ac7144051ab65c"} Oct 02 02:44:14 crc kubenswrapper[4885]: I1002 02:44:14.419502 4885 scope.go:117] "RemoveContainer" containerID="02916a559cc5f6ab8f5bc76ee73c22c5f71f201c585eeb74e3c14d51fc63cccb" Oct 02 02:44:15 crc kubenswrapper[4885]: I1002 02:44:15.436054 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l8jsc" event={"ID":"bb50ccc8-3964-427b-95b7-fbe6d29bd7f7","Type":"ContainerStarted","Data":"7f8e7416946f778d050fcb1cde342a1ba48997ed1598021a9d00aeedcd1fd256"} Oct 02 02:44:15 crc kubenswrapper[4885]: I1002 02:44:15.465288 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-l8jsc" podStartSLOduration=2.770583828 podStartE2EDuration="5.465241486s" podCreationTimestamp="2025-10-02 02:44:10 +0000 UTC" firstStartedPulling="2025-10-02 02:44:12.382577036 +0000 UTC m=+3441.194324475" lastFinishedPulling="2025-10-02 02:44:15.077234724 +0000 UTC m=+3443.888982133" observedRunningTime="2025-10-02 02:44:15.456596723 +0000 UTC m=+3444.268344122" watchObservedRunningTime="2025-10-02 02:44:15.465241486 +0000 UTC m=+3444.276988915" Oct 02 02:44:21 crc kubenswrapper[4885]: I1002 02:44:21.279333 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-l8jsc" Oct 02 02:44:21 crc kubenswrapper[4885]: I1002 02:44:21.279956 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-l8jsc" Oct 02 02:44:21 crc kubenswrapper[4885]: I1002 02:44:21.355814 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-l8jsc" Oct 02 02:44:21 crc kubenswrapper[4885]: I1002 02:44:21.566786 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-l8jsc" Oct 02 02:44:21 crc kubenswrapper[4885]: I1002 02:44:21.626484 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-l8jsc"] Oct 02 02:44:23 crc kubenswrapper[4885]: I1002 02:44:23.525223 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-l8jsc" podUID="bb50ccc8-3964-427b-95b7-fbe6d29bd7f7" containerName="registry-server" containerID="cri-o://7f8e7416946f778d050fcb1cde342a1ba48997ed1598021a9d00aeedcd1fd256" gracePeriod=2 Oct 02 02:44:24 crc kubenswrapper[4885]: I1002 02:44:24.089491 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-l8jsc" Oct 02 02:44:24 crc kubenswrapper[4885]: I1002 02:44:24.252483 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-49qll\" (UniqueName: \"kubernetes.io/projected/bb50ccc8-3964-427b-95b7-fbe6d29bd7f7-kube-api-access-49qll\") pod \"bb50ccc8-3964-427b-95b7-fbe6d29bd7f7\" (UID: \"bb50ccc8-3964-427b-95b7-fbe6d29bd7f7\") " Oct 02 02:44:24 crc kubenswrapper[4885]: I1002 02:44:24.253150 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb50ccc8-3964-427b-95b7-fbe6d29bd7f7-catalog-content\") pod \"bb50ccc8-3964-427b-95b7-fbe6d29bd7f7\" (UID: \"bb50ccc8-3964-427b-95b7-fbe6d29bd7f7\") " Oct 02 02:44:24 crc kubenswrapper[4885]: I1002 02:44:24.253333 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb50ccc8-3964-427b-95b7-fbe6d29bd7f7-utilities\") pod \"bb50ccc8-3964-427b-95b7-fbe6d29bd7f7\" (UID: \"bb50ccc8-3964-427b-95b7-fbe6d29bd7f7\") " Oct 02 02:44:24 crc kubenswrapper[4885]: I1002 02:44:24.254397 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bb50ccc8-3964-427b-95b7-fbe6d29bd7f7-utilities" (OuterVolumeSpecName: "utilities") pod "bb50ccc8-3964-427b-95b7-fbe6d29bd7f7" (UID: "bb50ccc8-3964-427b-95b7-fbe6d29bd7f7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:44:24 crc kubenswrapper[4885]: I1002 02:44:24.262591 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb50ccc8-3964-427b-95b7-fbe6d29bd7f7-kube-api-access-49qll" (OuterVolumeSpecName: "kube-api-access-49qll") pod "bb50ccc8-3964-427b-95b7-fbe6d29bd7f7" (UID: "bb50ccc8-3964-427b-95b7-fbe6d29bd7f7"). InnerVolumeSpecName "kube-api-access-49qll". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:44:24 crc kubenswrapper[4885]: I1002 02:44:24.329301 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bb50ccc8-3964-427b-95b7-fbe6d29bd7f7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bb50ccc8-3964-427b-95b7-fbe6d29bd7f7" (UID: "bb50ccc8-3964-427b-95b7-fbe6d29bd7f7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:44:24 crc kubenswrapper[4885]: I1002 02:44:24.355349 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-49qll\" (UniqueName: \"kubernetes.io/projected/bb50ccc8-3964-427b-95b7-fbe6d29bd7f7-kube-api-access-49qll\") on node \"crc\" DevicePath \"\"" Oct 02 02:44:24 crc kubenswrapper[4885]: I1002 02:44:24.355643 4885 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb50ccc8-3964-427b-95b7-fbe6d29bd7f7-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 02:44:24 crc kubenswrapper[4885]: I1002 02:44:24.355727 4885 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb50ccc8-3964-427b-95b7-fbe6d29bd7f7-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 02:44:24 crc kubenswrapper[4885]: I1002 02:44:24.542910 4885 generic.go:334] "Generic (PLEG): container finished" podID="bb50ccc8-3964-427b-95b7-fbe6d29bd7f7" containerID="7f8e7416946f778d050fcb1cde342a1ba48997ed1598021a9d00aeedcd1fd256" exitCode=0 Oct 02 02:44:24 crc kubenswrapper[4885]: I1002 02:44:24.542978 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l8jsc" event={"ID":"bb50ccc8-3964-427b-95b7-fbe6d29bd7f7","Type":"ContainerDied","Data":"7f8e7416946f778d050fcb1cde342a1ba48997ed1598021a9d00aeedcd1fd256"} Oct 02 02:44:24 crc kubenswrapper[4885]: I1002 02:44:24.543009 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-l8jsc" Oct 02 02:44:24 crc kubenswrapper[4885]: I1002 02:44:24.543046 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l8jsc" event={"ID":"bb50ccc8-3964-427b-95b7-fbe6d29bd7f7","Type":"ContainerDied","Data":"70483fdb65322b49d30f420cef91464aa902fce032d42479c393f1b8a8922661"} Oct 02 02:44:24 crc kubenswrapper[4885]: I1002 02:44:24.543080 4885 scope.go:117] "RemoveContainer" containerID="7f8e7416946f778d050fcb1cde342a1ba48997ed1598021a9d00aeedcd1fd256" Oct 02 02:44:24 crc kubenswrapper[4885]: I1002 02:44:24.601961 4885 scope.go:117] "RemoveContainer" containerID="dd3d6f108d96fd2c241bfa82683d3497a4bfe9a0fc30323bc0ed3325db018f90" Oct 02 02:44:24 crc kubenswrapper[4885]: I1002 02:44:24.613350 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-l8jsc"] Oct 02 02:44:24 crc kubenswrapper[4885]: I1002 02:44:24.624846 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-l8jsc"] Oct 02 02:44:24 crc kubenswrapper[4885]: I1002 02:44:24.636843 4885 scope.go:117] "RemoveContainer" containerID="cdd1988a65ab0ecc9eb58da66a2d1edddc41abd212e92af838e7ad57bf203a41" Oct 02 02:44:24 crc kubenswrapper[4885]: I1002 02:44:24.699689 4885 scope.go:117] "RemoveContainer" containerID="7f8e7416946f778d050fcb1cde342a1ba48997ed1598021a9d00aeedcd1fd256" Oct 02 02:44:24 crc kubenswrapper[4885]: E1002 02:44:24.700291 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7f8e7416946f778d050fcb1cde342a1ba48997ed1598021a9d00aeedcd1fd256\": container with ID starting with 7f8e7416946f778d050fcb1cde342a1ba48997ed1598021a9d00aeedcd1fd256 not found: ID does not exist" containerID="7f8e7416946f778d050fcb1cde342a1ba48997ed1598021a9d00aeedcd1fd256" Oct 02 02:44:24 crc kubenswrapper[4885]: I1002 02:44:24.700330 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f8e7416946f778d050fcb1cde342a1ba48997ed1598021a9d00aeedcd1fd256"} err="failed to get container status \"7f8e7416946f778d050fcb1cde342a1ba48997ed1598021a9d00aeedcd1fd256\": rpc error: code = NotFound desc = could not find container \"7f8e7416946f778d050fcb1cde342a1ba48997ed1598021a9d00aeedcd1fd256\": container with ID starting with 7f8e7416946f778d050fcb1cde342a1ba48997ed1598021a9d00aeedcd1fd256 not found: ID does not exist" Oct 02 02:44:24 crc kubenswrapper[4885]: I1002 02:44:24.700356 4885 scope.go:117] "RemoveContainer" containerID="dd3d6f108d96fd2c241bfa82683d3497a4bfe9a0fc30323bc0ed3325db018f90" Oct 02 02:44:24 crc kubenswrapper[4885]: E1002 02:44:24.701015 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dd3d6f108d96fd2c241bfa82683d3497a4bfe9a0fc30323bc0ed3325db018f90\": container with ID starting with dd3d6f108d96fd2c241bfa82683d3497a4bfe9a0fc30323bc0ed3325db018f90 not found: ID does not exist" containerID="dd3d6f108d96fd2c241bfa82683d3497a4bfe9a0fc30323bc0ed3325db018f90" Oct 02 02:44:24 crc kubenswrapper[4885]: I1002 02:44:24.701057 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd3d6f108d96fd2c241bfa82683d3497a4bfe9a0fc30323bc0ed3325db018f90"} err="failed to get container status \"dd3d6f108d96fd2c241bfa82683d3497a4bfe9a0fc30323bc0ed3325db018f90\": rpc error: code = NotFound desc = could not find container \"dd3d6f108d96fd2c241bfa82683d3497a4bfe9a0fc30323bc0ed3325db018f90\": container with ID starting with dd3d6f108d96fd2c241bfa82683d3497a4bfe9a0fc30323bc0ed3325db018f90 not found: ID does not exist" Oct 02 02:44:24 crc kubenswrapper[4885]: I1002 02:44:24.701086 4885 scope.go:117] "RemoveContainer" containerID="cdd1988a65ab0ecc9eb58da66a2d1edddc41abd212e92af838e7ad57bf203a41" Oct 02 02:44:24 crc kubenswrapper[4885]: E1002 02:44:24.701647 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cdd1988a65ab0ecc9eb58da66a2d1edddc41abd212e92af838e7ad57bf203a41\": container with ID starting with cdd1988a65ab0ecc9eb58da66a2d1edddc41abd212e92af838e7ad57bf203a41 not found: ID does not exist" containerID="cdd1988a65ab0ecc9eb58da66a2d1edddc41abd212e92af838e7ad57bf203a41" Oct 02 02:44:24 crc kubenswrapper[4885]: I1002 02:44:24.701678 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cdd1988a65ab0ecc9eb58da66a2d1edddc41abd212e92af838e7ad57bf203a41"} err="failed to get container status \"cdd1988a65ab0ecc9eb58da66a2d1edddc41abd212e92af838e7ad57bf203a41\": rpc error: code = NotFound desc = could not find container \"cdd1988a65ab0ecc9eb58da66a2d1edddc41abd212e92af838e7ad57bf203a41\": container with ID starting with cdd1988a65ab0ecc9eb58da66a2d1edddc41abd212e92af838e7ad57bf203a41 not found: ID does not exist" Oct 02 02:44:26 crc kubenswrapper[4885]: I1002 02:44:26.068572 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb50ccc8-3964-427b-95b7-fbe6d29bd7f7" path="/var/lib/kubelet/pods/bb50ccc8-3964-427b-95b7-fbe6d29bd7f7/volumes" Oct 02 02:45:00 crc kubenswrapper[4885]: I1002 02:45:00.223012 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322885-877kp"] Oct 02 02:45:00 crc kubenswrapper[4885]: E1002 02:45:00.224524 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb50ccc8-3964-427b-95b7-fbe6d29bd7f7" containerName="extract-content" Oct 02 02:45:00 crc kubenswrapper[4885]: I1002 02:45:00.224554 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb50ccc8-3964-427b-95b7-fbe6d29bd7f7" containerName="extract-content" Oct 02 02:45:00 crc kubenswrapper[4885]: E1002 02:45:00.224588 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb50ccc8-3964-427b-95b7-fbe6d29bd7f7" containerName="registry-server" Oct 02 02:45:00 crc kubenswrapper[4885]: I1002 02:45:00.224602 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb50ccc8-3964-427b-95b7-fbe6d29bd7f7" containerName="registry-server" Oct 02 02:45:00 crc kubenswrapper[4885]: E1002 02:45:00.224657 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb50ccc8-3964-427b-95b7-fbe6d29bd7f7" containerName="extract-utilities" Oct 02 02:45:00 crc kubenswrapper[4885]: I1002 02:45:00.224670 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb50ccc8-3964-427b-95b7-fbe6d29bd7f7" containerName="extract-utilities" Oct 02 02:45:00 crc kubenswrapper[4885]: I1002 02:45:00.225023 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb50ccc8-3964-427b-95b7-fbe6d29bd7f7" containerName="registry-server" Oct 02 02:45:00 crc kubenswrapper[4885]: I1002 02:45:00.226111 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322885-877kp" Oct 02 02:45:00 crc kubenswrapper[4885]: I1002 02:45:00.231124 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 02 02:45:00 crc kubenswrapper[4885]: I1002 02:45:00.231990 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 02 02:45:00 crc kubenswrapper[4885]: I1002 02:45:00.237104 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322885-877kp"] Oct 02 02:45:00 crc kubenswrapper[4885]: I1002 02:45:00.310201 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a73d6d25-c7c8-4154-8fb8-252c512b40cc-config-volume\") pod \"collect-profiles-29322885-877kp\" (UID: \"a73d6d25-c7c8-4154-8fb8-252c512b40cc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322885-877kp" Oct 02 02:45:00 crc kubenswrapper[4885]: I1002 02:45:00.310463 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cv4jq\" (UniqueName: \"kubernetes.io/projected/a73d6d25-c7c8-4154-8fb8-252c512b40cc-kube-api-access-cv4jq\") pod \"collect-profiles-29322885-877kp\" (UID: \"a73d6d25-c7c8-4154-8fb8-252c512b40cc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322885-877kp" Oct 02 02:45:00 crc kubenswrapper[4885]: I1002 02:45:00.310500 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a73d6d25-c7c8-4154-8fb8-252c512b40cc-secret-volume\") pod \"collect-profiles-29322885-877kp\" (UID: \"a73d6d25-c7c8-4154-8fb8-252c512b40cc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322885-877kp" Oct 02 02:45:00 crc kubenswrapper[4885]: I1002 02:45:00.412347 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a73d6d25-c7c8-4154-8fb8-252c512b40cc-config-volume\") pod \"collect-profiles-29322885-877kp\" (UID: \"a73d6d25-c7c8-4154-8fb8-252c512b40cc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322885-877kp" Oct 02 02:45:00 crc kubenswrapper[4885]: I1002 02:45:00.413412 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cv4jq\" (UniqueName: \"kubernetes.io/projected/a73d6d25-c7c8-4154-8fb8-252c512b40cc-kube-api-access-cv4jq\") pod \"collect-profiles-29322885-877kp\" (UID: \"a73d6d25-c7c8-4154-8fb8-252c512b40cc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322885-877kp" Oct 02 02:45:00 crc kubenswrapper[4885]: I1002 02:45:00.413620 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a73d6d25-c7c8-4154-8fb8-252c512b40cc-secret-volume\") pod \"collect-profiles-29322885-877kp\" (UID: \"a73d6d25-c7c8-4154-8fb8-252c512b40cc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322885-877kp" Oct 02 02:45:00 crc kubenswrapper[4885]: I1002 02:45:00.413895 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a73d6d25-c7c8-4154-8fb8-252c512b40cc-config-volume\") pod \"collect-profiles-29322885-877kp\" (UID: \"a73d6d25-c7c8-4154-8fb8-252c512b40cc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322885-877kp" Oct 02 02:45:00 crc kubenswrapper[4885]: I1002 02:45:00.423121 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a73d6d25-c7c8-4154-8fb8-252c512b40cc-secret-volume\") pod \"collect-profiles-29322885-877kp\" (UID: \"a73d6d25-c7c8-4154-8fb8-252c512b40cc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322885-877kp" Oct 02 02:45:00 crc kubenswrapper[4885]: I1002 02:45:00.437486 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cv4jq\" (UniqueName: \"kubernetes.io/projected/a73d6d25-c7c8-4154-8fb8-252c512b40cc-kube-api-access-cv4jq\") pod \"collect-profiles-29322885-877kp\" (UID: \"a73d6d25-c7c8-4154-8fb8-252c512b40cc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322885-877kp" Oct 02 02:45:00 crc kubenswrapper[4885]: I1002 02:45:00.564111 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322885-877kp" Oct 02 02:45:01 crc kubenswrapper[4885]: I1002 02:45:01.077244 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322885-877kp"] Oct 02 02:45:01 crc kubenswrapper[4885]: I1002 02:45:01.953761 4885 generic.go:334] "Generic (PLEG): container finished" podID="a73d6d25-c7c8-4154-8fb8-252c512b40cc" containerID="38f480a997a3d059fca1483f23629f0f007a306e89d45696b4e3bf2ac5840d11" exitCode=0 Oct 02 02:45:01 crc kubenswrapper[4885]: I1002 02:45:01.953816 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29322885-877kp" event={"ID":"a73d6d25-c7c8-4154-8fb8-252c512b40cc","Type":"ContainerDied","Data":"38f480a997a3d059fca1483f23629f0f007a306e89d45696b4e3bf2ac5840d11"} Oct 02 02:45:01 crc kubenswrapper[4885]: I1002 02:45:01.954080 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29322885-877kp" event={"ID":"a73d6d25-c7c8-4154-8fb8-252c512b40cc","Type":"ContainerStarted","Data":"f6a0f4d55c79dc76085c984b68c3722ebca0dec7c90ffbc2db86b8425a6a8db4"} Oct 02 02:45:03 crc kubenswrapper[4885]: I1002 02:45:03.402239 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322885-877kp" Oct 02 02:45:03 crc kubenswrapper[4885]: I1002 02:45:03.556306 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a73d6d25-c7c8-4154-8fb8-252c512b40cc-config-volume\") pod \"a73d6d25-c7c8-4154-8fb8-252c512b40cc\" (UID: \"a73d6d25-c7c8-4154-8fb8-252c512b40cc\") " Oct 02 02:45:03 crc kubenswrapper[4885]: I1002 02:45:03.556749 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cv4jq\" (UniqueName: \"kubernetes.io/projected/a73d6d25-c7c8-4154-8fb8-252c512b40cc-kube-api-access-cv4jq\") pod \"a73d6d25-c7c8-4154-8fb8-252c512b40cc\" (UID: \"a73d6d25-c7c8-4154-8fb8-252c512b40cc\") " Oct 02 02:45:03 crc kubenswrapper[4885]: I1002 02:45:03.556874 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a73d6d25-c7c8-4154-8fb8-252c512b40cc-secret-volume\") pod \"a73d6d25-c7c8-4154-8fb8-252c512b40cc\" (UID: \"a73d6d25-c7c8-4154-8fb8-252c512b40cc\") " Oct 02 02:45:03 crc kubenswrapper[4885]: I1002 02:45:03.558236 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a73d6d25-c7c8-4154-8fb8-252c512b40cc-config-volume" (OuterVolumeSpecName: "config-volume") pod "a73d6d25-c7c8-4154-8fb8-252c512b40cc" (UID: "a73d6d25-c7c8-4154-8fb8-252c512b40cc"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:45:03 crc kubenswrapper[4885]: I1002 02:45:03.563016 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a73d6d25-c7c8-4154-8fb8-252c512b40cc-kube-api-access-cv4jq" (OuterVolumeSpecName: "kube-api-access-cv4jq") pod "a73d6d25-c7c8-4154-8fb8-252c512b40cc" (UID: "a73d6d25-c7c8-4154-8fb8-252c512b40cc"). InnerVolumeSpecName "kube-api-access-cv4jq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:45:03 crc kubenswrapper[4885]: I1002 02:45:03.568553 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a73d6d25-c7c8-4154-8fb8-252c512b40cc-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "a73d6d25-c7c8-4154-8fb8-252c512b40cc" (UID: "a73d6d25-c7c8-4154-8fb8-252c512b40cc"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:45:03 crc kubenswrapper[4885]: I1002 02:45:03.659569 4885 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a73d6d25-c7c8-4154-8fb8-252c512b40cc-config-volume\") on node \"crc\" DevicePath \"\"" Oct 02 02:45:03 crc kubenswrapper[4885]: I1002 02:45:03.659626 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cv4jq\" (UniqueName: \"kubernetes.io/projected/a73d6d25-c7c8-4154-8fb8-252c512b40cc-kube-api-access-cv4jq\") on node \"crc\" DevicePath \"\"" Oct 02 02:45:03 crc kubenswrapper[4885]: I1002 02:45:03.659648 4885 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a73d6d25-c7c8-4154-8fb8-252c512b40cc-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 02 02:45:03 crc kubenswrapper[4885]: I1002 02:45:03.992801 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29322885-877kp" event={"ID":"a73d6d25-c7c8-4154-8fb8-252c512b40cc","Type":"ContainerDied","Data":"f6a0f4d55c79dc76085c984b68c3722ebca0dec7c90ffbc2db86b8425a6a8db4"} Oct 02 02:45:03 crc kubenswrapper[4885]: I1002 02:45:03.992844 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322885-877kp" Oct 02 02:45:03 crc kubenswrapper[4885]: I1002 02:45:03.992863 4885 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f6a0f4d55c79dc76085c984b68c3722ebca0dec7c90ffbc2db86b8425a6a8db4" Oct 02 02:45:04 crc kubenswrapper[4885]: I1002 02:45:04.507481 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322840-56lsm"] Oct 02 02:45:04 crc kubenswrapper[4885]: I1002 02:45:04.514686 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322840-56lsm"] Oct 02 02:45:06 crc kubenswrapper[4885]: I1002 02:45:06.064828 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e32595a5-56a6-4fb8-8f7d-41363c7f57d2" path="/var/lib/kubelet/pods/e32595a5-56a6-4fb8-8f7d-41363c7f57d2/volumes" Oct 02 02:45:15 crc kubenswrapper[4885]: I1002 02:45:15.189940 4885 scope.go:117] "RemoveContainer" containerID="ecd566acdf250db83dd67b6c57afa0780263d338c8eee6e4a4cd7c37be95e031" Oct 02 02:46:13 crc kubenswrapper[4885]: I1002 02:46:13.266583 4885 patch_prober.go:28] interesting pod/machine-config-daemon-rttx8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 02:46:13 crc kubenswrapper[4885]: I1002 02:46:13.267252 4885 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 02:46:37 crc kubenswrapper[4885]: I1002 02:46:37.057169 4885 generic.go:334] "Generic (PLEG): container finished" podID="7a715636-06d0-4508-a9d0-fe83f4c05d01" containerID="226fd3dedfa1769e7bf88d810b0326fa5df173290057a082a896836458a0ef54" exitCode=0 Oct 02 02:46:37 crc kubenswrapper[4885]: I1002 02:46:37.057308 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"7a715636-06d0-4508-a9d0-fe83f4c05d01","Type":"ContainerDied","Data":"226fd3dedfa1769e7bf88d810b0326fa5df173290057a082a896836458a0ef54"} Oct 02 02:46:38 crc kubenswrapper[4885]: I1002 02:46:38.502903 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 02 02:46:38 crc kubenswrapper[4885]: I1002 02:46:38.667755 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vjxlh\" (UniqueName: \"kubernetes.io/projected/7a715636-06d0-4508-a9d0-fe83f4c05d01-kube-api-access-vjxlh\") pod \"7a715636-06d0-4508-a9d0-fe83f4c05d01\" (UID: \"7a715636-06d0-4508-a9d0-fe83f4c05d01\") " Oct 02 02:46:38 crc kubenswrapper[4885]: I1002 02:46:38.667867 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/7a715636-06d0-4508-a9d0-fe83f4c05d01-ca-certs\") pod \"7a715636-06d0-4508-a9d0-fe83f4c05d01\" (UID: \"7a715636-06d0-4508-a9d0-fe83f4c05d01\") " Oct 02 02:46:38 crc kubenswrapper[4885]: I1002 02:46:38.667967 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/7a715636-06d0-4508-a9d0-fe83f4c05d01-openstack-config\") pod \"7a715636-06d0-4508-a9d0-fe83f4c05d01\" (UID: \"7a715636-06d0-4508-a9d0-fe83f4c05d01\") " Oct 02 02:46:38 crc kubenswrapper[4885]: I1002 02:46:38.668079 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"7a715636-06d0-4508-a9d0-fe83f4c05d01\" (UID: \"7a715636-06d0-4508-a9d0-fe83f4c05d01\") " Oct 02 02:46:38 crc kubenswrapper[4885]: I1002 02:46:38.671147 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/7a715636-06d0-4508-a9d0-fe83f4c05d01-test-operator-ephemeral-workdir\") pod \"7a715636-06d0-4508-a9d0-fe83f4c05d01\" (UID: \"7a715636-06d0-4508-a9d0-fe83f4c05d01\") " Oct 02 02:46:38 crc kubenswrapper[4885]: I1002 02:46:38.671226 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7a715636-06d0-4508-a9d0-fe83f4c05d01-config-data\") pod \"7a715636-06d0-4508-a9d0-fe83f4c05d01\" (UID: \"7a715636-06d0-4508-a9d0-fe83f4c05d01\") " Oct 02 02:46:38 crc kubenswrapper[4885]: I1002 02:46:38.684481 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "test-operator-logs") pod "7a715636-06d0-4508-a9d0-fe83f4c05d01" (UID: "7a715636-06d0-4508-a9d0-fe83f4c05d01"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 02 02:46:38 crc kubenswrapper[4885]: I1002 02:46:38.685097 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7a715636-06d0-4508-a9d0-fe83f4c05d01-config-data" (OuterVolumeSpecName: "config-data") pod "7a715636-06d0-4508-a9d0-fe83f4c05d01" (UID: "7a715636-06d0-4508-a9d0-fe83f4c05d01"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:46:38 crc kubenswrapper[4885]: I1002 02:46:38.685189 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/7a715636-06d0-4508-a9d0-fe83f4c05d01-openstack-config-secret\") pod \"7a715636-06d0-4508-a9d0-fe83f4c05d01\" (UID: \"7a715636-06d0-4508-a9d0-fe83f4c05d01\") " Oct 02 02:46:38 crc kubenswrapper[4885]: I1002 02:46:38.685398 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/7a715636-06d0-4508-a9d0-fe83f4c05d01-test-operator-ephemeral-temporary\") pod \"7a715636-06d0-4508-a9d0-fe83f4c05d01\" (UID: \"7a715636-06d0-4508-a9d0-fe83f4c05d01\") " Oct 02 02:46:38 crc kubenswrapper[4885]: I1002 02:46:38.685442 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7a715636-06d0-4508-a9d0-fe83f4c05d01-ssh-key\") pod \"7a715636-06d0-4508-a9d0-fe83f4c05d01\" (UID: \"7a715636-06d0-4508-a9d0-fe83f4c05d01\") " Oct 02 02:46:38 crc kubenswrapper[4885]: I1002 02:46:38.686900 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7a715636-06d0-4508-a9d0-fe83f4c05d01-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "7a715636-06d0-4508-a9d0-fe83f4c05d01" (UID: "7a715636-06d0-4508-a9d0-fe83f4c05d01"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:46:38 crc kubenswrapper[4885]: I1002 02:46:38.687335 4885 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/7a715636-06d0-4508-a9d0-fe83f4c05d01-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Oct 02 02:46:38 crc kubenswrapper[4885]: I1002 02:46:38.687367 4885 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Oct 02 02:46:38 crc kubenswrapper[4885]: I1002 02:46:38.687378 4885 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7a715636-06d0-4508-a9d0-fe83f4c05d01-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 02:46:38 crc kubenswrapper[4885]: I1002 02:46:38.691692 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7a715636-06d0-4508-a9d0-fe83f4c05d01-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "7a715636-06d0-4508-a9d0-fe83f4c05d01" (UID: "7a715636-06d0-4508-a9d0-fe83f4c05d01"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:46:38 crc kubenswrapper[4885]: I1002 02:46:38.703484 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a715636-06d0-4508-a9d0-fe83f4c05d01-kube-api-access-vjxlh" (OuterVolumeSpecName: "kube-api-access-vjxlh") pod "7a715636-06d0-4508-a9d0-fe83f4c05d01" (UID: "7a715636-06d0-4508-a9d0-fe83f4c05d01"). InnerVolumeSpecName "kube-api-access-vjxlh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:46:38 crc kubenswrapper[4885]: I1002 02:46:38.709298 4885 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Oct 02 02:46:38 crc kubenswrapper[4885]: I1002 02:46:38.719456 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a715636-06d0-4508-a9d0-fe83f4c05d01-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "7a715636-06d0-4508-a9d0-fe83f4c05d01" (UID: "7a715636-06d0-4508-a9d0-fe83f4c05d01"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:46:38 crc kubenswrapper[4885]: I1002 02:46:38.721404 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a715636-06d0-4508-a9d0-fe83f4c05d01-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "7a715636-06d0-4508-a9d0-fe83f4c05d01" (UID: "7a715636-06d0-4508-a9d0-fe83f4c05d01"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:46:38 crc kubenswrapper[4885]: I1002 02:46:38.728077 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a715636-06d0-4508-a9d0-fe83f4c05d01-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "7a715636-06d0-4508-a9d0-fe83f4c05d01" (UID: "7a715636-06d0-4508-a9d0-fe83f4c05d01"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:46:38 crc kubenswrapper[4885]: I1002 02:46:38.751192 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7a715636-06d0-4508-a9d0-fe83f4c05d01-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "7a715636-06d0-4508-a9d0-fe83f4c05d01" (UID: "7a715636-06d0-4508-a9d0-fe83f4c05d01"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:46:38 crc kubenswrapper[4885]: I1002 02:46:38.788556 4885 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/7a715636-06d0-4508-a9d0-fe83f4c05d01-ca-certs\") on node \"crc\" DevicePath \"\"" Oct 02 02:46:38 crc kubenswrapper[4885]: I1002 02:46:38.788589 4885 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/7a715636-06d0-4508-a9d0-fe83f4c05d01-openstack-config\") on node \"crc\" DevicePath \"\"" Oct 02 02:46:38 crc kubenswrapper[4885]: I1002 02:46:38.788600 4885 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Oct 02 02:46:38 crc kubenswrapper[4885]: I1002 02:46:38.788610 4885 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/7a715636-06d0-4508-a9d0-fe83f4c05d01-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Oct 02 02:46:38 crc kubenswrapper[4885]: I1002 02:46:38.788621 4885 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/7a715636-06d0-4508-a9d0-fe83f4c05d01-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 02 02:46:38 crc kubenswrapper[4885]: I1002 02:46:38.788630 4885 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7a715636-06d0-4508-a9d0-fe83f4c05d01-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 02:46:38 crc kubenswrapper[4885]: I1002 02:46:38.788640 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vjxlh\" (UniqueName: \"kubernetes.io/projected/7a715636-06d0-4508-a9d0-fe83f4c05d01-kube-api-access-vjxlh\") on node \"crc\" DevicePath \"\"" Oct 02 02:46:39 crc kubenswrapper[4885]: I1002 02:46:39.091947 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"7a715636-06d0-4508-a9d0-fe83f4c05d01","Type":"ContainerDied","Data":"fe8572d94ee5941888c10edcfa3a7c885868c94bf5ad32a55ddb21de0cfacc4a"} Oct 02 02:46:39 crc kubenswrapper[4885]: I1002 02:46:39.091992 4885 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fe8572d94ee5941888c10edcfa3a7c885868c94bf5ad32a55ddb21de0cfacc4a" Oct 02 02:46:39 crc kubenswrapper[4885]: I1002 02:46:39.091998 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 02 02:46:43 crc kubenswrapper[4885]: I1002 02:46:43.265676 4885 patch_prober.go:28] interesting pod/machine-config-daemon-rttx8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 02:46:43 crc kubenswrapper[4885]: I1002 02:46:43.266369 4885 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 02:46:48 crc kubenswrapper[4885]: I1002 02:46:48.563198 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Oct 02 02:46:48 crc kubenswrapper[4885]: E1002 02:46:48.564674 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a715636-06d0-4508-a9d0-fe83f4c05d01" containerName="tempest-tests-tempest-tests-runner" Oct 02 02:46:48 crc kubenswrapper[4885]: I1002 02:46:48.564703 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a715636-06d0-4508-a9d0-fe83f4c05d01" containerName="tempest-tests-tempest-tests-runner" Oct 02 02:46:48 crc kubenswrapper[4885]: E1002 02:46:48.564780 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a73d6d25-c7c8-4154-8fb8-252c512b40cc" containerName="collect-profiles" Oct 02 02:46:48 crc kubenswrapper[4885]: I1002 02:46:48.564794 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="a73d6d25-c7c8-4154-8fb8-252c512b40cc" containerName="collect-profiles" Oct 02 02:46:48 crc kubenswrapper[4885]: I1002 02:46:48.565121 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a715636-06d0-4508-a9d0-fe83f4c05d01" containerName="tempest-tests-tempest-tests-runner" Oct 02 02:46:48 crc kubenswrapper[4885]: I1002 02:46:48.565169 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="a73d6d25-c7c8-4154-8fb8-252c512b40cc" containerName="collect-profiles" Oct 02 02:46:48 crc kubenswrapper[4885]: I1002 02:46:48.566532 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 02 02:46:48 crc kubenswrapper[4885]: I1002 02:46:48.575646 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-mg457" Oct 02 02:46:48 crc kubenswrapper[4885]: I1002 02:46:48.580429 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Oct 02 02:46:48 crc kubenswrapper[4885]: I1002 02:46:48.728794 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6r6s8\" (UniqueName: \"kubernetes.io/projected/9cf08307-1f33-4989-b935-67c8f4a76cc4-kube-api-access-6r6s8\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"9cf08307-1f33-4989-b935-67c8f4a76cc4\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 02 02:46:48 crc kubenswrapper[4885]: I1002 02:46:48.729084 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"9cf08307-1f33-4989-b935-67c8f4a76cc4\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 02 02:46:48 crc kubenswrapper[4885]: I1002 02:46:48.831389 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"9cf08307-1f33-4989-b935-67c8f4a76cc4\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 02 02:46:48 crc kubenswrapper[4885]: I1002 02:46:48.832000 4885 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"9cf08307-1f33-4989-b935-67c8f4a76cc4\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 02 02:46:48 crc kubenswrapper[4885]: I1002 02:46:48.832131 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6r6s8\" (UniqueName: \"kubernetes.io/projected/9cf08307-1f33-4989-b935-67c8f4a76cc4-kube-api-access-6r6s8\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"9cf08307-1f33-4989-b935-67c8f4a76cc4\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 02 02:46:48 crc kubenswrapper[4885]: I1002 02:46:48.869094 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6r6s8\" (UniqueName: \"kubernetes.io/projected/9cf08307-1f33-4989-b935-67c8f4a76cc4-kube-api-access-6r6s8\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"9cf08307-1f33-4989-b935-67c8f4a76cc4\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 02 02:46:48 crc kubenswrapper[4885]: I1002 02:46:48.869913 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"9cf08307-1f33-4989-b935-67c8f4a76cc4\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 02 02:46:48 crc kubenswrapper[4885]: I1002 02:46:48.901511 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 02 02:46:49 crc kubenswrapper[4885]: I1002 02:46:49.406205 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Oct 02 02:46:50 crc kubenswrapper[4885]: I1002 02:46:50.214348 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"9cf08307-1f33-4989-b935-67c8f4a76cc4","Type":"ContainerStarted","Data":"448aa3205d2d6fcc874cfe6591dc54b028f345702ecd08f9ec273458f6167f0b"} Oct 02 02:46:51 crc kubenswrapper[4885]: I1002 02:46:51.244775 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"9cf08307-1f33-4989-b935-67c8f4a76cc4","Type":"ContainerStarted","Data":"8ed23e3ca752b866f6c192a65ce80a5fc4af378d5995ce477df63504b7ea20af"} Oct 02 02:46:51 crc kubenswrapper[4885]: I1002 02:46:51.274916 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=2.314569887 podStartE2EDuration="3.274891494s" podCreationTimestamp="2025-10-02 02:46:48 +0000 UTC" firstStartedPulling="2025-10-02 02:46:49.41029191 +0000 UTC m=+3598.222039349" lastFinishedPulling="2025-10-02 02:46:50.370613517 +0000 UTC m=+3599.182360956" observedRunningTime="2025-10-02 02:46:51.265111886 +0000 UTC m=+3600.076859315" watchObservedRunningTime="2025-10-02 02:46:51.274891494 +0000 UTC m=+3600.086638933" Oct 02 02:47:07 crc kubenswrapper[4885]: I1002 02:47:07.350989 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-28twm/must-gather-s92fx"] Oct 02 02:47:07 crc kubenswrapper[4885]: I1002 02:47:07.353246 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-28twm/must-gather-s92fx" Oct 02 02:47:07 crc kubenswrapper[4885]: I1002 02:47:07.355097 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-28twm"/"kube-root-ca.crt" Oct 02 02:47:07 crc kubenswrapper[4885]: I1002 02:47:07.355437 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-28twm"/"openshift-service-ca.crt" Oct 02 02:47:07 crc kubenswrapper[4885]: I1002 02:47:07.358887 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-28twm/must-gather-s92fx"] Oct 02 02:47:07 crc kubenswrapper[4885]: I1002 02:47:07.359596 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-28twm"/"default-dockercfg-9dvp2" Oct 02 02:47:07 crc kubenswrapper[4885]: I1002 02:47:07.467671 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q9pkh\" (UniqueName: \"kubernetes.io/projected/408c4e51-14de-441d-a876-e89a483ceef4-kube-api-access-q9pkh\") pod \"must-gather-s92fx\" (UID: \"408c4e51-14de-441d-a876-e89a483ceef4\") " pod="openshift-must-gather-28twm/must-gather-s92fx" Oct 02 02:47:07 crc kubenswrapper[4885]: I1002 02:47:07.467989 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/408c4e51-14de-441d-a876-e89a483ceef4-must-gather-output\") pod \"must-gather-s92fx\" (UID: \"408c4e51-14de-441d-a876-e89a483ceef4\") " pod="openshift-must-gather-28twm/must-gather-s92fx" Oct 02 02:47:07 crc kubenswrapper[4885]: I1002 02:47:07.569919 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q9pkh\" (UniqueName: \"kubernetes.io/projected/408c4e51-14de-441d-a876-e89a483ceef4-kube-api-access-q9pkh\") pod \"must-gather-s92fx\" (UID: \"408c4e51-14de-441d-a876-e89a483ceef4\") " pod="openshift-must-gather-28twm/must-gather-s92fx" Oct 02 02:47:07 crc kubenswrapper[4885]: I1002 02:47:07.570350 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/408c4e51-14de-441d-a876-e89a483ceef4-must-gather-output\") pod \"must-gather-s92fx\" (UID: \"408c4e51-14de-441d-a876-e89a483ceef4\") " pod="openshift-must-gather-28twm/must-gather-s92fx" Oct 02 02:47:07 crc kubenswrapper[4885]: I1002 02:47:07.570698 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/408c4e51-14de-441d-a876-e89a483ceef4-must-gather-output\") pod \"must-gather-s92fx\" (UID: \"408c4e51-14de-441d-a876-e89a483ceef4\") " pod="openshift-must-gather-28twm/must-gather-s92fx" Oct 02 02:47:07 crc kubenswrapper[4885]: I1002 02:47:07.601405 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q9pkh\" (UniqueName: \"kubernetes.io/projected/408c4e51-14de-441d-a876-e89a483ceef4-kube-api-access-q9pkh\") pod \"must-gather-s92fx\" (UID: \"408c4e51-14de-441d-a876-e89a483ceef4\") " pod="openshift-must-gather-28twm/must-gather-s92fx" Oct 02 02:47:07 crc kubenswrapper[4885]: I1002 02:47:07.670487 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-28twm/must-gather-s92fx" Oct 02 02:47:08 crc kubenswrapper[4885]: I1002 02:47:08.265961 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-28twm/must-gather-s92fx"] Oct 02 02:47:08 crc kubenswrapper[4885]: I1002 02:47:08.440503 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-28twm/must-gather-s92fx" event={"ID":"408c4e51-14de-441d-a876-e89a483ceef4","Type":"ContainerStarted","Data":"da421f643a879533ca52657f9e779d154be43bc18a9dc2d14ab2bfcbee3e11b7"} Oct 02 02:47:13 crc kubenswrapper[4885]: I1002 02:47:13.265667 4885 patch_prober.go:28] interesting pod/machine-config-daemon-rttx8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 02:47:13 crc kubenswrapper[4885]: I1002 02:47:13.266163 4885 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 02:47:13 crc kubenswrapper[4885]: I1002 02:47:13.266213 4885 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" Oct 02 02:47:13 crc kubenswrapper[4885]: I1002 02:47:13.266984 4885 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"999b3e75477511af3aa11f1e9d932c1ffb739b6f215cc19f45ac7144051ab65c"} pod="openshift-machine-config-operator/machine-config-daemon-rttx8" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 02:47:13 crc kubenswrapper[4885]: I1002 02:47:13.267036 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" containerName="machine-config-daemon" containerID="cri-o://999b3e75477511af3aa11f1e9d932c1ffb739b6f215cc19f45ac7144051ab65c" gracePeriod=600 Oct 02 02:47:13 crc kubenswrapper[4885]: I1002 02:47:13.487770 4885 generic.go:334] "Generic (PLEG): container finished" podID="d169c8ba-000a-4218-8e3b-4ae53035b110" containerID="999b3e75477511af3aa11f1e9d932c1ffb739b6f215cc19f45ac7144051ab65c" exitCode=0 Oct 02 02:47:13 crc kubenswrapper[4885]: I1002 02:47:13.488065 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" event={"ID":"d169c8ba-000a-4218-8e3b-4ae53035b110","Type":"ContainerDied","Data":"999b3e75477511af3aa11f1e9d932c1ffb739b6f215cc19f45ac7144051ab65c"} Oct 02 02:47:13 crc kubenswrapper[4885]: I1002 02:47:13.488100 4885 scope.go:117] "RemoveContainer" containerID="263b36668b739af6c8b8dd19175897e6595f755c49d0d124cba769111393cb93" Oct 02 02:47:14 crc kubenswrapper[4885]: E1002 02:47:14.329223 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:47:14 crc kubenswrapper[4885]: I1002 02:47:14.501496 4885 scope.go:117] "RemoveContainer" containerID="999b3e75477511af3aa11f1e9d932c1ffb739b6f215cc19f45ac7144051ab65c" Oct 02 02:47:14 crc kubenswrapper[4885]: E1002 02:47:14.501807 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:47:15 crc kubenswrapper[4885]: I1002 02:47:15.512961 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-28twm/must-gather-s92fx" event={"ID":"408c4e51-14de-441d-a876-e89a483ceef4","Type":"ContainerStarted","Data":"dfa749ea2232255db4d7ab8402d46b0523e517e5abd73867d391cb29be8c11c3"} Oct 02 02:47:15 crc kubenswrapper[4885]: I1002 02:47:15.513493 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-28twm/must-gather-s92fx" event={"ID":"408c4e51-14de-441d-a876-e89a483ceef4","Type":"ContainerStarted","Data":"e9a0e101e717ebc532f542697393207e356a45656d9b3b7b793fd1312ab2b4be"} Oct 02 02:47:15 crc kubenswrapper[4885]: I1002 02:47:15.535731 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-28twm/must-gather-s92fx" podStartSLOduration=2.343122423 podStartE2EDuration="8.535665663s" podCreationTimestamp="2025-10-02 02:47:07 +0000 UTC" firstStartedPulling="2025-10-02 02:47:08.26623316 +0000 UTC m=+3617.077980589" lastFinishedPulling="2025-10-02 02:47:14.45877642 +0000 UTC m=+3623.270523829" observedRunningTime="2025-10-02 02:47:15.527126862 +0000 UTC m=+3624.338874301" watchObservedRunningTime="2025-10-02 02:47:15.535665663 +0000 UTC m=+3624.347413112" Oct 02 02:47:17 crc kubenswrapper[4885]: E1002 02:47:17.652601 4885 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.39:52092->38.102.83.39:35229: write tcp 38.102.83.39:52092->38.102.83.39:35229: write: connection reset by peer Oct 02 02:47:18 crc kubenswrapper[4885]: I1002 02:47:18.299711 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-28twm/crc-debug-mglll"] Oct 02 02:47:18 crc kubenswrapper[4885]: I1002 02:47:18.300939 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-28twm/crc-debug-mglll" Oct 02 02:47:18 crc kubenswrapper[4885]: I1002 02:47:18.376374 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gkztp\" (UniqueName: \"kubernetes.io/projected/12ee7c20-678b-4f11-8d6e-96bd3a4a98f2-kube-api-access-gkztp\") pod \"crc-debug-mglll\" (UID: \"12ee7c20-678b-4f11-8d6e-96bd3a4a98f2\") " pod="openshift-must-gather-28twm/crc-debug-mglll" Oct 02 02:47:18 crc kubenswrapper[4885]: I1002 02:47:18.376461 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/12ee7c20-678b-4f11-8d6e-96bd3a4a98f2-host\") pod \"crc-debug-mglll\" (UID: \"12ee7c20-678b-4f11-8d6e-96bd3a4a98f2\") " pod="openshift-must-gather-28twm/crc-debug-mglll" Oct 02 02:47:18 crc kubenswrapper[4885]: I1002 02:47:18.479091 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gkztp\" (UniqueName: \"kubernetes.io/projected/12ee7c20-678b-4f11-8d6e-96bd3a4a98f2-kube-api-access-gkztp\") pod \"crc-debug-mglll\" (UID: \"12ee7c20-678b-4f11-8d6e-96bd3a4a98f2\") " pod="openshift-must-gather-28twm/crc-debug-mglll" Oct 02 02:47:18 crc kubenswrapper[4885]: I1002 02:47:18.479215 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/12ee7c20-678b-4f11-8d6e-96bd3a4a98f2-host\") pod \"crc-debug-mglll\" (UID: \"12ee7c20-678b-4f11-8d6e-96bd3a4a98f2\") " pod="openshift-must-gather-28twm/crc-debug-mglll" Oct 02 02:47:18 crc kubenswrapper[4885]: I1002 02:47:18.479345 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/12ee7c20-678b-4f11-8d6e-96bd3a4a98f2-host\") pod \"crc-debug-mglll\" (UID: \"12ee7c20-678b-4f11-8d6e-96bd3a4a98f2\") " pod="openshift-must-gather-28twm/crc-debug-mglll" Oct 02 02:47:18 crc kubenswrapper[4885]: I1002 02:47:18.505023 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gkztp\" (UniqueName: \"kubernetes.io/projected/12ee7c20-678b-4f11-8d6e-96bd3a4a98f2-kube-api-access-gkztp\") pod \"crc-debug-mglll\" (UID: \"12ee7c20-678b-4f11-8d6e-96bd3a4a98f2\") " pod="openshift-must-gather-28twm/crc-debug-mglll" Oct 02 02:47:18 crc kubenswrapper[4885]: I1002 02:47:18.618933 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-28twm/crc-debug-mglll" Oct 02 02:47:18 crc kubenswrapper[4885]: W1002 02:47:18.657060 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod12ee7c20_678b_4f11_8d6e_96bd3a4a98f2.slice/crio-0198a32fa697ca27fee4f885d9d530f15e11c8f8d65d55a801ff7feff5eeb5f5 WatchSource:0}: Error finding container 0198a32fa697ca27fee4f885d9d530f15e11c8f8d65d55a801ff7feff5eeb5f5: Status 404 returned error can't find the container with id 0198a32fa697ca27fee4f885d9d530f15e11c8f8d65d55a801ff7feff5eeb5f5 Oct 02 02:47:19 crc kubenswrapper[4885]: I1002 02:47:19.558664 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-28twm/crc-debug-mglll" event={"ID":"12ee7c20-678b-4f11-8d6e-96bd3a4a98f2","Type":"ContainerStarted","Data":"0198a32fa697ca27fee4f885d9d530f15e11c8f8d65d55a801ff7feff5eeb5f5"} Oct 02 02:47:29 crc kubenswrapper[4885]: I1002 02:47:29.047035 4885 scope.go:117] "RemoveContainer" containerID="999b3e75477511af3aa11f1e9d932c1ffb739b6f215cc19f45ac7144051ab65c" Oct 02 02:47:29 crc kubenswrapper[4885]: E1002 02:47:29.048063 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:47:30 crc kubenswrapper[4885]: I1002 02:47:30.646112 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-28twm/crc-debug-mglll" event={"ID":"12ee7c20-678b-4f11-8d6e-96bd3a4a98f2","Type":"ContainerStarted","Data":"a663470245620171417c0f2c6b2c2043717fcae45eacf2e2b432c2eca245f639"} Oct 02 02:47:30 crc kubenswrapper[4885]: I1002 02:47:30.663639 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-28twm/crc-debug-mglll" podStartSLOduration=1.661476264 podStartE2EDuration="12.663614152s" podCreationTimestamp="2025-10-02 02:47:18 +0000 UTC" firstStartedPulling="2025-10-02 02:47:18.659286881 +0000 UTC m=+3627.471034280" lastFinishedPulling="2025-10-02 02:47:29.661424729 +0000 UTC m=+3638.473172168" observedRunningTime="2025-10-02 02:47:30.657739213 +0000 UTC m=+3639.469486612" watchObservedRunningTime="2025-10-02 02:47:30.663614152 +0000 UTC m=+3639.475361581" Oct 02 02:47:41 crc kubenswrapper[4885]: I1002 02:47:41.047008 4885 scope.go:117] "RemoveContainer" containerID="999b3e75477511af3aa11f1e9d932c1ffb739b6f215cc19f45ac7144051ab65c" Oct 02 02:47:41 crc kubenswrapper[4885]: E1002 02:47:41.047896 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:47:52 crc kubenswrapper[4885]: I1002 02:47:52.058513 4885 scope.go:117] "RemoveContainer" containerID="999b3e75477511af3aa11f1e9d932c1ffb739b6f215cc19f45ac7144051ab65c" Oct 02 02:47:52 crc kubenswrapper[4885]: E1002 02:47:52.059201 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:48:06 crc kubenswrapper[4885]: I1002 02:48:06.048844 4885 scope.go:117] "RemoveContainer" containerID="999b3e75477511af3aa11f1e9d932c1ffb739b6f215cc19f45ac7144051ab65c" Oct 02 02:48:06 crc kubenswrapper[4885]: E1002 02:48:06.050190 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:48:16 crc kubenswrapper[4885]: I1002 02:48:16.604759 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-6b5d8cdf54-gr28p_af70565f-bc37-43bc-97a3-3942c35a9fb2/barbican-api-log/0.log" Oct 02 02:48:16 crc kubenswrapper[4885]: I1002 02:48:16.631013 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-6b5d8cdf54-gr28p_af70565f-bc37-43bc-97a3-3942c35a9fb2/barbican-api/0.log" Oct 02 02:48:16 crc kubenswrapper[4885]: I1002 02:48:16.861053 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-6b9ddcb5cd-2r5wd_c416d7da-12d2-4aee-a9c3-5f171a791b43/barbican-keystone-listener-log/0.log" Oct 02 02:48:16 crc kubenswrapper[4885]: I1002 02:48:16.864134 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-6b9ddcb5cd-2r5wd_c416d7da-12d2-4aee-a9c3-5f171a791b43/barbican-keystone-listener/0.log" Oct 02 02:48:17 crc kubenswrapper[4885]: I1002 02:48:17.038834 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5c6b9856b7-vhzx6_075adfcf-950d-44e6-ab5e-826f3b06c05e/barbican-worker-log/0.log" Oct 02 02:48:17 crc kubenswrapper[4885]: I1002 02:48:17.043379 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5c6b9856b7-vhzx6_075adfcf-950d-44e6-ab5e-826f3b06c05e/barbican-worker/0.log" Oct 02 02:48:17 crc kubenswrapper[4885]: I1002 02:48:17.473609 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-rgbbj_8cba1665-9cca-44ef-a790-5cd9b174e726/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 02:48:17 crc kubenswrapper[4885]: I1002 02:48:17.498423 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_0fab80f5-cc23-4ed6-8c0b-1604d9809341/ceilometer-notification-agent/0.log" Oct 02 02:48:17 crc kubenswrapper[4885]: I1002 02:48:17.501443 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_0fab80f5-cc23-4ed6-8c0b-1604d9809341/ceilometer-central-agent/0.log" Oct 02 02:48:17 crc kubenswrapper[4885]: I1002 02:48:17.688232 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_0fab80f5-cc23-4ed6-8c0b-1604d9809341/proxy-httpd/0.log" Oct 02 02:48:17 crc kubenswrapper[4885]: I1002 02:48:17.691839 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_0fab80f5-cc23-4ed6-8c0b-1604d9809341/sg-core/0.log" Oct 02 02:48:17 crc kubenswrapper[4885]: I1002 02:48:17.872447 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_ca40a36d-76b5-4cfe-8f8c-ccb4625d6f56/cinder-api-log/0.log" Oct 02 02:48:17 crc kubenswrapper[4885]: I1002 02:48:17.913180 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_ca40a36d-76b5-4cfe-8f8c-ccb4625d6f56/cinder-api/0.log" Oct 02 02:48:18 crc kubenswrapper[4885]: I1002 02:48:18.046982 4885 scope.go:117] "RemoveContainer" containerID="999b3e75477511af3aa11f1e9d932c1ffb739b6f215cc19f45ac7144051ab65c" Oct 02 02:48:18 crc kubenswrapper[4885]: E1002 02:48:18.047234 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:48:18 crc kubenswrapper[4885]: I1002 02:48:18.077192 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_0609d820-e495-4845-a50f-166717c55f93/cinder-scheduler/0.log" Oct 02 02:48:18 crc kubenswrapper[4885]: I1002 02:48:18.128398 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_0609d820-e495-4845-a50f-166717c55f93/probe/0.log" Oct 02 02:48:18 crc kubenswrapper[4885]: I1002 02:48:18.323696 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-zlzq7_fd55e079-473a-43d3-aca2-ee0d91c06aca/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 02:48:18 crc kubenswrapper[4885]: I1002 02:48:18.951726 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-wvh9j_93d5499a-81ec-4dbc-b769-94d3e8690fbe/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 02:48:18 crc kubenswrapper[4885]: I1002 02:48:18.954412 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-cb6ffcf87-46bgc_58f621da-0934-482b-8f10-4273021207f0/init/0.log" Oct 02 02:48:18 crc kubenswrapper[4885]: I1002 02:48:18.989379 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-mv2tw_c6157c99-0cbf-4edb-9a16-04c99b7b2385/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 02:48:19 crc kubenswrapper[4885]: I1002 02:48:19.159726 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-cb6ffcf87-46bgc_58f621da-0934-482b-8f10-4273021207f0/dnsmasq-dns/0.log" Oct 02 02:48:19 crc kubenswrapper[4885]: I1002 02:48:19.366124 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-cb6ffcf87-46bgc_58f621da-0934-482b-8f10-4273021207f0/init/0.log" Oct 02 02:48:19 crc kubenswrapper[4885]: I1002 02:48:19.397911 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_d6e0fd15-e949-444a-86e2-894963c1267d/glance-httpd/0.log" Oct 02 02:48:19 crc kubenswrapper[4885]: I1002 02:48:19.478977 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-c2g9p_c1ca4902-a7b7-4cfa-a8df-5953754142e0/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 02:48:19 crc kubenswrapper[4885]: I1002 02:48:19.530851 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_d6e0fd15-e949-444a-86e2-894963c1267d/glance-log/0.log" Oct 02 02:48:19 crc kubenswrapper[4885]: I1002 02:48:19.665797 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_511f1564-a39d-48be-9c35-ed3343e4650c/glance-log/0.log" Oct 02 02:48:19 crc kubenswrapper[4885]: I1002 02:48:19.678106 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_511f1564-a39d-48be-9c35-ed3343e4650c/glance-httpd/0.log" Oct 02 02:48:20 crc kubenswrapper[4885]: I1002 02:48:20.173735 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-prq96_bd405392-0e1a-42c5-a81a-cab2d229ea55/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 02:48:20 crc kubenswrapper[4885]: I1002 02:48:20.460916 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-6d7c47744-vjcbs_607de89d-b68e-49e8-beb4-7664a37c6105/horizon/0.log" Oct 02 02:48:20 crc kubenswrapper[4885]: I1002 02:48:20.501404 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-p9p9g_48ef0e0a-ae52-4d11-9023-d53f4ab716ac/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 02:48:20 crc kubenswrapper[4885]: I1002 02:48:20.718437 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-6d7c47744-vjcbs_607de89d-b68e-49e8-beb4-7664a37c6105/horizon-log/0.log" Oct 02 02:48:20 crc kubenswrapper[4885]: I1002 02:48:20.920538 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_b5cb9a09-d2c8-4138-9eea-f404bd2e5034/kube-state-metrics/0.log" Oct 02 02:48:20 crc kubenswrapper[4885]: I1002 02:48:20.981860 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-59f7bf489d-7tfk4_5f5ca9c7-0cfd-4c95-a160-4d909f8f3d00/keystone-api/0.log" Oct 02 02:48:21 crc kubenswrapper[4885]: I1002 02:48:21.381291 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-x47dm_54cf89d0-a777-4ac9-968b-49836d0b13b1/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 02:48:21 crc kubenswrapper[4885]: I1002 02:48:21.442381 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-684bd7f7b9-p7lvf_ddb14564-1e3d-43d7-9c9d-2665b53124b6/neutron-httpd/0.log" Oct 02 02:48:21 crc kubenswrapper[4885]: I1002 02:48:21.498955 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-684bd7f7b9-p7lvf_ddb14564-1e3d-43d7-9c9d-2665b53124b6/neutron-api/0.log" Oct 02 02:48:21 crc kubenswrapper[4885]: I1002 02:48:21.932571 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-2hp4t_f6edca36-b630-4682-93ed-2c0c2dcadecb/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 02:48:22 crc kubenswrapper[4885]: I1002 02:48:22.471328 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_504d46bb-b842-4362-9a00-4917a01cf509/nova-cell0-conductor-conductor/0.log" Oct 02 02:48:22 crc kubenswrapper[4885]: I1002 02:48:22.574167 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_71959149-27b9-4b03-aa11-1c2164677d5c/nova-api-api/0.log" Oct 02 02:48:22 crc kubenswrapper[4885]: I1002 02:48:22.613912 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_71959149-27b9-4b03-aa11-1c2164677d5c/nova-api-log/0.log" Oct 02 02:48:22 crc kubenswrapper[4885]: I1002 02:48:22.940002 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_b6f37337-ce77-444d-9efa-852d0db3aabb/nova-cell1-conductor-conductor/0.log" Oct 02 02:48:23 crc kubenswrapper[4885]: I1002 02:48:23.011501 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_f6249fa5-32a1-4353-a521-67d1fde191d4/nova-cell1-novncproxy-novncproxy/0.log" Oct 02 02:48:23 crc kubenswrapper[4885]: I1002 02:48:23.363900 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-c6gw4_14bdd01b-529b-45b4-9ac3-188fdf3d3855/nova-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 02:48:23 crc kubenswrapper[4885]: I1002 02:48:23.370341 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_179284e1-8a2f-4987-a0b4-7b24b6e3cf94/nova-metadata-log/0.log" Oct 02 02:48:23 crc kubenswrapper[4885]: I1002 02:48:23.883427 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_50278f9c-b6fc-437c-ba3a-972b857493e3/nova-scheduler-scheduler/0.log" Oct 02 02:48:24 crc kubenswrapper[4885]: I1002 02:48:24.141189 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_87196135-8d40-47cb-b918-98aea04c7c17/mysql-bootstrap/0.log" Oct 02 02:48:24 crc kubenswrapper[4885]: I1002 02:48:24.302366 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_87196135-8d40-47cb-b918-98aea04c7c17/mysql-bootstrap/0.log" Oct 02 02:48:24 crc kubenswrapper[4885]: I1002 02:48:24.387868 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_87196135-8d40-47cb-b918-98aea04c7c17/galera/0.log" Oct 02 02:48:24 crc kubenswrapper[4885]: I1002 02:48:24.688839 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_179284e1-8a2f-4987-a0b4-7b24b6e3cf94/nova-metadata-metadata/0.log" Oct 02 02:48:24 crc kubenswrapper[4885]: I1002 02:48:24.726122 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_8427e298-965b-48f0-821f-bd078fa2e96f/mysql-bootstrap/0.log" Oct 02 02:48:24 crc kubenswrapper[4885]: I1002 02:48:24.993064 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_8427e298-965b-48f0-821f-bd078fa2e96f/galera/0.log" Oct 02 02:48:25 crc kubenswrapper[4885]: I1002 02:48:25.039182 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_8427e298-965b-48f0-821f-bd078fa2e96f/mysql-bootstrap/0.log" Oct 02 02:48:25 crc kubenswrapper[4885]: I1002 02:48:25.191084 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_ab251703-d0d6-49cf-bd2a-b74dee9e48f7/openstackclient/0.log" Oct 02 02:48:25 crc kubenswrapper[4885]: I1002 02:48:25.373098 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-fgvpt_924fb321-4073-449a-a546-811b046ed26a/ovn-controller/0.log" Oct 02 02:48:25 crc kubenswrapper[4885]: I1002 02:48:25.470299 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-sphmn_eb166198-03b5-458f-b86b-5706afa88d7b/openstack-network-exporter/0.log" Oct 02 02:48:25 crc kubenswrapper[4885]: I1002 02:48:25.786074 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-q92rb_b65df3c4-cc33-43da-9bf1-8f46a480f7eb/ovsdb-server-init/0.log" Oct 02 02:48:25 crc kubenswrapper[4885]: I1002 02:48:25.794295 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-q92rb_b65df3c4-cc33-43da-9bf1-8f46a480f7eb/ovs-vswitchd/0.log" Oct 02 02:48:25 crc kubenswrapper[4885]: I1002 02:48:25.848142 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-q92rb_b65df3c4-cc33-43da-9bf1-8f46a480f7eb/ovsdb-server-init/0.log" Oct 02 02:48:25 crc kubenswrapper[4885]: I1002 02:48:25.953604 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-q92rb_b65df3c4-cc33-43da-9bf1-8f46a480f7eb/ovsdb-server/0.log" Oct 02 02:48:26 crc kubenswrapper[4885]: I1002 02:48:26.213408 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_d02dcd41-585a-411f-ace9-97c1a662fdc8/openstack-network-exporter/0.log" Oct 02 02:48:26 crc kubenswrapper[4885]: I1002 02:48:26.269142 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-xkwk4_d8d7cf22-9bf7-417b-b4e2-d199cfccf949/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 02:48:26 crc kubenswrapper[4885]: I1002 02:48:26.420276 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_d02dcd41-585a-411f-ace9-97c1a662fdc8/ovn-northd/0.log" Oct 02 02:48:26 crc kubenswrapper[4885]: I1002 02:48:26.479553 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_079e125d-c938-4962-8f41-09b3109fb0ae/openstack-network-exporter/0.log" Oct 02 02:48:26 crc kubenswrapper[4885]: I1002 02:48:26.582869 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_079e125d-c938-4962-8f41-09b3109fb0ae/ovsdbserver-nb/0.log" Oct 02 02:48:26 crc kubenswrapper[4885]: I1002 02:48:26.712508 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_33286960-496c-4f14-a334-c6dc765ca956/openstack-network-exporter/0.log" Oct 02 02:48:26 crc kubenswrapper[4885]: I1002 02:48:26.775207 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_33286960-496c-4f14-a334-c6dc765ca956/ovsdbserver-sb/0.log" Oct 02 02:48:26 crc kubenswrapper[4885]: I1002 02:48:26.968745 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-5d65d69454-8bzbn_9bcc69ce-fb1a-4265-a6de-b2e2a10ac733/placement-api/0.log" Oct 02 02:48:27 crc kubenswrapper[4885]: I1002 02:48:27.079365 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-5d65d69454-8bzbn_9bcc69ce-fb1a-4265-a6de-b2e2a10ac733/placement-log/0.log" Oct 02 02:48:27 crc kubenswrapper[4885]: I1002 02:48:27.174859 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_79f97e3d-d825-4982-9f76-91089df7a67b/setup-container/0.log" Oct 02 02:48:27 crc kubenswrapper[4885]: I1002 02:48:27.407706 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_79f97e3d-d825-4982-9f76-91089df7a67b/setup-container/0.log" Oct 02 02:48:27 crc kubenswrapper[4885]: I1002 02:48:27.452860 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_79f97e3d-d825-4982-9f76-91089df7a67b/rabbitmq/0.log" Oct 02 02:48:27 crc kubenswrapper[4885]: I1002 02:48:27.597237 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_18642896-a2b6-412c-beb1-8010484744a4/setup-container/0.log" Oct 02 02:48:27 crc kubenswrapper[4885]: I1002 02:48:27.837139 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_18642896-a2b6-412c-beb1-8010484744a4/rabbitmq/0.log" Oct 02 02:48:27 crc kubenswrapper[4885]: I1002 02:48:27.847251 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_18642896-a2b6-412c-beb1-8010484744a4/setup-container/0.log" Oct 02 02:48:28 crc kubenswrapper[4885]: I1002 02:48:28.041531 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-2rt6h_2d16884c-61ae-4288-b629-2a7c3be23ea8/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 02:48:28 crc kubenswrapper[4885]: I1002 02:48:28.105019 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-6nvvk_379db8cf-874f-4d6e-9e3c-795457fdddb9/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 02:48:28 crc kubenswrapper[4885]: I1002 02:48:28.257473 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-vmf4j_6602e994-3938-456d-8bb3-e092b6350c8a/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 02:48:28 crc kubenswrapper[4885]: I1002 02:48:28.507079 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-jnpmf_3093cd67-4ea2-4256-bcf4-7e2533870ba6/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 02:48:28 crc kubenswrapper[4885]: I1002 02:48:28.621765 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-njjfc_e6dc3bce-d7cf-4ce8-9577-dd1f6813bff1/ssh-known-hosts-edpm-deployment/0.log" Oct 02 02:48:28 crc kubenswrapper[4885]: I1002 02:48:28.843427 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-7474b946b7-s2w7j_3b22bebc-2f6c-4981-8166-4ccaf19f9727/proxy-server/0.log" Oct 02 02:48:28 crc kubenswrapper[4885]: I1002 02:48:28.995010 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-7474b946b7-s2w7j_3b22bebc-2f6c-4981-8166-4ccaf19f9727/proxy-httpd/0.log" Oct 02 02:48:29 crc kubenswrapper[4885]: I1002 02:48:29.055741 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-9wd49_2d471d2b-912c-4f49-a7ec-c07b85c02bd9/swift-ring-rebalance/0.log" Oct 02 02:48:29 crc kubenswrapper[4885]: I1002 02:48:29.187857 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0c8d183a-f48b-4a69-8685-90c834e678ce/account-auditor/0.log" Oct 02 02:48:29 crc kubenswrapper[4885]: I1002 02:48:29.249360 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0c8d183a-f48b-4a69-8685-90c834e678ce/account-reaper/0.log" Oct 02 02:48:29 crc kubenswrapper[4885]: I1002 02:48:29.430133 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0c8d183a-f48b-4a69-8685-90c834e678ce/account-replicator/0.log" Oct 02 02:48:29 crc kubenswrapper[4885]: I1002 02:48:29.446561 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0c8d183a-f48b-4a69-8685-90c834e678ce/container-auditor/0.log" Oct 02 02:48:29 crc kubenswrapper[4885]: I1002 02:48:29.488653 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0c8d183a-f48b-4a69-8685-90c834e678ce/account-server/0.log" Oct 02 02:48:29 crc kubenswrapper[4885]: I1002 02:48:29.657688 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0c8d183a-f48b-4a69-8685-90c834e678ce/container-server/0.log" Oct 02 02:48:29 crc kubenswrapper[4885]: I1002 02:48:29.663471 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0c8d183a-f48b-4a69-8685-90c834e678ce/container-replicator/0.log" Oct 02 02:48:29 crc kubenswrapper[4885]: I1002 02:48:29.705037 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0c8d183a-f48b-4a69-8685-90c834e678ce/container-updater/0.log" Oct 02 02:48:29 crc kubenswrapper[4885]: I1002 02:48:29.895785 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0c8d183a-f48b-4a69-8685-90c834e678ce/object-expirer/0.log" Oct 02 02:48:29 crc kubenswrapper[4885]: I1002 02:48:29.909326 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0c8d183a-f48b-4a69-8685-90c834e678ce/object-auditor/0.log" Oct 02 02:48:29 crc kubenswrapper[4885]: I1002 02:48:29.910393 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0c8d183a-f48b-4a69-8685-90c834e678ce/object-replicator/0.log" Oct 02 02:48:30 crc kubenswrapper[4885]: I1002 02:48:30.046229 4885 scope.go:117] "RemoveContainer" containerID="999b3e75477511af3aa11f1e9d932c1ffb739b6f215cc19f45ac7144051ab65c" Oct 02 02:48:30 crc kubenswrapper[4885]: E1002 02:48:30.046476 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:48:30 crc kubenswrapper[4885]: I1002 02:48:30.088666 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0c8d183a-f48b-4a69-8685-90c834e678ce/object-server/0.log" Oct 02 02:48:30 crc kubenswrapper[4885]: I1002 02:48:30.089711 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0c8d183a-f48b-4a69-8685-90c834e678ce/object-updater/0.log" Oct 02 02:48:30 crc kubenswrapper[4885]: I1002 02:48:30.111768 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0c8d183a-f48b-4a69-8685-90c834e678ce/rsync/0.log" Oct 02 02:48:30 crc kubenswrapper[4885]: I1002 02:48:30.277989 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0c8d183a-f48b-4a69-8685-90c834e678ce/swift-recon-cron/0.log" Oct 02 02:48:30 crc kubenswrapper[4885]: I1002 02:48:30.410205 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-v9h2z_ce63c1c1-34cc-4cba-a553-33b66431bc3c/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 02:48:30 crc kubenswrapper[4885]: I1002 02:48:30.515993 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_7a715636-06d0-4508-a9d0-fe83f4c05d01/tempest-tests-tempest-tests-runner/0.log" Oct 02 02:48:30 crc kubenswrapper[4885]: I1002 02:48:30.726869 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_9cf08307-1f33-4989-b935-67c8f4a76cc4/test-operator-logs-container/0.log" Oct 02 02:48:30 crc kubenswrapper[4885]: I1002 02:48:30.905418 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-5l99v_829f7f0e-4288-42a2-b585-99beef8cf451/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 02:48:38 crc kubenswrapper[4885]: I1002 02:48:38.275598 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_d798d658-f60b-4e23-8753-b3de580e44b0/memcached/0.log" Oct 02 02:48:44 crc kubenswrapper[4885]: I1002 02:48:44.047402 4885 scope.go:117] "RemoveContainer" containerID="999b3e75477511af3aa11f1e9d932c1ffb739b6f215cc19f45ac7144051ab65c" Oct 02 02:48:44 crc kubenswrapper[4885]: E1002 02:48:44.049044 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:48:56 crc kubenswrapper[4885]: I1002 02:48:56.047707 4885 scope.go:117] "RemoveContainer" containerID="999b3e75477511af3aa11f1e9d932c1ffb739b6f215cc19f45ac7144051ab65c" Oct 02 02:48:56 crc kubenswrapper[4885]: E1002 02:48:56.048854 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:49:08 crc kubenswrapper[4885]: I1002 02:49:08.050241 4885 scope.go:117] "RemoveContainer" containerID="999b3e75477511af3aa11f1e9d932c1ffb739b6f215cc19f45ac7144051ab65c" Oct 02 02:49:08 crc kubenswrapper[4885]: E1002 02:49:08.051019 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:49:20 crc kubenswrapper[4885]: I1002 02:49:20.717037 4885 generic.go:334] "Generic (PLEG): container finished" podID="12ee7c20-678b-4f11-8d6e-96bd3a4a98f2" containerID="a663470245620171417c0f2c6b2c2043717fcae45eacf2e2b432c2eca245f639" exitCode=0 Oct 02 02:49:20 crc kubenswrapper[4885]: I1002 02:49:20.717138 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-28twm/crc-debug-mglll" event={"ID":"12ee7c20-678b-4f11-8d6e-96bd3a4a98f2","Type":"ContainerDied","Data":"a663470245620171417c0f2c6b2c2043717fcae45eacf2e2b432c2eca245f639"} Oct 02 02:49:21 crc kubenswrapper[4885]: I1002 02:49:21.859533 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-28twm/crc-debug-mglll" Oct 02 02:49:21 crc kubenswrapper[4885]: I1002 02:49:21.907016 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-28twm/crc-debug-mglll"] Oct 02 02:49:21 crc kubenswrapper[4885]: I1002 02:49:21.917287 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-28twm/crc-debug-mglll"] Oct 02 02:49:21 crc kubenswrapper[4885]: I1002 02:49:21.963662 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gkztp\" (UniqueName: \"kubernetes.io/projected/12ee7c20-678b-4f11-8d6e-96bd3a4a98f2-kube-api-access-gkztp\") pod \"12ee7c20-678b-4f11-8d6e-96bd3a4a98f2\" (UID: \"12ee7c20-678b-4f11-8d6e-96bd3a4a98f2\") " Oct 02 02:49:21 crc kubenswrapper[4885]: I1002 02:49:21.963957 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/12ee7c20-678b-4f11-8d6e-96bd3a4a98f2-host\") pod \"12ee7c20-678b-4f11-8d6e-96bd3a4a98f2\" (UID: \"12ee7c20-678b-4f11-8d6e-96bd3a4a98f2\") " Oct 02 02:49:21 crc kubenswrapper[4885]: I1002 02:49:21.964031 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/12ee7c20-678b-4f11-8d6e-96bd3a4a98f2-host" (OuterVolumeSpecName: "host") pod "12ee7c20-678b-4f11-8d6e-96bd3a4a98f2" (UID: "12ee7c20-678b-4f11-8d6e-96bd3a4a98f2"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 02:49:21 crc kubenswrapper[4885]: I1002 02:49:21.964555 4885 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/12ee7c20-678b-4f11-8d6e-96bd3a4a98f2-host\") on node \"crc\" DevicePath \"\"" Oct 02 02:49:21 crc kubenswrapper[4885]: I1002 02:49:21.979628 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/12ee7c20-678b-4f11-8d6e-96bd3a4a98f2-kube-api-access-gkztp" (OuterVolumeSpecName: "kube-api-access-gkztp") pod "12ee7c20-678b-4f11-8d6e-96bd3a4a98f2" (UID: "12ee7c20-678b-4f11-8d6e-96bd3a4a98f2"). InnerVolumeSpecName "kube-api-access-gkztp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:49:22 crc kubenswrapper[4885]: I1002 02:49:22.065720 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gkztp\" (UniqueName: \"kubernetes.io/projected/12ee7c20-678b-4f11-8d6e-96bd3a4a98f2-kube-api-access-gkztp\") on node \"crc\" DevicePath \"\"" Oct 02 02:49:22 crc kubenswrapper[4885]: I1002 02:49:22.067987 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="12ee7c20-678b-4f11-8d6e-96bd3a4a98f2" path="/var/lib/kubelet/pods/12ee7c20-678b-4f11-8d6e-96bd3a4a98f2/volumes" Oct 02 02:49:22 crc kubenswrapper[4885]: I1002 02:49:22.737992 4885 scope.go:117] "RemoveContainer" containerID="a663470245620171417c0f2c6b2c2043717fcae45eacf2e2b432c2eca245f639" Oct 02 02:49:22 crc kubenswrapper[4885]: I1002 02:49:22.738056 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-28twm/crc-debug-mglll" Oct 02 02:49:23 crc kubenswrapper[4885]: I1002 02:49:23.047027 4885 scope.go:117] "RemoveContainer" containerID="999b3e75477511af3aa11f1e9d932c1ffb739b6f215cc19f45ac7144051ab65c" Oct 02 02:49:23 crc kubenswrapper[4885]: E1002 02:49:23.048044 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:49:23 crc kubenswrapper[4885]: I1002 02:49:23.787141 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-28twm/crc-debug-vm9wp"] Oct 02 02:49:23 crc kubenswrapper[4885]: E1002 02:49:23.788937 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12ee7c20-678b-4f11-8d6e-96bd3a4a98f2" containerName="container-00" Oct 02 02:49:23 crc kubenswrapper[4885]: I1002 02:49:23.788956 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="12ee7c20-678b-4f11-8d6e-96bd3a4a98f2" containerName="container-00" Oct 02 02:49:23 crc kubenswrapper[4885]: I1002 02:49:23.789517 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="12ee7c20-678b-4f11-8d6e-96bd3a4a98f2" containerName="container-00" Oct 02 02:49:23 crc kubenswrapper[4885]: I1002 02:49:23.790516 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-28twm/crc-debug-vm9wp" Oct 02 02:49:23 crc kubenswrapper[4885]: I1002 02:49:23.923728 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fb09afe1-6de6-4aa0-8410-a5ebe7328924-host\") pod \"crc-debug-vm9wp\" (UID: \"fb09afe1-6de6-4aa0-8410-a5ebe7328924\") " pod="openshift-must-gather-28twm/crc-debug-vm9wp" Oct 02 02:49:23 crc kubenswrapper[4885]: I1002 02:49:23.923788 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-92gtv\" (UniqueName: \"kubernetes.io/projected/fb09afe1-6de6-4aa0-8410-a5ebe7328924-kube-api-access-92gtv\") pod \"crc-debug-vm9wp\" (UID: \"fb09afe1-6de6-4aa0-8410-a5ebe7328924\") " pod="openshift-must-gather-28twm/crc-debug-vm9wp" Oct 02 02:49:24 crc kubenswrapper[4885]: I1002 02:49:24.024440 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fb09afe1-6de6-4aa0-8410-a5ebe7328924-host\") pod \"crc-debug-vm9wp\" (UID: \"fb09afe1-6de6-4aa0-8410-a5ebe7328924\") " pod="openshift-must-gather-28twm/crc-debug-vm9wp" Oct 02 02:49:24 crc kubenswrapper[4885]: I1002 02:49:24.024706 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-92gtv\" (UniqueName: \"kubernetes.io/projected/fb09afe1-6de6-4aa0-8410-a5ebe7328924-kube-api-access-92gtv\") pod \"crc-debug-vm9wp\" (UID: \"fb09afe1-6de6-4aa0-8410-a5ebe7328924\") " pod="openshift-must-gather-28twm/crc-debug-vm9wp" Oct 02 02:49:24 crc kubenswrapper[4885]: I1002 02:49:24.024602 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fb09afe1-6de6-4aa0-8410-a5ebe7328924-host\") pod \"crc-debug-vm9wp\" (UID: \"fb09afe1-6de6-4aa0-8410-a5ebe7328924\") " pod="openshift-must-gather-28twm/crc-debug-vm9wp" Oct 02 02:49:24 crc kubenswrapper[4885]: I1002 02:49:24.066887 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-92gtv\" (UniqueName: \"kubernetes.io/projected/fb09afe1-6de6-4aa0-8410-a5ebe7328924-kube-api-access-92gtv\") pod \"crc-debug-vm9wp\" (UID: \"fb09afe1-6de6-4aa0-8410-a5ebe7328924\") " pod="openshift-must-gather-28twm/crc-debug-vm9wp" Oct 02 02:49:24 crc kubenswrapper[4885]: I1002 02:49:24.113509 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-28twm/crc-debug-vm9wp" Oct 02 02:49:24 crc kubenswrapper[4885]: I1002 02:49:24.755580 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-28twm/crc-debug-vm9wp" event={"ID":"fb09afe1-6de6-4aa0-8410-a5ebe7328924","Type":"ContainerStarted","Data":"46552ac844213a16ef1a7a1cee33dfab4092e42f05442399558109f44192ac62"} Oct 02 02:49:24 crc kubenswrapper[4885]: I1002 02:49:24.755822 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-28twm/crc-debug-vm9wp" event={"ID":"fb09afe1-6de6-4aa0-8410-a5ebe7328924","Type":"ContainerStarted","Data":"02dbaa42f20867ea842d0ceaa6acd232ff18523ce488bb6e0fd30ee6c516139b"} Oct 02 02:49:24 crc kubenswrapper[4885]: I1002 02:49:24.778138 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-28twm/crc-debug-vm9wp" podStartSLOduration=1.778110098 podStartE2EDuration="1.778110098s" podCreationTimestamp="2025-10-02 02:49:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:49:24.766145923 +0000 UTC m=+3753.577893372" watchObservedRunningTime="2025-10-02 02:49:24.778110098 +0000 UTC m=+3753.589857537" Oct 02 02:49:25 crc kubenswrapper[4885]: I1002 02:49:25.768841 4885 generic.go:334] "Generic (PLEG): container finished" podID="fb09afe1-6de6-4aa0-8410-a5ebe7328924" containerID="46552ac844213a16ef1a7a1cee33dfab4092e42f05442399558109f44192ac62" exitCode=0 Oct 02 02:49:25 crc kubenswrapper[4885]: I1002 02:49:25.769331 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-28twm/crc-debug-vm9wp" event={"ID":"fb09afe1-6de6-4aa0-8410-a5ebe7328924","Type":"ContainerDied","Data":"46552ac844213a16ef1a7a1cee33dfab4092e42f05442399558109f44192ac62"} Oct 02 02:49:26 crc kubenswrapper[4885]: I1002 02:49:26.907942 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-28twm/crc-debug-vm9wp" Oct 02 02:49:27 crc kubenswrapper[4885]: I1002 02:49:27.068056 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-92gtv\" (UniqueName: \"kubernetes.io/projected/fb09afe1-6de6-4aa0-8410-a5ebe7328924-kube-api-access-92gtv\") pod \"fb09afe1-6de6-4aa0-8410-a5ebe7328924\" (UID: \"fb09afe1-6de6-4aa0-8410-a5ebe7328924\") " Oct 02 02:49:27 crc kubenswrapper[4885]: I1002 02:49:27.068396 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fb09afe1-6de6-4aa0-8410-a5ebe7328924-host\") pod \"fb09afe1-6de6-4aa0-8410-a5ebe7328924\" (UID: \"fb09afe1-6de6-4aa0-8410-a5ebe7328924\") " Oct 02 02:49:27 crc kubenswrapper[4885]: I1002 02:49:27.069034 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fb09afe1-6de6-4aa0-8410-a5ebe7328924-host" (OuterVolumeSpecName: "host") pod "fb09afe1-6de6-4aa0-8410-a5ebe7328924" (UID: "fb09afe1-6de6-4aa0-8410-a5ebe7328924"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 02:49:27 crc kubenswrapper[4885]: I1002 02:49:27.075949 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fb09afe1-6de6-4aa0-8410-a5ebe7328924-kube-api-access-92gtv" (OuterVolumeSpecName: "kube-api-access-92gtv") pod "fb09afe1-6de6-4aa0-8410-a5ebe7328924" (UID: "fb09afe1-6de6-4aa0-8410-a5ebe7328924"). InnerVolumeSpecName "kube-api-access-92gtv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:49:27 crc kubenswrapper[4885]: I1002 02:49:27.171197 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-92gtv\" (UniqueName: \"kubernetes.io/projected/fb09afe1-6de6-4aa0-8410-a5ebe7328924-kube-api-access-92gtv\") on node \"crc\" DevicePath \"\"" Oct 02 02:49:27 crc kubenswrapper[4885]: I1002 02:49:27.171225 4885 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fb09afe1-6de6-4aa0-8410-a5ebe7328924-host\") on node \"crc\" DevicePath \"\"" Oct 02 02:49:27 crc kubenswrapper[4885]: I1002 02:49:27.794822 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-28twm/crc-debug-vm9wp" event={"ID":"fb09afe1-6de6-4aa0-8410-a5ebe7328924","Type":"ContainerDied","Data":"02dbaa42f20867ea842d0ceaa6acd232ff18523ce488bb6e0fd30ee6c516139b"} Oct 02 02:49:27 crc kubenswrapper[4885]: I1002 02:49:27.794860 4885 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="02dbaa42f20867ea842d0ceaa6acd232ff18523ce488bb6e0fd30ee6c516139b" Oct 02 02:49:27 crc kubenswrapper[4885]: I1002 02:49:27.794887 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-28twm/crc-debug-vm9wp" Oct 02 02:49:31 crc kubenswrapper[4885]: I1002 02:49:31.348782 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-28twm/crc-debug-vm9wp"] Oct 02 02:49:31 crc kubenswrapper[4885]: I1002 02:49:31.357135 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-28twm/crc-debug-vm9wp"] Oct 02 02:49:32 crc kubenswrapper[4885]: I1002 02:49:32.061435 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fb09afe1-6de6-4aa0-8410-a5ebe7328924" path="/var/lib/kubelet/pods/fb09afe1-6de6-4aa0-8410-a5ebe7328924/volumes" Oct 02 02:49:32 crc kubenswrapper[4885]: I1002 02:49:32.586293 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-28twm/crc-debug-c6vcb"] Oct 02 02:49:32 crc kubenswrapper[4885]: E1002 02:49:32.586969 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb09afe1-6de6-4aa0-8410-a5ebe7328924" containerName="container-00" Oct 02 02:49:32 crc kubenswrapper[4885]: I1002 02:49:32.586983 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb09afe1-6de6-4aa0-8410-a5ebe7328924" containerName="container-00" Oct 02 02:49:32 crc kubenswrapper[4885]: I1002 02:49:32.587220 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb09afe1-6de6-4aa0-8410-a5ebe7328924" containerName="container-00" Oct 02 02:49:32 crc kubenswrapper[4885]: I1002 02:49:32.587971 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-28twm/crc-debug-c6vcb" Oct 02 02:49:32 crc kubenswrapper[4885]: I1002 02:49:32.669963 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4dwgd\" (UniqueName: \"kubernetes.io/projected/1eb5a090-12a5-4f21-a5c8-f23bcd3a34f7-kube-api-access-4dwgd\") pod \"crc-debug-c6vcb\" (UID: \"1eb5a090-12a5-4f21-a5c8-f23bcd3a34f7\") " pod="openshift-must-gather-28twm/crc-debug-c6vcb" Oct 02 02:49:32 crc kubenswrapper[4885]: I1002 02:49:32.670313 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1eb5a090-12a5-4f21-a5c8-f23bcd3a34f7-host\") pod \"crc-debug-c6vcb\" (UID: \"1eb5a090-12a5-4f21-a5c8-f23bcd3a34f7\") " pod="openshift-must-gather-28twm/crc-debug-c6vcb" Oct 02 02:49:32 crc kubenswrapper[4885]: I1002 02:49:32.772056 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4dwgd\" (UniqueName: \"kubernetes.io/projected/1eb5a090-12a5-4f21-a5c8-f23bcd3a34f7-kube-api-access-4dwgd\") pod \"crc-debug-c6vcb\" (UID: \"1eb5a090-12a5-4f21-a5c8-f23bcd3a34f7\") " pod="openshift-must-gather-28twm/crc-debug-c6vcb" Oct 02 02:49:32 crc kubenswrapper[4885]: I1002 02:49:32.772400 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1eb5a090-12a5-4f21-a5c8-f23bcd3a34f7-host\") pod \"crc-debug-c6vcb\" (UID: \"1eb5a090-12a5-4f21-a5c8-f23bcd3a34f7\") " pod="openshift-must-gather-28twm/crc-debug-c6vcb" Oct 02 02:49:32 crc kubenswrapper[4885]: I1002 02:49:32.772519 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1eb5a090-12a5-4f21-a5c8-f23bcd3a34f7-host\") pod \"crc-debug-c6vcb\" (UID: \"1eb5a090-12a5-4f21-a5c8-f23bcd3a34f7\") " pod="openshift-must-gather-28twm/crc-debug-c6vcb" Oct 02 02:49:32 crc kubenswrapper[4885]: I1002 02:49:32.817849 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4dwgd\" (UniqueName: \"kubernetes.io/projected/1eb5a090-12a5-4f21-a5c8-f23bcd3a34f7-kube-api-access-4dwgd\") pod \"crc-debug-c6vcb\" (UID: \"1eb5a090-12a5-4f21-a5c8-f23bcd3a34f7\") " pod="openshift-must-gather-28twm/crc-debug-c6vcb" Oct 02 02:49:32 crc kubenswrapper[4885]: I1002 02:49:32.954303 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-28twm/crc-debug-c6vcb" Oct 02 02:49:33 crc kubenswrapper[4885]: I1002 02:49:33.853795 4885 generic.go:334] "Generic (PLEG): container finished" podID="1eb5a090-12a5-4f21-a5c8-f23bcd3a34f7" containerID="8cd0b89680e0d29daf95cdbaf2dd8d1956f30dd988ed8c0e0443405b46e95f8c" exitCode=0 Oct 02 02:49:33 crc kubenswrapper[4885]: I1002 02:49:33.853951 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-28twm/crc-debug-c6vcb" event={"ID":"1eb5a090-12a5-4f21-a5c8-f23bcd3a34f7","Type":"ContainerDied","Data":"8cd0b89680e0d29daf95cdbaf2dd8d1956f30dd988ed8c0e0443405b46e95f8c"} Oct 02 02:49:33 crc kubenswrapper[4885]: I1002 02:49:33.854219 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-28twm/crc-debug-c6vcb" event={"ID":"1eb5a090-12a5-4f21-a5c8-f23bcd3a34f7","Type":"ContainerStarted","Data":"5ed9a2b66ec98c02d7bbf4e9f576fe603d408bbf155010da1ab46d2f17d98260"} Oct 02 02:49:33 crc kubenswrapper[4885]: I1002 02:49:33.905469 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-28twm/crc-debug-c6vcb"] Oct 02 02:49:33 crc kubenswrapper[4885]: I1002 02:49:33.913876 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-28twm/crc-debug-c6vcb"] Oct 02 02:49:34 crc kubenswrapper[4885]: I1002 02:49:34.966017 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-28twm/crc-debug-c6vcb" Oct 02 02:49:35 crc kubenswrapper[4885]: I1002 02:49:35.023087 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4dwgd\" (UniqueName: \"kubernetes.io/projected/1eb5a090-12a5-4f21-a5c8-f23bcd3a34f7-kube-api-access-4dwgd\") pod \"1eb5a090-12a5-4f21-a5c8-f23bcd3a34f7\" (UID: \"1eb5a090-12a5-4f21-a5c8-f23bcd3a34f7\") " Oct 02 02:49:35 crc kubenswrapper[4885]: I1002 02:49:35.023228 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1eb5a090-12a5-4f21-a5c8-f23bcd3a34f7-host\") pod \"1eb5a090-12a5-4f21-a5c8-f23bcd3a34f7\" (UID: \"1eb5a090-12a5-4f21-a5c8-f23bcd3a34f7\") " Oct 02 02:49:35 crc kubenswrapper[4885]: I1002 02:49:35.023392 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1eb5a090-12a5-4f21-a5c8-f23bcd3a34f7-host" (OuterVolumeSpecName: "host") pod "1eb5a090-12a5-4f21-a5c8-f23bcd3a34f7" (UID: "1eb5a090-12a5-4f21-a5c8-f23bcd3a34f7"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 02:49:35 crc kubenswrapper[4885]: I1002 02:49:35.023914 4885 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1eb5a090-12a5-4f21-a5c8-f23bcd3a34f7-host\") on node \"crc\" DevicePath \"\"" Oct 02 02:49:35 crc kubenswrapper[4885]: I1002 02:49:35.028342 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1eb5a090-12a5-4f21-a5c8-f23bcd3a34f7-kube-api-access-4dwgd" (OuterVolumeSpecName: "kube-api-access-4dwgd") pod "1eb5a090-12a5-4f21-a5c8-f23bcd3a34f7" (UID: "1eb5a090-12a5-4f21-a5c8-f23bcd3a34f7"). InnerVolumeSpecName "kube-api-access-4dwgd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:49:35 crc kubenswrapper[4885]: I1002 02:49:35.125279 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4dwgd\" (UniqueName: \"kubernetes.io/projected/1eb5a090-12a5-4f21-a5c8-f23bcd3a34f7-kube-api-access-4dwgd\") on node \"crc\" DevicePath \"\"" Oct 02 02:49:35 crc kubenswrapper[4885]: I1002 02:49:35.411003 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-6ff8b75857-wnxfk_8113ab96-dde5-4f9f-b922-687e7a17afa7/kube-rbac-proxy/0.log" Oct 02 02:49:35 crc kubenswrapper[4885]: I1002 02:49:35.484528 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-6ff8b75857-wnxfk_8113ab96-dde5-4f9f-b922-687e7a17afa7/manager/0.log" Oct 02 02:49:35 crc kubenswrapper[4885]: I1002 02:49:35.593755 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cbbec88b66c61e3d10a0bbc868e336160e79c262c68f9747aec27f06598t8b5_5b1a29e8-0ed9-4c6a-bcc5-65849560bd7a/util/0.log" Oct 02 02:49:35 crc kubenswrapper[4885]: I1002 02:49:35.759474 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cbbec88b66c61e3d10a0bbc868e336160e79c262c68f9747aec27f06598t8b5_5b1a29e8-0ed9-4c6a-bcc5-65849560bd7a/pull/0.log" Oct 02 02:49:35 crc kubenswrapper[4885]: I1002 02:49:35.760223 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cbbec88b66c61e3d10a0bbc868e336160e79c262c68f9747aec27f06598t8b5_5b1a29e8-0ed9-4c6a-bcc5-65849560bd7a/util/0.log" Oct 02 02:49:35 crc kubenswrapper[4885]: I1002 02:49:35.763681 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cbbec88b66c61e3d10a0bbc868e336160e79c262c68f9747aec27f06598t8b5_5b1a29e8-0ed9-4c6a-bcc5-65849560bd7a/pull/0.log" Oct 02 02:49:35 crc kubenswrapper[4885]: I1002 02:49:35.889050 4885 scope.go:117] "RemoveContainer" containerID="8cd0b89680e0d29daf95cdbaf2dd8d1956f30dd988ed8c0e0443405b46e95f8c" Oct 02 02:49:35 crc kubenswrapper[4885]: I1002 02:49:35.889086 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-28twm/crc-debug-c6vcb" Oct 02 02:49:35 crc kubenswrapper[4885]: I1002 02:49:35.948726 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cbbec88b66c61e3d10a0bbc868e336160e79c262c68f9747aec27f06598t8b5_5b1a29e8-0ed9-4c6a-bcc5-65849560bd7a/pull/0.log" Oct 02 02:49:35 crc kubenswrapper[4885]: I1002 02:49:35.954497 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cbbec88b66c61e3d10a0bbc868e336160e79c262c68f9747aec27f06598t8b5_5b1a29e8-0ed9-4c6a-bcc5-65849560bd7a/util/0.log" Oct 02 02:49:35 crc kubenswrapper[4885]: I1002 02:49:35.974151 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cbbec88b66c61e3d10a0bbc868e336160e79c262c68f9747aec27f06598t8b5_5b1a29e8-0ed9-4c6a-bcc5-65849560bd7a/extract/0.log" Oct 02 02:49:36 crc kubenswrapper[4885]: I1002 02:49:36.047279 4885 scope.go:117] "RemoveContainer" containerID="999b3e75477511af3aa11f1e9d932c1ffb739b6f215cc19f45ac7144051ab65c" Oct 02 02:49:36 crc kubenswrapper[4885]: E1002 02:49:36.047524 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:49:36 crc kubenswrapper[4885]: I1002 02:49:36.086459 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1eb5a090-12a5-4f21-a5c8-f23bcd3a34f7" path="/var/lib/kubelet/pods/1eb5a090-12a5-4f21-a5c8-f23bcd3a34f7/volumes" Oct 02 02:49:36 crc kubenswrapper[4885]: I1002 02:49:36.116798 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-644bddb6d8-kw92z_e5874982-d78d-4b7c-bf0a-19a7572dddad/kube-rbac-proxy/0.log" Oct 02 02:49:36 crc kubenswrapper[4885]: I1002 02:49:36.144513 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-644bddb6d8-kw92z_e5874982-d78d-4b7c-bf0a-19a7572dddad/manager/0.log" Oct 02 02:49:36 crc kubenswrapper[4885]: I1002 02:49:36.206281 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-84f4f7b77b-mw6h7_cc08d74a-adca-4ae7-98c6-16c7037bc0ca/kube-rbac-proxy/0.log" Oct 02 02:49:36 crc kubenswrapper[4885]: I1002 02:49:36.295622 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-84f4f7b77b-mw6h7_cc08d74a-adca-4ae7-98c6-16c7037bc0ca/manager/0.log" Oct 02 02:49:36 crc kubenswrapper[4885]: I1002 02:49:36.344585 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-84958c4d49-2dsv9_ff10db28-28ca-4f65-8c24-2ed813d7cc5c/kube-rbac-proxy/0.log" Oct 02 02:49:36 crc kubenswrapper[4885]: I1002 02:49:36.432778 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-84958c4d49-2dsv9_ff10db28-28ca-4f65-8c24-2ed813d7cc5c/manager/0.log" Oct 02 02:49:36 crc kubenswrapper[4885]: I1002 02:49:36.528984 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5d889d78cf-862d6_78117b64-3b3a-4892-90de-0dea12fa2602/manager/0.log" Oct 02 02:49:36 crc kubenswrapper[4885]: I1002 02:49:36.547435 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5d889d78cf-862d6_78117b64-3b3a-4892-90de-0dea12fa2602/kube-rbac-proxy/0.log" Oct 02 02:49:36 crc kubenswrapper[4885]: I1002 02:49:36.694571 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-9f4696d94-pv94t_1e64a13b-65af-4460-80a8-4a4659edd8aa/kube-rbac-proxy/0.log" Oct 02 02:49:36 crc kubenswrapper[4885]: I1002 02:49:36.719032 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-9f4696d94-pv94t_1e64a13b-65af-4460-80a8-4a4659edd8aa/manager/0.log" Oct 02 02:49:36 crc kubenswrapper[4885]: I1002 02:49:36.813010 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-9d6c5db85-9fjl5_45e282ed-8195-40dc-86cf-1177f3ebeb65/kube-rbac-proxy/0.log" Oct 02 02:49:36 crc kubenswrapper[4885]: I1002 02:49:36.931232 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-5cd4858477-mjtzp_bbd60151-4081-46a8-85c0-0783604392a1/kube-rbac-proxy/0.log" Oct 02 02:49:36 crc kubenswrapper[4885]: I1002 02:49:36.991190 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-5cd4858477-mjtzp_bbd60151-4081-46a8-85c0-0783604392a1/manager/0.log" Oct 02 02:49:36 crc kubenswrapper[4885]: I1002 02:49:36.996203 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-9d6c5db85-9fjl5_45e282ed-8195-40dc-86cf-1177f3ebeb65/manager/0.log" Oct 02 02:49:37 crc kubenswrapper[4885]: I1002 02:49:37.109697 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-5bd55b4bff-pww46_d2263844-1435-4352-b773-04e8d10f35a0/kube-rbac-proxy/0.log" Oct 02 02:49:37 crc kubenswrapper[4885]: I1002 02:49:37.238677 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-5bd55b4bff-pww46_d2263844-1435-4352-b773-04e8d10f35a0/manager/0.log" Oct 02 02:49:37 crc kubenswrapper[4885]: I1002 02:49:37.253295 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-6d68dbc695-9zf8p_bcb6c44c-c620-465c-8db3-6d770deda522/kube-rbac-proxy/0.log" Oct 02 02:49:37 crc kubenswrapper[4885]: I1002 02:49:37.305666 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-6d68dbc695-9zf8p_bcb6c44c-c620-465c-8db3-6d770deda522/manager/0.log" Oct 02 02:49:37 crc kubenswrapper[4885]: I1002 02:49:37.398160 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-88c7-nz599_0f1b84a4-59d6-45fe-9c2f-0bdff82fde6e/kube-rbac-proxy/0.log" Oct 02 02:49:37 crc kubenswrapper[4885]: I1002 02:49:37.437910 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-88c7-nz599_0f1b84a4-59d6-45fe-9c2f-0bdff82fde6e/manager/0.log" Oct 02 02:49:37 crc kubenswrapper[4885]: I1002 02:49:37.535789 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-849d5b9b84-clsdk_ae73c7b6-6b7e-4ca9-8a4e-0694ab34d9a9/kube-rbac-proxy/0.log" Oct 02 02:49:37 crc kubenswrapper[4885]: I1002 02:49:37.622305 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-849d5b9b84-clsdk_ae73c7b6-6b7e-4ca9-8a4e-0694ab34d9a9/manager/0.log" Oct 02 02:49:37 crc kubenswrapper[4885]: I1002 02:49:37.656435 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-64cd67b5cb-hwrj4_e6673002-4d33-4c32-8eb4-f1727339b76d/kube-rbac-proxy/0.log" Oct 02 02:49:37 crc kubenswrapper[4885]: I1002 02:49:37.785483 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-64cd67b5cb-hwrj4_e6673002-4d33-4c32-8eb4-f1727339b76d/manager/0.log" Oct 02 02:49:37 crc kubenswrapper[4885]: I1002 02:49:37.868967 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-7b787867f4-kcpkv_9d724768-b5a0-4a55-951a-d350127061cb/manager/0.log" Oct 02 02:49:37 crc kubenswrapper[4885]: I1002 02:49:37.889441 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-7b787867f4-kcpkv_9d724768-b5a0-4a55-951a-d350127061cb/kube-rbac-proxy/0.log" Oct 02 02:49:38 crc kubenswrapper[4885]: I1002 02:49:38.009986 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-5869cb545-wgfjr_dc61bc9b-923b-41cb-b724-62c72caf0586/kube-rbac-proxy/0.log" Oct 02 02:49:38 crc kubenswrapper[4885]: I1002 02:49:38.064745 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-5869cb545-wgfjr_dc61bc9b-923b-41cb-b724-62c72caf0586/manager/0.log" Oct 02 02:49:38 crc kubenswrapper[4885]: I1002 02:49:38.158159 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-557f5d867b-ndfjx_8b18e6b1-82f2-46cd-99c8-85af816df20f/kube-rbac-proxy/0.log" Oct 02 02:49:38 crc kubenswrapper[4885]: I1002 02:49:38.308422 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-6bcc7488c-bc4gx_02e27ea3-228f-4f78-9135-deaff888afd2/kube-rbac-proxy/0.log" Oct 02 02:49:38 crc kubenswrapper[4885]: I1002 02:49:38.509498 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-6bcc7488c-bc4gx_02e27ea3-228f-4f78-9135-deaff888afd2/operator/0.log" Oct 02 02:49:38 crc kubenswrapper[4885]: I1002 02:49:38.546276 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-zgthg_12809968-8a60-4ffa-9506-0c89ce063382/registry-server/0.log" Oct 02 02:49:38 crc kubenswrapper[4885]: I1002 02:49:38.618441 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-9976ff44c-gl72q_c540bfce-7653-4976-8d6e-36b479c3c711/kube-rbac-proxy/0.log" Oct 02 02:49:38 crc kubenswrapper[4885]: I1002 02:49:38.772897 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-589c58c6c-r65hz_6e740063-fb49-4847-ba6d-77b6c089ba50/kube-rbac-proxy/0.log" Oct 02 02:49:38 crc kubenswrapper[4885]: I1002 02:49:38.786382 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-9976ff44c-gl72q_c540bfce-7653-4976-8d6e-36b479c3c711/manager/0.log" Oct 02 02:49:38 crc kubenswrapper[4885]: I1002 02:49:38.882996 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-589c58c6c-r65hz_6e740063-fb49-4847-ba6d-77b6c089ba50/manager/0.log" Oct 02 02:49:38 crc kubenswrapper[4885]: I1002 02:49:38.959677 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-5f97d8c699-bngfl_d2768afc-17ee-4141-aa68-7366faaa0145/operator/0.log" Oct 02 02:49:39 crc kubenswrapper[4885]: I1002 02:49:39.081728 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-84d6b4b759-6f74g_f80d2e7d-10a3-492f-a045-b48126a02490/kube-rbac-proxy/0.log" Oct 02 02:49:39 crc kubenswrapper[4885]: I1002 02:49:39.156875 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-84d6b4b759-6f74g_f80d2e7d-10a3-492f-a045-b48126a02490/manager/0.log" Oct 02 02:49:39 crc kubenswrapper[4885]: I1002 02:49:39.245168 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-b8d54b5d7-26g2l_dd09a3bb-1217-4a3a-822b-2b2f04171271/kube-rbac-proxy/0.log" Oct 02 02:49:39 crc kubenswrapper[4885]: I1002 02:49:39.329227 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-557f5d867b-ndfjx_8b18e6b1-82f2-46cd-99c8-85af816df20f/manager/0.log" Oct 02 02:49:39 crc kubenswrapper[4885]: I1002 02:49:39.364791 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-b8d54b5d7-26g2l_dd09a3bb-1217-4a3a-822b-2b2f04171271/manager/0.log" Oct 02 02:49:39 crc kubenswrapper[4885]: I1002 02:49:39.438431 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-85777745bb-94284_74a07e3e-5061-43f9-b239-eb480bade999/manager/0.log" Oct 02 02:49:39 crc kubenswrapper[4885]: I1002 02:49:39.457923 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-85777745bb-94284_74a07e3e-5061-43f9-b239-eb480bade999/kube-rbac-proxy/0.log" Oct 02 02:49:39 crc kubenswrapper[4885]: I1002 02:49:39.547616 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-6b9957f54f-mb6mf_c5dd164b-699d-4f14-85d0-91f138a42c95/kube-rbac-proxy/0.log" Oct 02 02:49:39 crc kubenswrapper[4885]: I1002 02:49:39.562177 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-6b9957f54f-mb6mf_c5dd164b-699d-4f14-85d0-91f138a42c95/manager/0.log" Oct 02 02:49:49 crc kubenswrapper[4885]: I1002 02:49:49.046763 4885 scope.go:117] "RemoveContainer" containerID="999b3e75477511af3aa11f1e9d932c1ffb739b6f215cc19f45ac7144051ab65c" Oct 02 02:49:49 crc kubenswrapper[4885]: E1002 02:49:49.048539 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:49:56 crc kubenswrapper[4885]: I1002 02:49:56.637176 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-dpqkj_cc6bb484-0230-46d9-9507-b641d2a5f330/control-plane-machine-set-operator/0.log" Oct 02 02:49:56 crc kubenswrapper[4885]: I1002 02:49:56.756880 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-5wldz_76fb6dce-01dd-4aef-8213-e03be7a5e055/kube-rbac-proxy/0.log" Oct 02 02:49:56 crc kubenswrapper[4885]: I1002 02:49:56.779577 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-5wldz_76fb6dce-01dd-4aef-8213-e03be7a5e055/machine-api-operator/0.log" Oct 02 02:50:01 crc kubenswrapper[4885]: I1002 02:50:01.047247 4885 scope.go:117] "RemoveContainer" containerID="999b3e75477511af3aa11f1e9d932c1ffb739b6f215cc19f45ac7144051ab65c" Oct 02 02:50:01 crc kubenswrapper[4885]: E1002 02:50:01.047893 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:50:10 crc kubenswrapper[4885]: I1002 02:50:10.678177 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-pj866_0b6eafa9-42ce-4e5b-9ded-180a669d26e2/cert-manager-controller/0.log" Oct 02 02:50:10 crc kubenswrapper[4885]: I1002 02:50:10.729255 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-rbccc_e50cfe49-c47f-4f2a-9799-0c79123fac70/cert-manager-cainjector/0.log" Oct 02 02:50:10 crc kubenswrapper[4885]: I1002 02:50:10.811219 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-vh2n2_18c66206-ad33-47d2-a37a-b3b50c2bf984/cert-manager-webhook/0.log" Oct 02 02:50:14 crc kubenswrapper[4885]: I1002 02:50:14.046565 4885 scope.go:117] "RemoveContainer" containerID="999b3e75477511af3aa11f1e9d932c1ffb739b6f215cc19f45ac7144051ab65c" Oct 02 02:50:14 crc kubenswrapper[4885]: E1002 02:50:14.047513 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:50:23 crc kubenswrapper[4885]: I1002 02:50:23.560895 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-6b874cbd85-dvlk4_c7f8ff02-de9d-456d-ba5e-a794f568c5cf/nmstate-console-plugin/0.log" Oct 02 02:50:23 crc kubenswrapper[4885]: I1002 02:50:23.660582 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-4xzpg_4376783f-66f6-41a4-a42e-2666f1cc9978/nmstate-handler/0.log" Oct 02 02:50:23 crc kubenswrapper[4885]: I1002 02:50:23.707529 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-sr7mv_c2a4785b-6378-4f85-ae1d-46e27903af36/kube-rbac-proxy/0.log" Oct 02 02:50:23 crc kubenswrapper[4885]: I1002 02:50:23.745467 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-sr7mv_c2a4785b-6378-4f85-ae1d-46e27903af36/nmstate-metrics/0.log" Oct 02 02:50:23 crc kubenswrapper[4885]: I1002 02:50:23.851286 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-858ddd8f98-r4dhw_6f1fd2a7-4a15-46b4-818f-c24f9fcd2bcd/nmstate-operator/0.log" Oct 02 02:50:23 crc kubenswrapper[4885]: I1002 02:50:23.921816 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6cdbc54649-jk45f_f75137b4-4551-43cd-b20c-6044d13d27a1/nmstate-webhook/0.log" Oct 02 02:50:26 crc kubenswrapper[4885]: I1002 02:50:26.045941 4885 scope.go:117] "RemoveContainer" containerID="999b3e75477511af3aa11f1e9d932c1ffb739b6f215cc19f45ac7144051ab65c" Oct 02 02:50:26 crc kubenswrapper[4885]: E1002 02:50:26.046806 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:50:38 crc kubenswrapper[4885]: I1002 02:50:38.552662 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-djznn_47f27b52-1453-4566-954f-94e6b02b2221/kube-rbac-proxy/0.log" Oct 02 02:50:38 crc kubenswrapper[4885]: I1002 02:50:38.677335 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-djznn_47f27b52-1453-4566-954f-94e6b02b2221/controller/0.log" Oct 02 02:50:38 crc kubenswrapper[4885]: I1002 02:50:38.765897 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ggsh4_05715b27-682b-46a1-80b5-667667147ce0/cp-frr-files/0.log" Oct 02 02:50:38 crc kubenswrapper[4885]: I1002 02:50:38.906741 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ggsh4_05715b27-682b-46a1-80b5-667667147ce0/cp-frr-files/0.log" Oct 02 02:50:38 crc kubenswrapper[4885]: I1002 02:50:38.924141 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ggsh4_05715b27-682b-46a1-80b5-667667147ce0/cp-reloader/0.log" Oct 02 02:50:38 crc kubenswrapper[4885]: I1002 02:50:38.955455 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ggsh4_05715b27-682b-46a1-80b5-667667147ce0/cp-metrics/0.log" Oct 02 02:50:38 crc kubenswrapper[4885]: I1002 02:50:38.966019 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ggsh4_05715b27-682b-46a1-80b5-667667147ce0/cp-reloader/0.log" Oct 02 02:50:39 crc kubenswrapper[4885]: I1002 02:50:39.113423 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ggsh4_05715b27-682b-46a1-80b5-667667147ce0/cp-frr-files/0.log" Oct 02 02:50:39 crc kubenswrapper[4885]: I1002 02:50:39.121714 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ggsh4_05715b27-682b-46a1-80b5-667667147ce0/cp-metrics/0.log" Oct 02 02:50:39 crc kubenswrapper[4885]: I1002 02:50:39.139724 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ggsh4_05715b27-682b-46a1-80b5-667667147ce0/cp-reloader/0.log" Oct 02 02:50:39 crc kubenswrapper[4885]: I1002 02:50:39.164090 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ggsh4_05715b27-682b-46a1-80b5-667667147ce0/cp-metrics/0.log" Oct 02 02:50:39 crc kubenswrapper[4885]: I1002 02:50:39.375044 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ggsh4_05715b27-682b-46a1-80b5-667667147ce0/cp-reloader/0.log" Oct 02 02:50:39 crc kubenswrapper[4885]: I1002 02:50:39.384575 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ggsh4_05715b27-682b-46a1-80b5-667667147ce0/cp-metrics/0.log" Oct 02 02:50:39 crc kubenswrapper[4885]: I1002 02:50:39.386502 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ggsh4_05715b27-682b-46a1-80b5-667667147ce0/cp-frr-files/0.log" Oct 02 02:50:39 crc kubenswrapper[4885]: I1002 02:50:39.413100 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ggsh4_05715b27-682b-46a1-80b5-667667147ce0/controller/0.log" Oct 02 02:50:39 crc kubenswrapper[4885]: I1002 02:50:39.562453 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ggsh4_05715b27-682b-46a1-80b5-667667147ce0/frr-metrics/0.log" Oct 02 02:50:39 crc kubenswrapper[4885]: I1002 02:50:39.563992 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ggsh4_05715b27-682b-46a1-80b5-667667147ce0/kube-rbac-proxy/0.log" Oct 02 02:50:39 crc kubenswrapper[4885]: I1002 02:50:39.624908 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ggsh4_05715b27-682b-46a1-80b5-667667147ce0/kube-rbac-proxy-frr/0.log" Oct 02 02:50:39 crc kubenswrapper[4885]: I1002 02:50:39.824883 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ggsh4_05715b27-682b-46a1-80b5-667667147ce0/reloader/0.log" Oct 02 02:50:39 crc kubenswrapper[4885]: I1002 02:50:39.875666 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-64bf5d555-q7xtq_66663f53-9790-48b4-8d16-4675661f340d/frr-k8s-webhook-server/0.log" Oct 02 02:50:40 crc kubenswrapper[4885]: I1002 02:50:40.018072 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-6c54f448c5-tknpj_7dc51fda-276d-4739-975d-e88c9fcec5d3/manager/0.log" Oct 02 02:50:40 crc kubenswrapper[4885]: I1002 02:50:40.046381 4885 scope.go:117] "RemoveContainer" containerID="999b3e75477511af3aa11f1e9d932c1ffb739b6f215cc19f45ac7144051ab65c" Oct 02 02:50:40 crc kubenswrapper[4885]: E1002 02:50:40.046620 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:50:40 crc kubenswrapper[4885]: I1002 02:50:40.163123 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-fd987dcb7-f2dx2_e0315b09-692d-4c5f-8861-0863f90338fa/webhook-server/0.log" Oct 02 02:50:40 crc kubenswrapper[4885]: I1002 02:50:40.349814 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-5h9mk_e3d6e1b0-5963-4e8f-bf3d-eb0e88ec0395/kube-rbac-proxy/0.log" Oct 02 02:50:40 crc kubenswrapper[4885]: I1002 02:50:40.797505 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-5h9mk_e3d6e1b0-5963-4e8f-bf3d-eb0e88ec0395/speaker/0.log" Oct 02 02:50:40 crc kubenswrapper[4885]: I1002 02:50:40.854252 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ggsh4_05715b27-682b-46a1-80b5-667667147ce0/frr/0.log" Oct 02 02:50:53 crc kubenswrapper[4885]: I1002 02:50:53.918705 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2ms2kw_65b0be27-bd89-4ddd-a578-7977b676f674/util/0.log" Oct 02 02:50:54 crc kubenswrapper[4885]: I1002 02:50:54.046816 4885 scope.go:117] "RemoveContainer" containerID="999b3e75477511af3aa11f1e9d932c1ffb739b6f215cc19f45ac7144051ab65c" Oct 02 02:50:54 crc kubenswrapper[4885]: E1002 02:50:54.047073 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:50:54 crc kubenswrapper[4885]: I1002 02:50:54.057801 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2ms2kw_65b0be27-bd89-4ddd-a578-7977b676f674/util/0.log" Oct 02 02:50:54 crc kubenswrapper[4885]: I1002 02:50:54.073343 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2ms2kw_65b0be27-bd89-4ddd-a578-7977b676f674/pull/0.log" Oct 02 02:50:54 crc kubenswrapper[4885]: I1002 02:50:54.089393 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2ms2kw_65b0be27-bd89-4ddd-a578-7977b676f674/pull/0.log" Oct 02 02:50:54 crc kubenswrapper[4885]: I1002 02:50:54.234685 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2ms2kw_65b0be27-bd89-4ddd-a578-7977b676f674/pull/0.log" Oct 02 02:50:54 crc kubenswrapper[4885]: I1002 02:50:54.254840 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2ms2kw_65b0be27-bd89-4ddd-a578-7977b676f674/util/0.log" Oct 02 02:50:54 crc kubenswrapper[4885]: I1002 02:50:54.308238 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2ms2kw_65b0be27-bd89-4ddd-a578-7977b676f674/extract/0.log" Oct 02 02:50:54 crc kubenswrapper[4885]: I1002 02:50:54.424568 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-sn6xg_56c1dc36-0146-45ea-9396-990a988a6d9e/extract-utilities/0.log" Oct 02 02:50:54 crc kubenswrapper[4885]: I1002 02:50:54.592098 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-sn6xg_56c1dc36-0146-45ea-9396-990a988a6d9e/extract-utilities/0.log" Oct 02 02:50:54 crc kubenswrapper[4885]: I1002 02:50:54.597049 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-sn6xg_56c1dc36-0146-45ea-9396-990a988a6d9e/extract-content/0.log" Oct 02 02:50:54 crc kubenswrapper[4885]: I1002 02:50:54.638869 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-sn6xg_56c1dc36-0146-45ea-9396-990a988a6d9e/extract-content/0.log" Oct 02 02:50:54 crc kubenswrapper[4885]: I1002 02:50:54.752136 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-sn6xg_56c1dc36-0146-45ea-9396-990a988a6d9e/extract-utilities/0.log" Oct 02 02:50:54 crc kubenswrapper[4885]: I1002 02:50:54.757123 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-sn6xg_56c1dc36-0146-45ea-9396-990a988a6d9e/extract-content/0.log" Oct 02 02:50:54 crc kubenswrapper[4885]: I1002 02:50:54.977636 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-xbmgl_d8e59d2f-c3ac-4a7a-b8a1-b1de59eb0724/extract-utilities/0.log" Oct 02 02:50:55 crc kubenswrapper[4885]: I1002 02:50:55.124611 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-xbmgl_d8e59d2f-c3ac-4a7a-b8a1-b1de59eb0724/extract-utilities/0.log" Oct 02 02:50:55 crc kubenswrapper[4885]: I1002 02:50:55.145412 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-xbmgl_d8e59d2f-c3ac-4a7a-b8a1-b1de59eb0724/extract-content/0.log" Oct 02 02:50:55 crc kubenswrapper[4885]: I1002 02:50:55.189395 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-sn6xg_56c1dc36-0146-45ea-9396-990a988a6d9e/registry-server/0.log" Oct 02 02:50:55 crc kubenswrapper[4885]: I1002 02:50:55.201295 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-xbmgl_d8e59d2f-c3ac-4a7a-b8a1-b1de59eb0724/extract-content/0.log" Oct 02 02:50:55 crc kubenswrapper[4885]: I1002 02:50:55.352488 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-xbmgl_d8e59d2f-c3ac-4a7a-b8a1-b1de59eb0724/extract-content/0.log" Oct 02 02:50:55 crc kubenswrapper[4885]: I1002 02:50:55.366756 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-xbmgl_d8e59d2f-c3ac-4a7a-b8a1-b1de59eb0724/extract-utilities/0.log" Oct 02 02:50:55 crc kubenswrapper[4885]: I1002 02:50:55.548906 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ctlhn5_533f5d4b-02ed-4235-8eed-715dd68a58fd/util/0.log" Oct 02 02:50:55 crc kubenswrapper[4885]: I1002 02:50:55.643005 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-xbmgl_d8e59d2f-c3ac-4a7a-b8a1-b1de59eb0724/registry-server/0.log" Oct 02 02:50:55 crc kubenswrapper[4885]: I1002 02:50:55.731862 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ctlhn5_533f5d4b-02ed-4235-8eed-715dd68a58fd/util/0.log" Oct 02 02:50:55 crc kubenswrapper[4885]: I1002 02:50:55.761787 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ctlhn5_533f5d4b-02ed-4235-8eed-715dd68a58fd/pull/0.log" Oct 02 02:50:55 crc kubenswrapper[4885]: I1002 02:50:55.769932 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ctlhn5_533f5d4b-02ed-4235-8eed-715dd68a58fd/pull/0.log" Oct 02 02:50:55 crc kubenswrapper[4885]: I1002 02:50:55.911388 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ctlhn5_533f5d4b-02ed-4235-8eed-715dd68a58fd/util/0.log" Oct 02 02:50:55 crc kubenswrapper[4885]: I1002 02:50:55.941061 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ctlhn5_533f5d4b-02ed-4235-8eed-715dd68a58fd/extract/0.log" Oct 02 02:50:55 crc kubenswrapper[4885]: I1002 02:50:55.954592 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ctlhn5_533f5d4b-02ed-4235-8eed-715dd68a58fd/pull/0.log" Oct 02 02:50:56 crc kubenswrapper[4885]: I1002 02:50:56.100191 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-z7fcl_b0909ffb-180c-4e74-9916-98de47f2419d/marketplace-operator/0.log" Oct 02 02:50:56 crc kubenswrapper[4885]: I1002 02:50:56.149995 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-sbsjt_eb7219fa-b79f-48be-9c5d-6453330d9527/extract-utilities/0.log" Oct 02 02:50:56 crc kubenswrapper[4885]: I1002 02:50:56.349700 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-sbsjt_eb7219fa-b79f-48be-9c5d-6453330d9527/extract-content/0.log" Oct 02 02:50:56 crc kubenswrapper[4885]: I1002 02:50:56.356168 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-sbsjt_eb7219fa-b79f-48be-9c5d-6453330d9527/extract-utilities/0.log" Oct 02 02:50:56 crc kubenswrapper[4885]: I1002 02:50:56.365746 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-sbsjt_eb7219fa-b79f-48be-9c5d-6453330d9527/extract-content/0.log" Oct 02 02:50:56 crc kubenswrapper[4885]: I1002 02:50:56.512454 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-sbsjt_eb7219fa-b79f-48be-9c5d-6453330d9527/extract-utilities/0.log" Oct 02 02:50:56 crc kubenswrapper[4885]: I1002 02:50:56.523397 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-sbsjt_eb7219fa-b79f-48be-9c5d-6453330d9527/extract-content/0.log" Oct 02 02:50:56 crc kubenswrapper[4885]: I1002 02:50:56.646464 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-sbsjt_eb7219fa-b79f-48be-9c5d-6453330d9527/registry-server/0.log" Oct 02 02:50:56 crc kubenswrapper[4885]: I1002 02:50:56.717184 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-9tmpc_747efce9-9b31-4e1b-a7b5-94019bcce7f4/extract-utilities/0.log" Oct 02 02:50:56 crc kubenswrapper[4885]: I1002 02:50:56.854089 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-9tmpc_747efce9-9b31-4e1b-a7b5-94019bcce7f4/extract-content/0.log" Oct 02 02:50:56 crc kubenswrapper[4885]: I1002 02:50:56.890572 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-9tmpc_747efce9-9b31-4e1b-a7b5-94019bcce7f4/extract-content/0.log" Oct 02 02:50:56 crc kubenswrapper[4885]: I1002 02:50:56.891549 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-9tmpc_747efce9-9b31-4e1b-a7b5-94019bcce7f4/extract-utilities/0.log" Oct 02 02:50:57 crc kubenswrapper[4885]: I1002 02:50:57.047660 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-9tmpc_747efce9-9b31-4e1b-a7b5-94019bcce7f4/extract-utilities/0.log" Oct 02 02:50:57 crc kubenswrapper[4885]: I1002 02:50:57.050123 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-9tmpc_747efce9-9b31-4e1b-a7b5-94019bcce7f4/extract-content/0.log" Oct 02 02:50:57 crc kubenswrapper[4885]: I1002 02:50:57.588007 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-9tmpc_747efce9-9b31-4e1b-a7b5-94019bcce7f4/registry-server/0.log" Oct 02 02:51:09 crc kubenswrapper[4885]: I1002 02:51:09.047362 4885 scope.go:117] "RemoveContainer" containerID="999b3e75477511af3aa11f1e9d932c1ffb739b6f215cc19f45ac7144051ab65c" Oct 02 02:51:09 crc kubenswrapper[4885]: E1002 02:51:09.048286 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:51:23 crc kubenswrapper[4885]: I1002 02:51:23.046658 4885 scope.go:117] "RemoveContainer" containerID="999b3e75477511af3aa11f1e9d932c1ffb739b6f215cc19f45ac7144051ab65c" Oct 02 02:51:23 crc kubenswrapper[4885]: E1002 02:51:23.048310 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:51:37 crc kubenswrapper[4885]: I1002 02:51:37.048918 4885 scope.go:117] "RemoveContainer" containerID="999b3e75477511af3aa11f1e9d932c1ffb739b6f215cc19f45ac7144051ab65c" Oct 02 02:51:37 crc kubenswrapper[4885]: E1002 02:51:37.050056 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:51:49 crc kubenswrapper[4885]: I1002 02:51:49.046215 4885 scope.go:117] "RemoveContainer" containerID="999b3e75477511af3aa11f1e9d932c1ffb739b6f215cc19f45ac7144051ab65c" Oct 02 02:51:49 crc kubenswrapper[4885]: E1002 02:51:49.046968 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:52:00 crc kubenswrapper[4885]: I1002 02:52:00.047725 4885 scope.go:117] "RemoveContainer" containerID="999b3e75477511af3aa11f1e9d932c1ffb739b6f215cc19f45ac7144051ab65c" Oct 02 02:52:00 crc kubenswrapper[4885]: E1002 02:52:00.049017 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:52:11 crc kubenswrapper[4885]: I1002 02:52:11.046667 4885 scope.go:117] "RemoveContainer" containerID="999b3e75477511af3aa11f1e9d932c1ffb739b6f215cc19f45ac7144051ab65c" Oct 02 02:52:11 crc kubenswrapper[4885]: E1002 02:52:11.047632 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:52:26 crc kubenswrapper[4885]: I1002 02:52:26.047064 4885 scope.go:117] "RemoveContainer" containerID="999b3e75477511af3aa11f1e9d932c1ffb739b6f215cc19f45ac7144051ab65c" Oct 02 02:52:26 crc kubenswrapper[4885]: I1002 02:52:26.828806 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" event={"ID":"d169c8ba-000a-4218-8e3b-4ae53035b110","Type":"ContainerStarted","Data":"514bc2a11d8477ee5f25c8b041d1d9f8fe40c82d21890ace02ef6041f3710407"} Oct 02 02:52:50 crc kubenswrapper[4885]: I1002 02:52:50.107385 4885 generic.go:334] "Generic (PLEG): container finished" podID="408c4e51-14de-441d-a876-e89a483ceef4" containerID="e9a0e101e717ebc532f542697393207e356a45656d9b3b7b793fd1312ab2b4be" exitCode=0 Oct 02 02:52:50 crc kubenswrapper[4885]: I1002 02:52:50.107433 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-28twm/must-gather-s92fx" event={"ID":"408c4e51-14de-441d-a876-e89a483ceef4","Type":"ContainerDied","Data":"e9a0e101e717ebc532f542697393207e356a45656d9b3b7b793fd1312ab2b4be"} Oct 02 02:52:50 crc kubenswrapper[4885]: I1002 02:52:50.108341 4885 scope.go:117] "RemoveContainer" containerID="e9a0e101e717ebc532f542697393207e356a45656d9b3b7b793fd1312ab2b4be" Oct 02 02:52:50 crc kubenswrapper[4885]: I1002 02:52:50.581395 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-28twm_must-gather-s92fx_408c4e51-14de-441d-a876-e89a483ceef4/gather/0.log" Oct 02 02:52:58 crc kubenswrapper[4885]: I1002 02:52:58.431717 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-28twm/must-gather-s92fx"] Oct 02 02:52:58 crc kubenswrapper[4885]: I1002 02:52:58.432580 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-28twm/must-gather-s92fx" podUID="408c4e51-14de-441d-a876-e89a483ceef4" containerName="copy" containerID="cri-o://dfa749ea2232255db4d7ab8402d46b0523e517e5abd73867d391cb29be8c11c3" gracePeriod=2 Oct 02 02:52:58 crc kubenswrapper[4885]: I1002 02:52:58.451847 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-28twm/must-gather-s92fx"] Oct 02 02:52:58 crc kubenswrapper[4885]: I1002 02:52:58.912244 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-28twm_must-gather-s92fx_408c4e51-14de-441d-a876-e89a483ceef4/copy/0.log" Oct 02 02:52:58 crc kubenswrapper[4885]: I1002 02:52:58.915342 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-28twm/must-gather-s92fx" Oct 02 02:52:58 crc kubenswrapper[4885]: I1002 02:52:58.988386 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/408c4e51-14de-441d-a876-e89a483ceef4-must-gather-output\") pod \"408c4e51-14de-441d-a876-e89a483ceef4\" (UID: \"408c4e51-14de-441d-a876-e89a483ceef4\") " Oct 02 02:52:58 crc kubenswrapper[4885]: I1002 02:52:58.988471 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q9pkh\" (UniqueName: \"kubernetes.io/projected/408c4e51-14de-441d-a876-e89a483ceef4-kube-api-access-q9pkh\") pod \"408c4e51-14de-441d-a876-e89a483ceef4\" (UID: \"408c4e51-14de-441d-a876-e89a483ceef4\") " Oct 02 02:52:59 crc kubenswrapper[4885]: I1002 02:52:59.000635 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/408c4e51-14de-441d-a876-e89a483ceef4-kube-api-access-q9pkh" (OuterVolumeSpecName: "kube-api-access-q9pkh") pod "408c4e51-14de-441d-a876-e89a483ceef4" (UID: "408c4e51-14de-441d-a876-e89a483ceef4"). InnerVolumeSpecName "kube-api-access-q9pkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:52:59 crc kubenswrapper[4885]: I1002 02:52:59.091086 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q9pkh\" (UniqueName: \"kubernetes.io/projected/408c4e51-14de-441d-a876-e89a483ceef4-kube-api-access-q9pkh\") on node \"crc\" DevicePath \"\"" Oct 02 02:52:59 crc kubenswrapper[4885]: I1002 02:52:59.130167 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/408c4e51-14de-441d-a876-e89a483ceef4-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "408c4e51-14de-441d-a876-e89a483ceef4" (UID: "408c4e51-14de-441d-a876-e89a483ceef4"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:52:59 crc kubenswrapper[4885]: I1002 02:52:59.192417 4885 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/408c4e51-14de-441d-a876-e89a483ceef4-must-gather-output\") on node \"crc\" DevicePath \"\"" Oct 02 02:52:59 crc kubenswrapper[4885]: I1002 02:52:59.204776 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-28twm_must-gather-s92fx_408c4e51-14de-441d-a876-e89a483ceef4/copy/0.log" Oct 02 02:52:59 crc kubenswrapper[4885]: I1002 02:52:59.205326 4885 generic.go:334] "Generic (PLEG): container finished" podID="408c4e51-14de-441d-a876-e89a483ceef4" containerID="dfa749ea2232255db4d7ab8402d46b0523e517e5abd73867d391cb29be8c11c3" exitCode=143 Oct 02 02:52:59 crc kubenswrapper[4885]: I1002 02:52:59.205383 4885 scope.go:117] "RemoveContainer" containerID="dfa749ea2232255db4d7ab8402d46b0523e517e5abd73867d391cb29be8c11c3" Oct 02 02:52:59 crc kubenswrapper[4885]: I1002 02:52:59.205517 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-28twm/must-gather-s92fx" Oct 02 02:52:59 crc kubenswrapper[4885]: I1002 02:52:59.231609 4885 scope.go:117] "RemoveContainer" containerID="e9a0e101e717ebc532f542697393207e356a45656d9b3b7b793fd1312ab2b4be" Oct 02 02:52:59 crc kubenswrapper[4885]: I1002 02:52:59.289421 4885 scope.go:117] "RemoveContainer" containerID="dfa749ea2232255db4d7ab8402d46b0523e517e5abd73867d391cb29be8c11c3" Oct 02 02:52:59 crc kubenswrapper[4885]: E1002 02:52:59.289893 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dfa749ea2232255db4d7ab8402d46b0523e517e5abd73867d391cb29be8c11c3\": container with ID starting with dfa749ea2232255db4d7ab8402d46b0523e517e5abd73867d391cb29be8c11c3 not found: ID does not exist" containerID="dfa749ea2232255db4d7ab8402d46b0523e517e5abd73867d391cb29be8c11c3" Oct 02 02:52:59 crc kubenswrapper[4885]: I1002 02:52:59.289920 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dfa749ea2232255db4d7ab8402d46b0523e517e5abd73867d391cb29be8c11c3"} err="failed to get container status \"dfa749ea2232255db4d7ab8402d46b0523e517e5abd73867d391cb29be8c11c3\": rpc error: code = NotFound desc = could not find container \"dfa749ea2232255db4d7ab8402d46b0523e517e5abd73867d391cb29be8c11c3\": container with ID starting with dfa749ea2232255db4d7ab8402d46b0523e517e5abd73867d391cb29be8c11c3 not found: ID does not exist" Oct 02 02:52:59 crc kubenswrapper[4885]: I1002 02:52:59.289939 4885 scope.go:117] "RemoveContainer" containerID="e9a0e101e717ebc532f542697393207e356a45656d9b3b7b793fd1312ab2b4be" Oct 02 02:52:59 crc kubenswrapper[4885]: E1002 02:52:59.290377 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e9a0e101e717ebc532f542697393207e356a45656d9b3b7b793fd1312ab2b4be\": container with ID starting with e9a0e101e717ebc532f542697393207e356a45656d9b3b7b793fd1312ab2b4be not found: ID does not exist" containerID="e9a0e101e717ebc532f542697393207e356a45656d9b3b7b793fd1312ab2b4be" Oct 02 02:52:59 crc kubenswrapper[4885]: I1002 02:52:59.290392 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e9a0e101e717ebc532f542697393207e356a45656d9b3b7b793fd1312ab2b4be"} err="failed to get container status \"e9a0e101e717ebc532f542697393207e356a45656d9b3b7b793fd1312ab2b4be\": rpc error: code = NotFound desc = could not find container \"e9a0e101e717ebc532f542697393207e356a45656d9b3b7b793fd1312ab2b4be\": container with ID starting with e9a0e101e717ebc532f542697393207e356a45656d9b3b7b793fd1312ab2b4be not found: ID does not exist" Oct 02 02:53:00 crc kubenswrapper[4885]: I1002 02:53:00.060674 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="408c4e51-14de-441d-a876-e89a483ceef4" path="/var/lib/kubelet/pods/408c4e51-14de-441d-a876-e89a483ceef4/volumes" Oct 02 02:53:32 crc kubenswrapper[4885]: I1002 02:53:32.918544 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-hv6gh"] Oct 02 02:53:32 crc kubenswrapper[4885]: E1002 02:53:32.919630 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1eb5a090-12a5-4f21-a5c8-f23bcd3a34f7" containerName="container-00" Oct 02 02:53:32 crc kubenswrapper[4885]: I1002 02:53:32.919648 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="1eb5a090-12a5-4f21-a5c8-f23bcd3a34f7" containerName="container-00" Oct 02 02:53:32 crc kubenswrapper[4885]: E1002 02:53:32.919663 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="408c4e51-14de-441d-a876-e89a483ceef4" containerName="copy" Oct 02 02:53:32 crc kubenswrapper[4885]: I1002 02:53:32.919671 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="408c4e51-14de-441d-a876-e89a483ceef4" containerName="copy" Oct 02 02:53:32 crc kubenswrapper[4885]: E1002 02:53:32.919696 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="408c4e51-14de-441d-a876-e89a483ceef4" containerName="gather" Oct 02 02:53:32 crc kubenswrapper[4885]: I1002 02:53:32.919704 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="408c4e51-14de-441d-a876-e89a483ceef4" containerName="gather" Oct 02 02:53:32 crc kubenswrapper[4885]: I1002 02:53:32.919963 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="1eb5a090-12a5-4f21-a5c8-f23bcd3a34f7" containerName="container-00" Oct 02 02:53:32 crc kubenswrapper[4885]: I1002 02:53:32.919989 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="408c4e51-14de-441d-a876-e89a483ceef4" containerName="gather" Oct 02 02:53:32 crc kubenswrapper[4885]: I1002 02:53:32.920008 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="408c4e51-14de-441d-a876-e89a483ceef4" containerName="copy" Oct 02 02:53:32 crc kubenswrapper[4885]: I1002 02:53:32.925722 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hv6gh" Oct 02 02:53:32 crc kubenswrapper[4885]: I1002 02:53:32.931558 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hv6gh"] Oct 02 02:53:33 crc kubenswrapper[4885]: I1002 02:53:33.006979 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08218259-0111-4549-bdbe-18e8cfe529ed-utilities\") pod \"redhat-operators-hv6gh\" (UID: \"08218259-0111-4549-bdbe-18e8cfe529ed\") " pod="openshift-marketplace/redhat-operators-hv6gh" Oct 02 02:53:33 crc kubenswrapper[4885]: I1002 02:53:33.007099 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08218259-0111-4549-bdbe-18e8cfe529ed-catalog-content\") pod \"redhat-operators-hv6gh\" (UID: \"08218259-0111-4549-bdbe-18e8cfe529ed\") " pod="openshift-marketplace/redhat-operators-hv6gh" Oct 02 02:53:33 crc kubenswrapper[4885]: I1002 02:53:33.007206 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lfwjg\" (UniqueName: \"kubernetes.io/projected/08218259-0111-4549-bdbe-18e8cfe529ed-kube-api-access-lfwjg\") pod \"redhat-operators-hv6gh\" (UID: \"08218259-0111-4549-bdbe-18e8cfe529ed\") " pod="openshift-marketplace/redhat-operators-hv6gh" Oct 02 02:53:33 crc kubenswrapper[4885]: I1002 02:53:33.108675 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lfwjg\" (UniqueName: \"kubernetes.io/projected/08218259-0111-4549-bdbe-18e8cfe529ed-kube-api-access-lfwjg\") pod \"redhat-operators-hv6gh\" (UID: \"08218259-0111-4549-bdbe-18e8cfe529ed\") " pod="openshift-marketplace/redhat-operators-hv6gh" Oct 02 02:53:33 crc kubenswrapper[4885]: I1002 02:53:33.108786 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08218259-0111-4549-bdbe-18e8cfe529ed-utilities\") pod \"redhat-operators-hv6gh\" (UID: \"08218259-0111-4549-bdbe-18e8cfe529ed\") " pod="openshift-marketplace/redhat-operators-hv6gh" Oct 02 02:53:33 crc kubenswrapper[4885]: I1002 02:53:33.108844 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08218259-0111-4549-bdbe-18e8cfe529ed-catalog-content\") pod \"redhat-operators-hv6gh\" (UID: \"08218259-0111-4549-bdbe-18e8cfe529ed\") " pod="openshift-marketplace/redhat-operators-hv6gh" Oct 02 02:53:33 crc kubenswrapper[4885]: I1002 02:53:33.109295 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08218259-0111-4549-bdbe-18e8cfe529ed-catalog-content\") pod \"redhat-operators-hv6gh\" (UID: \"08218259-0111-4549-bdbe-18e8cfe529ed\") " pod="openshift-marketplace/redhat-operators-hv6gh" Oct 02 02:53:33 crc kubenswrapper[4885]: I1002 02:53:33.109332 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08218259-0111-4549-bdbe-18e8cfe529ed-utilities\") pod \"redhat-operators-hv6gh\" (UID: \"08218259-0111-4549-bdbe-18e8cfe529ed\") " pod="openshift-marketplace/redhat-operators-hv6gh" Oct 02 02:53:33 crc kubenswrapper[4885]: I1002 02:53:33.141028 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lfwjg\" (UniqueName: \"kubernetes.io/projected/08218259-0111-4549-bdbe-18e8cfe529ed-kube-api-access-lfwjg\") pod \"redhat-operators-hv6gh\" (UID: \"08218259-0111-4549-bdbe-18e8cfe529ed\") " pod="openshift-marketplace/redhat-operators-hv6gh" Oct 02 02:53:33 crc kubenswrapper[4885]: I1002 02:53:33.274252 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hv6gh" Oct 02 02:53:33 crc kubenswrapper[4885]: I1002 02:53:33.755096 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hv6gh"] Oct 02 02:53:34 crc kubenswrapper[4885]: I1002 02:53:34.654179 4885 generic.go:334] "Generic (PLEG): container finished" podID="08218259-0111-4549-bdbe-18e8cfe529ed" containerID="baf4e03bb70c7073fcabd66d69e1c4f4c68192adea5ff489a8a808b24f13a223" exitCode=0 Oct 02 02:53:34 crc kubenswrapper[4885]: I1002 02:53:34.654301 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hv6gh" event={"ID":"08218259-0111-4549-bdbe-18e8cfe529ed","Type":"ContainerDied","Data":"baf4e03bb70c7073fcabd66d69e1c4f4c68192adea5ff489a8a808b24f13a223"} Oct 02 02:53:34 crc kubenswrapper[4885]: I1002 02:53:34.654563 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hv6gh" event={"ID":"08218259-0111-4549-bdbe-18e8cfe529ed","Type":"ContainerStarted","Data":"635fb98e4f33dc88740e2692b5c5088f48fed66fd7d3c47b847e027ec683ad12"} Oct 02 02:53:34 crc kubenswrapper[4885]: I1002 02:53:34.657011 4885 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 02:53:36 crc kubenswrapper[4885]: I1002 02:53:36.690108 4885 generic.go:334] "Generic (PLEG): container finished" podID="08218259-0111-4549-bdbe-18e8cfe529ed" containerID="956489b831ec40e357a72f59b81d7141433c91e98951d8bb57a4e6c0b8aaf5e1" exitCode=0 Oct 02 02:53:36 crc kubenswrapper[4885]: I1002 02:53:36.690315 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hv6gh" event={"ID":"08218259-0111-4549-bdbe-18e8cfe529ed","Type":"ContainerDied","Data":"956489b831ec40e357a72f59b81d7141433c91e98951d8bb57a4e6c0b8aaf5e1"} Oct 02 02:53:37 crc kubenswrapper[4885]: I1002 02:53:37.705896 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hv6gh" event={"ID":"08218259-0111-4549-bdbe-18e8cfe529ed","Type":"ContainerStarted","Data":"9583e4b035aec75ac750403d669046f32bb9764530877f306f68e4f5021d874a"} Oct 02 02:53:37 crc kubenswrapper[4885]: I1002 02:53:37.746725 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-hv6gh" podStartSLOduration=3.259197318 podStartE2EDuration="5.746706756s" podCreationTimestamp="2025-10-02 02:53:32 +0000 UTC" firstStartedPulling="2025-10-02 02:53:34.656650614 +0000 UTC m=+4003.468398023" lastFinishedPulling="2025-10-02 02:53:37.144160022 +0000 UTC m=+4005.955907461" observedRunningTime="2025-10-02 02:53:37.738714122 +0000 UTC m=+4006.550461531" watchObservedRunningTime="2025-10-02 02:53:37.746706756 +0000 UTC m=+4006.558454165" Oct 02 02:53:43 crc kubenswrapper[4885]: I1002 02:53:43.275162 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-hv6gh" Oct 02 02:53:43 crc kubenswrapper[4885]: I1002 02:53:43.276132 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-hv6gh" Oct 02 02:53:44 crc kubenswrapper[4885]: I1002 02:53:44.157967 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-wl5qp"] Oct 02 02:53:44 crc kubenswrapper[4885]: I1002 02:53:44.161541 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wl5qp" Oct 02 02:53:44 crc kubenswrapper[4885]: I1002 02:53:44.172793 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wl5qp"] Oct 02 02:53:44 crc kubenswrapper[4885]: I1002 02:53:44.277948 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a53169d-5a5b-466c-988c-3e60ceef0111-catalog-content\") pod \"redhat-marketplace-wl5qp\" (UID: \"2a53169d-5a5b-466c-988c-3e60ceef0111\") " pod="openshift-marketplace/redhat-marketplace-wl5qp" Oct 02 02:53:44 crc kubenswrapper[4885]: I1002 02:53:44.278392 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nttjj\" (UniqueName: \"kubernetes.io/projected/2a53169d-5a5b-466c-988c-3e60ceef0111-kube-api-access-nttjj\") pod \"redhat-marketplace-wl5qp\" (UID: \"2a53169d-5a5b-466c-988c-3e60ceef0111\") " pod="openshift-marketplace/redhat-marketplace-wl5qp" Oct 02 02:53:44 crc kubenswrapper[4885]: I1002 02:53:44.278433 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a53169d-5a5b-466c-988c-3e60ceef0111-utilities\") pod \"redhat-marketplace-wl5qp\" (UID: \"2a53169d-5a5b-466c-988c-3e60ceef0111\") " pod="openshift-marketplace/redhat-marketplace-wl5qp" Oct 02 02:53:44 crc kubenswrapper[4885]: I1002 02:53:44.340461 4885 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-hv6gh" podUID="08218259-0111-4549-bdbe-18e8cfe529ed" containerName="registry-server" probeResult="failure" output=< Oct 02 02:53:44 crc kubenswrapper[4885]: timeout: failed to connect service ":50051" within 1s Oct 02 02:53:44 crc kubenswrapper[4885]: > Oct 02 02:53:44 crc kubenswrapper[4885]: I1002 02:53:44.379714 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nttjj\" (UniqueName: \"kubernetes.io/projected/2a53169d-5a5b-466c-988c-3e60ceef0111-kube-api-access-nttjj\") pod \"redhat-marketplace-wl5qp\" (UID: \"2a53169d-5a5b-466c-988c-3e60ceef0111\") " pod="openshift-marketplace/redhat-marketplace-wl5qp" Oct 02 02:53:44 crc kubenswrapper[4885]: I1002 02:53:44.379766 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a53169d-5a5b-466c-988c-3e60ceef0111-utilities\") pod \"redhat-marketplace-wl5qp\" (UID: \"2a53169d-5a5b-466c-988c-3e60ceef0111\") " pod="openshift-marketplace/redhat-marketplace-wl5qp" Oct 02 02:53:44 crc kubenswrapper[4885]: I1002 02:53:44.379820 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a53169d-5a5b-466c-988c-3e60ceef0111-catalog-content\") pod \"redhat-marketplace-wl5qp\" (UID: \"2a53169d-5a5b-466c-988c-3e60ceef0111\") " pod="openshift-marketplace/redhat-marketplace-wl5qp" Oct 02 02:53:44 crc kubenswrapper[4885]: I1002 02:53:44.380381 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a53169d-5a5b-466c-988c-3e60ceef0111-catalog-content\") pod \"redhat-marketplace-wl5qp\" (UID: \"2a53169d-5a5b-466c-988c-3e60ceef0111\") " pod="openshift-marketplace/redhat-marketplace-wl5qp" Oct 02 02:53:44 crc kubenswrapper[4885]: I1002 02:53:44.380573 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a53169d-5a5b-466c-988c-3e60ceef0111-utilities\") pod \"redhat-marketplace-wl5qp\" (UID: \"2a53169d-5a5b-466c-988c-3e60ceef0111\") " pod="openshift-marketplace/redhat-marketplace-wl5qp" Oct 02 02:53:44 crc kubenswrapper[4885]: I1002 02:53:44.398130 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nttjj\" (UniqueName: \"kubernetes.io/projected/2a53169d-5a5b-466c-988c-3e60ceef0111-kube-api-access-nttjj\") pod \"redhat-marketplace-wl5qp\" (UID: \"2a53169d-5a5b-466c-988c-3e60ceef0111\") " pod="openshift-marketplace/redhat-marketplace-wl5qp" Oct 02 02:53:44 crc kubenswrapper[4885]: I1002 02:53:44.499944 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wl5qp" Oct 02 02:53:45 crc kubenswrapper[4885]: I1002 02:53:45.008274 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wl5qp"] Oct 02 02:53:45 crc kubenswrapper[4885]: W1002 02:53:45.017425 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2a53169d_5a5b_466c_988c_3e60ceef0111.slice/crio-917b17f8d483d829ded496a9c30d7e0163c18f9f0b4046a2313564e28eadee9f WatchSource:0}: Error finding container 917b17f8d483d829ded496a9c30d7e0163c18f9f0b4046a2313564e28eadee9f: Status 404 returned error can't find the container with id 917b17f8d483d829ded496a9c30d7e0163c18f9f0b4046a2313564e28eadee9f Oct 02 02:53:45 crc kubenswrapper[4885]: I1002 02:53:45.807818 4885 generic.go:334] "Generic (PLEG): container finished" podID="2a53169d-5a5b-466c-988c-3e60ceef0111" containerID="cb415bc03dfcfa249ab84b81298dd8fded1699a33ef5c770bc05a8f2aae446a1" exitCode=0 Oct 02 02:53:45 crc kubenswrapper[4885]: I1002 02:53:45.807943 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wl5qp" event={"ID":"2a53169d-5a5b-466c-988c-3e60ceef0111","Type":"ContainerDied","Data":"cb415bc03dfcfa249ab84b81298dd8fded1699a33ef5c770bc05a8f2aae446a1"} Oct 02 02:53:45 crc kubenswrapper[4885]: I1002 02:53:45.808330 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wl5qp" event={"ID":"2a53169d-5a5b-466c-988c-3e60ceef0111","Type":"ContainerStarted","Data":"917b17f8d483d829ded496a9c30d7e0163c18f9f0b4046a2313564e28eadee9f"} Oct 02 02:53:46 crc kubenswrapper[4885]: I1002 02:53:46.823308 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wl5qp" event={"ID":"2a53169d-5a5b-466c-988c-3e60ceef0111","Type":"ContainerStarted","Data":"21b94a8f784a3a8bf61ee39c3a63a2d1f45f817f9bcdaceff0b769e041c71f66"} Oct 02 02:53:48 crc kubenswrapper[4885]: I1002 02:53:48.847661 4885 generic.go:334] "Generic (PLEG): container finished" podID="2a53169d-5a5b-466c-988c-3e60ceef0111" containerID="21b94a8f784a3a8bf61ee39c3a63a2d1f45f817f9bcdaceff0b769e041c71f66" exitCode=0 Oct 02 02:53:48 crc kubenswrapper[4885]: I1002 02:53:48.847756 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wl5qp" event={"ID":"2a53169d-5a5b-466c-988c-3e60ceef0111","Type":"ContainerDied","Data":"21b94a8f784a3a8bf61ee39c3a63a2d1f45f817f9bcdaceff0b769e041c71f66"} Oct 02 02:53:50 crc kubenswrapper[4885]: I1002 02:53:50.874558 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wl5qp" event={"ID":"2a53169d-5a5b-466c-988c-3e60ceef0111","Type":"ContainerStarted","Data":"2024120544b22f474f05d7b052cd999260fcba3f94e9daf154b9fdd6fcc3b883"} Oct 02 02:53:50 crc kubenswrapper[4885]: I1002 02:53:50.907361 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-wl5qp" podStartSLOduration=3.415995326 podStartE2EDuration="6.907336558s" podCreationTimestamp="2025-10-02 02:53:44 +0000 UTC" firstStartedPulling="2025-10-02 02:53:45.810690859 +0000 UTC m=+4014.622438288" lastFinishedPulling="2025-10-02 02:53:49.302032101 +0000 UTC m=+4018.113779520" observedRunningTime="2025-10-02 02:53:50.903707427 +0000 UTC m=+4019.715454876" watchObservedRunningTime="2025-10-02 02:53:50.907336558 +0000 UTC m=+4019.719083977" Oct 02 02:53:53 crc kubenswrapper[4885]: I1002 02:53:53.358179 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-hv6gh" Oct 02 02:53:53 crc kubenswrapper[4885]: I1002 02:53:53.427999 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-hv6gh" Oct 02 02:53:53 crc kubenswrapper[4885]: I1002 02:53:53.598820 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hv6gh"] Oct 02 02:53:54 crc kubenswrapper[4885]: I1002 02:53:54.319657 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-4xx9f/must-gather-fbjpr"] Oct 02 02:53:54 crc kubenswrapper[4885]: I1002 02:53:54.321554 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4xx9f/must-gather-fbjpr" Oct 02 02:53:54 crc kubenswrapper[4885]: I1002 02:53:54.330598 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-4xx9f"/"openshift-service-ca.crt" Oct 02 02:53:54 crc kubenswrapper[4885]: I1002 02:53:54.330699 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-4xx9f"/"kube-root-ca.crt" Oct 02 02:53:54 crc kubenswrapper[4885]: I1002 02:53:54.350375 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-4xx9f/must-gather-fbjpr"] Oct 02 02:53:54 crc kubenswrapper[4885]: I1002 02:53:54.413350 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/a06901b1-5b99-4529-909b-2dfa182f0196-must-gather-output\") pod \"must-gather-fbjpr\" (UID: \"a06901b1-5b99-4529-909b-2dfa182f0196\") " pod="openshift-must-gather-4xx9f/must-gather-fbjpr" Oct 02 02:53:54 crc kubenswrapper[4885]: I1002 02:53:54.413597 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q4pqr\" (UniqueName: \"kubernetes.io/projected/a06901b1-5b99-4529-909b-2dfa182f0196-kube-api-access-q4pqr\") pod \"must-gather-fbjpr\" (UID: \"a06901b1-5b99-4529-909b-2dfa182f0196\") " pod="openshift-must-gather-4xx9f/must-gather-fbjpr" Oct 02 02:53:54 crc kubenswrapper[4885]: I1002 02:53:54.500196 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-wl5qp" Oct 02 02:53:54 crc kubenswrapper[4885]: I1002 02:53:54.500247 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-wl5qp" Oct 02 02:53:54 crc kubenswrapper[4885]: I1002 02:53:54.515050 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/a06901b1-5b99-4529-909b-2dfa182f0196-must-gather-output\") pod \"must-gather-fbjpr\" (UID: \"a06901b1-5b99-4529-909b-2dfa182f0196\") " pod="openshift-must-gather-4xx9f/must-gather-fbjpr" Oct 02 02:53:54 crc kubenswrapper[4885]: I1002 02:53:54.515139 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q4pqr\" (UniqueName: \"kubernetes.io/projected/a06901b1-5b99-4529-909b-2dfa182f0196-kube-api-access-q4pqr\") pod \"must-gather-fbjpr\" (UID: \"a06901b1-5b99-4529-909b-2dfa182f0196\") " pod="openshift-must-gather-4xx9f/must-gather-fbjpr" Oct 02 02:53:54 crc kubenswrapper[4885]: I1002 02:53:54.515856 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/a06901b1-5b99-4529-909b-2dfa182f0196-must-gather-output\") pod \"must-gather-fbjpr\" (UID: \"a06901b1-5b99-4529-909b-2dfa182f0196\") " pod="openshift-must-gather-4xx9f/must-gather-fbjpr" Oct 02 02:53:54 crc kubenswrapper[4885]: I1002 02:53:54.540212 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q4pqr\" (UniqueName: \"kubernetes.io/projected/a06901b1-5b99-4529-909b-2dfa182f0196-kube-api-access-q4pqr\") pod \"must-gather-fbjpr\" (UID: \"a06901b1-5b99-4529-909b-2dfa182f0196\") " pod="openshift-must-gather-4xx9f/must-gather-fbjpr" Oct 02 02:53:54 crc kubenswrapper[4885]: I1002 02:53:54.545996 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-wl5qp" Oct 02 02:53:54 crc kubenswrapper[4885]: I1002 02:53:54.637974 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4xx9f/must-gather-fbjpr" Oct 02 02:53:54 crc kubenswrapper[4885]: I1002 02:53:54.919379 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-hv6gh" podUID="08218259-0111-4549-bdbe-18e8cfe529ed" containerName="registry-server" containerID="cri-o://9583e4b035aec75ac750403d669046f32bb9764530877f306f68e4f5021d874a" gracePeriod=2 Oct 02 02:53:55 crc kubenswrapper[4885]: I1002 02:53:55.144497 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-4xx9f/must-gather-fbjpr"] Oct 02 02:53:55 crc kubenswrapper[4885]: I1002 02:53:55.369918 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hv6gh" Oct 02 02:53:55 crc kubenswrapper[4885]: I1002 02:53:55.429423 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08218259-0111-4549-bdbe-18e8cfe529ed-utilities\") pod \"08218259-0111-4549-bdbe-18e8cfe529ed\" (UID: \"08218259-0111-4549-bdbe-18e8cfe529ed\") " Oct 02 02:53:55 crc kubenswrapper[4885]: I1002 02:53:55.429562 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lfwjg\" (UniqueName: \"kubernetes.io/projected/08218259-0111-4549-bdbe-18e8cfe529ed-kube-api-access-lfwjg\") pod \"08218259-0111-4549-bdbe-18e8cfe529ed\" (UID: \"08218259-0111-4549-bdbe-18e8cfe529ed\") " Oct 02 02:53:55 crc kubenswrapper[4885]: I1002 02:53:55.429622 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08218259-0111-4549-bdbe-18e8cfe529ed-catalog-content\") pod \"08218259-0111-4549-bdbe-18e8cfe529ed\" (UID: \"08218259-0111-4549-bdbe-18e8cfe529ed\") " Oct 02 02:53:55 crc kubenswrapper[4885]: I1002 02:53:55.430327 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/08218259-0111-4549-bdbe-18e8cfe529ed-utilities" (OuterVolumeSpecName: "utilities") pod "08218259-0111-4549-bdbe-18e8cfe529ed" (UID: "08218259-0111-4549-bdbe-18e8cfe529ed"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:53:55 crc kubenswrapper[4885]: I1002 02:53:55.437860 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08218259-0111-4549-bdbe-18e8cfe529ed-kube-api-access-lfwjg" (OuterVolumeSpecName: "kube-api-access-lfwjg") pod "08218259-0111-4549-bdbe-18e8cfe529ed" (UID: "08218259-0111-4549-bdbe-18e8cfe529ed"). InnerVolumeSpecName "kube-api-access-lfwjg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:53:55 crc kubenswrapper[4885]: I1002 02:53:55.534190 4885 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08218259-0111-4549-bdbe-18e8cfe529ed-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 02:53:55 crc kubenswrapper[4885]: I1002 02:53:55.534229 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lfwjg\" (UniqueName: \"kubernetes.io/projected/08218259-0111-4549-bdbe-18e8cfe529ed-kube-api-access-lfwjg\") on node \"crc\" DevicePath \"\"" Oct 02 02:53:55 crc kubenswrapper[4885]: I1002 02:53:55.539064 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/08218259-0111-4549-bdbe-18e8cfe529ed-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "08218259-0111-4549-bdbe-18e8cfe529ed" (UID: "08218259-0111-4549-bdbe-18e8cfe529ed"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:53:55 crc kubenswrapper[4885]: I1002 02:53:55.636005 4885 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08218259-0111-4549-bdbe-18e8cfe529ed-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 02:53:55 crc kubenswrapper[4885]: I1002 02:53:55.931479 4885 generic.go:334] "Generic (PLEG): container finished" podID="08218259-0111-4549-bdbe-18e8cfe529ed" containerID="9583e4b035aec75ac750403d669046f32bb9764530877f306f68e4f5021d874a" exitCode=0 Oct 02 02:53:55 crc kubenswrapper[4885]: I1002 02:53:55.931562 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hv6gh" event={"ID":"08218259-0111-4549-bdbe-18e8cfe529ed","Type":"ContainerDied","Data":"9583e4b035aec75ac750403d669046f32bb9764530877f306f68e4f5021d874a"} Oct 02 02:53:55 crc kubenswrapper[4885]: I1002 02:53:55.931611 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hv6gh" event={"ID":"08218259-0111-4549-bdbe-18e8cfe529ed","Type":"ContainerDied","Data":"635fb98e4f33dc88740e2692b5c5088f48fed66fd7d3c47b847e027ec683ad12"} Oct 02 02:53:55 crc kubenswrapper[4885]: I1002 02:53:55.931631 4885 scope.go:117] "RemoveContainer" containerID="9583e4b035aec75ac750403d669046f32bb9764530877f306f68e4f5021d874a" Oct 02 02:53:55 crc kubenswrapper[4885]: I1002 02:53:55.931638 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hv6gh" Oct 02 02:53:55 crc kubenswrapper[4885]: I1002 02:53:55.934926 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4xx9f/must-gather-fbjpr" event={"ID":"a06901b1-5b99-4529-909b-2dfa182f0196","Type":"ContainerStarted","Data":"4d031c535ae33d7da37cabe51f3a634fe02db4affc1d117cb08c8012dcf9cdbf"} Oct 02 02:53:55 crc kubenswrapper[4885]: I1002 02:53:55.934957 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4xx9f/must-gather-fbjpr" event={"ID":"a06901b1-5b99-4529-909b-2dfa182f0196","Type":"ContainerStarted","Data":"2b6f9985b46f54a5c49e86cbf2a3ce430da3cd71ff838ecae9d919b670ed9e7c"} Oct 02 02:53:55 crc kubenswrapper[4885]: I1002 02:53:55.934966 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4xx9f/must-gather-fbjpr" event={"ID":"a06901b1-5b99-4529-909b-2dfa182f0196","Type":"ContainerStarted","Data":"0d488ad894f5d3258bb0d841e3297088051f1722fe5321f50f20fa85a53c3d81"} Oct 02 02:53:55 crc kubenswrapper[4885]: I1002 02:53:55.967421 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-4xx9f/must-gather-fbjpr" podStartSLOduration=1.967378238 podStartE2EDuration="1.967378238s" podCreationTimestamp="2025-10-02 02:53:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:53:55.959333963 +0000 UTC m=+4024.771081362" watchObservedRunningTime="2025-10-02 02:53:55.967378238 +0000 UTC m=+4024.779125657" Oct 02 02:53:55 crc kubenswrapper[4885]: I1002 02:53:55.978923 4885 scope.go:117] "RemoveContainer" containerID="956489b831ec40e357a72f59b81d7141433c91e98951d8bb57a4e6c0b8aaf5e1" Oct 02 02:53:56 crc kubenswrapper[4885]: I1002 02:53:56.003312 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hv6gh"] Oct 02 02:53:56 crc kubenswrapper[4885]: I1002 02:53:56.018236 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-hv6gh"] Oct 02 02:53:56 crc kubenswrapper[4885]: I1002 02:53:56.018678 4885 scope.go:117] "RemoveContainer" containerID="baf4e03bb70c7073fcabd66d69e1c4f4c68192adea5ff489a8a808b24f13a223" Oct 02 02:53:56 crc kubenswrapper[4885]: I1002 02:53:56.047571 4885 scope.go:117] "RemoveContainer" containerID="9583e4b035aec75ac750403d669046f32bb9764530877f306f68e4f5021d874a" Oct 02 02:53:56 crc kubenswrapper[4885]: E1002 02:53:56.049404 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9583e4b035aec75ac750403d669046f32bb9764530877f306f68e4f5021d874a\": container with ID starting with 9583e4b035aec75ac750403d669046f32bb9764530877f306f68e4f5021d874a not found: ID does not exist" containerID="9583e4b035aec75ac750403d669046f32bb9764530877f306f68e4f5021d874a" Oct 02 02:53:56 crc kubenswrapper[4885]: I1002 02:53:56.049454 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9583e4b035aec75ac750403d669046f32bb9764530877f306f68e4f5021d874a"} err="failed to get container status \"9583e4b035aec75ac750403d669046f32bb9764530877f306f68e4f5021d874a\": rpc error: code = NotFound desc = could not find container \"9583e4b035aec75ac750403d669046f32bb9764530877f306f68e4f5021d874a\": container with ID starting with 9583e4b035aec75ac750403d669046f32bb9764530877f306f68e4f5021d874a not found: ID does not exist" Oct 02 02:53:56 crc kubenswrapper[4885]: I1002 02:53:56.049521 4885 scope.go:117] "RemoveContainer" containerID="956489b831ec40e357a72f59b81d7141433c91e98951d8bb57a4e6c0b8aaf5e1" Oct 02 02:53:56 crc kubenswrapper[4885]: E1002 02:53:56.049861 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"956489b831ec40e357a72f59b81d7141433c91e98951d8bb57a4e6c0b8aaf5e1\": container with ID starting with 956489b831ec40e357a72f59b81d7141433c91e98951d8bb57a4e6c0b8aaf5e1 not found: ID does not exist" containerID="956489b831ec40e357a72f59b81d7141433c91e98951d8bb57a4e6c0b8aaf5e1" Oct 02 02:53:56 crc kubenswrapper[4885]: I1002 02:53:56.049919 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"956489b831ec40e357a72f59b81d7141433c91e98951d8bb57a4e6c0b8aaf5e1"} err="failed to get container status \"956489b831ec40e357a72f59b81d7141433c91e98951d8bb57a4e6c0b8aaf5e1\": rpc error: code = NotFound desc = could not find container \"956489b831ec40e357a72f59b81d7141433c91e98951d8bb57a4e6c0b8aaf5e1\": container with ID starting with 956489b831ec40e357a72f59b81d7141433c91e98951d8bb57a4e6c0b8aaf5e1 not found: ID does not exist" Oct 02 02:53:56 crc kubenswrapper[4885]: I1002 02:53:56.049958 4885 scope.go:117] "RemoveContainer" containerID="baf4e03bb70c7073fcabd66d69e1c4f4c68192adea5ff489a8a808b24f13a223" Oct 02 02:53:56 crc kubenswrapper[4885]: E1002 02:53:56.050323 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"baf4e03bb70c7073fcabd66d69e1c4f4c68192adea5ff489a8a808b24f13a223\": container with ID starting with baf4e03bb70c7073fcabd66d69e1c4f4c68192adea5ff489a8a808b24f13a223 not found: ID does not exist" containerID="baf4e03bb70c7073fcabd66d69e1c4f4c68192adea5ff489a8a808b24f13a223" Oct 02 02:53:56 crc kubenswrapper[4885]: I1002 02:53:56.050380 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"baf4e03bb70c7073fcabd66d69e1c4f4c68192adea5ff489a8a808b24f13a223"} err="failed to get container status \"baf4e03bb70c7073fcabd66d69e1c4f4c68192adea5ff489a8a808b24f13a223\": rpc error: code = NotFound desc = could not find container \"baf4e03bb70c7073fcabd66d69e1c4f4c68192adea5ff489a8a808b24f13a223\": container with ID starting with baf4e03bb70c7073fcabd66d69e1c4f4c68192adea5ff489a8a808b24f13a223 not found: ID does not exist" Oct 02 02:53:56 crc kubenswrapper[4885]: I1002 02:53:56.067740 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="08218259-0111-4549-bdbe-18e8cfe529ed" path="/var/lib/kubelet/pods/08218259-0111-4549-bdbe-18e8cfe529ed/volumes" Oct 02 02:53:59 crc kubenswrapper[4885]: I1002 02:53:59.543138 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-4xx9f/crc-debug-w5jvs"] Oct 02 02:53:59 crc kubenswrapper[4885]: E1002 02:53:59.543855 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08218259-0111-4549-bdbe-18e8cfe529ed" containerName="extract-content" Oct 02 02:53:59 crc kubenswrapper[4885]: I1002 02:53:59.543867 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="08218259-0111-4549-bdbe-18e8cfe529ed" containerName="extract-content" Oct 02 02:53:59 crc kubenswrapper[4885]: E1002 02:53:59.543881 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08218259-0111-4549-bdbe-18e8cfe529ed" containerName="extract-utilities" Oct 02 02:53:59 crc kubenswrapper[4885]: I1002 02:53:59.543887 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="08218259-0111-4549-bdbe-18e8cfe529ed" containerName="extract-utilities" Oct 02 02:53:59 crc kubenswrapper[4885]: E1002 02:53:59.543921 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08218259-0111-4549-bdbe-18e8cfe529ed" containerName="registry-server" Oct 02 02:53:59 crc kubenswrapper[4885]: I1002 02:53:59.543926 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="08218259-0111-4549-bdbe-18e8cfe529ed" containerName="registry-server" Oct 02 02:53:59 crc kubenswrapper[4885]: I1002 02:53:59.544090 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="08218259-0111-4549-bdbe-18e8cfe529ed" containerName="registry-server" Oct 02 02:53:59 crc kubenswrapper[4885]: I1002 02:53:59.544687 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4xx9f/crc-debug-w5jvs" Oct 02 02:53:59 crc kubenswrapper[4885]: I1002 02:53:59.546834 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-4xx9f"/"default-dockercfg-5ljjn" Oct 02 02:53:59 crc kubenswrapper[4885]: I1002 02:53:59.616245 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b9f7878c-36a4-4cdf-9db7-c4e9d3548d74-host\") pod \"crc-debug-w5jvs\" (UID: \"b9f7878c-36a4-4cdf-9db7-c4e9d3548d74\") " pod="openshift-must-gather-4xx9f/crc-debug-w5jvs" Oct 02 02:53:59 crc kubenswrapper[4885]: I1002 02:53:59.616867 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z2vq2\" (UniqueName: \"kubernetes.io/projected/b9f7878c-36a4-4cdf-9db7-c4e9d3548d74-kube-api-access-z2vq2\") pod \"crc-debug-w5jvs\" (UID: \"b9f7878c-36a4-4cdf-9db7-c4e9d3548d74\") " pod="openshift-must-gather-4xx9f/crc-debug-w5jvs" Oct 02 02:53:59 crc kubenswrapper[4885]: I1002 02:53:59.718781 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z2vq2\" (UniqueName: \"kubernetes.io/projected/b9f7878c-36a4-4cdf-9db7-c4e9d3548d74-kube-api-access-z2vq2\") pod \"crc-debug-w5jvs\" (UID: \"b9f7878c-36a4-4cdf-9db7-c4e9d3548d74\") " pod="openshift-must-gather-4xx9f/crc-debug-w5jvs" Oct 02 02:53:59 crc kubenswrapper[4885]: I1002 02:53:59.718845 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b9f7878c-36a4-4cdf-9db7-c4e9d3548d74-host\") pod \"crc-debug-w5jvs\" (UID: \"b9f7878c-36a4-4cdf-9db7-c4e9d3548d74\") " pod="openshift-must-gather-4xx9f/crc-debug-w5jvs" Oct 02 02:53:59 crc kubenswrapper[4885]: I1002 02:53:59.718985 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b9f7878c-36a4-4cdf-9db7-c4e9d3548d74-host\") pod \"crc-debug-w5jvs\" (UID: \"b9f7878c-36a4-4cdf-9db7-c4e9d3548d74\") " pod="openshift-must-gather-4xx9f/crc-debug-w5jvs" Oct 02 02:53:59 crc kubenswrapper[4885]: I1002 02:53:59.738240 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z2vq2\" (UniqueName: \"kubernetes.io/projected/b9f7878c-36a4-4cdf-9db7-c4e9d3548d74-kube-api-access-z2vq2\") pod \"crc-debug-w5jvs\" (UID: \"b9f7878c-36a4-4cdf-9db7-c4e9d3548d74\") " pod="openshift-must-gather-4xx9f/crc-debug-w5jvs" Oct 02 02:53:59 crc kubenswrapper[4885]: I1002 02:53:59.862199 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4xx9f/crc-debug-w5jvs" Oct 02 02:53:59 crc kubenswrapper[4885]: I1002 02:53:59.977032 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4xx9f/crc-debug-w5jvs" event={"ID":"b9f7878c-36a4-4cdf-9db7-c4e9d3548d74","Type":"ContainerStarted","Data":"8d541779741db0adf7522ed1f1a8e8b64619cb6a1a067c4131d08717ef574058"} Oct 02 02:54:00 crc kubenswrapper[4885]: I1002 02:54:00.986627 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4xx9f/crc-debug-w5jvs" event={"ID":"b9f7878c-36a4-4cdf-9db7-c4e9d3548d74","Type":"ContainerStarted","Data":"66c41f69fd3d9160849345a74f233f238af50c6c776d9b1ca8e63877e1f24181"} Oct 02 02:54:01 crc kubenswrapper[4885]: I1002 02:54:01.004072 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-4xx9f/crc-debug-w5jvs" podStartSLOduration=2.004055916 podStartE2EDuration="2.004055916s" podCreationTimestamp="2025-10-02 02:53:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:54:01.00288636 +0000 UTC m=+4029.814633759" watchObservedRunningTime="2025-10-02 02:54:01.004055916 +0000 UTC m=+4029.815803315" Oct 02 02:54:04 crc kubenswrapper[4885]: I1002 02:54:04.572916 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-wl5qp" Oct 02 02:54:04 crc kubenswrapper[4885]: I1002 02:54:04.625601 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wl5qp"] Oct 02 02:54:05 crc kubenswrapper[4885]: I1002 02:54:05.055220 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-wl5qp" podUID="2a53169d-5a5b-466c-988c-3e60ceef0111" containerName="registry-server" containerID="cri-o://2024120544b22f474f05d7b052cd999260fcba3f94e9daf154b9fdd6fcc3b883" gracePeriod=2 Oct 02 02:54:05 crc kubenswrapper[4885]: I1002 02:54:05.536572 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wl5qp" Oct 02 02:54:05 crc kubenswrapper[4885]: I1002 02:54:05.621913 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nttjj\" (UniqueName: \"kubernetes.io/projected/2a53169d-5a5b-466c-988c-3e60ceef0111-kube-api-access-nttjj\") pod \"2a53169d-5a5b-466c-988c-3e60ceef0111\" (UID: \"2a53169d-5a5b-466c-988c-3e60ceef0111\") " Oct 02 02:54:05 crc kubenswrapper[4885]: I1002 02:54:05.621990 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a53169d-5a5b-466c-988c-3e60ceef0111-catalog-content\") pod \"2a53169d-5a5b-466c-988c-3e60ceef0111\" (UID: \"2a53169d-5a5b-466c-988c-3e60ceef0111\") " Oct 02 02:54:05 crc kubenswrapper[4885]: I1002 02:54:05.622132 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a53169d-5a5b-466c-988c-3e60ceef0111-utilities\") pod \"2a53169d-5a5b-466c-988c-3e60ceef0111\" (UID: \"2a53169d-5a5b-466c-988c-3e60ceef0111\") " Oct 02 02:54:05 crc kubenswrapper[4885]: I1002 02:54:05.623507 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2a53169d-5a5b-466c-988c-3e60ceef0111-utilities" (OuterVolumeSpecName: "utilities") pod "2a53169d-5a5b-466c-988c-3e60ceef0111" (UID: "2a53169d-5a5b-466c-988c-3e60ceef0111"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:54:05 crc kubenswrapper[4885]: I1002 02:54:05.632479 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a53169d-5a5b-466c-988c-3e60ceef0111-kube-api-access-nttjj" (OuterVolumeSpecName: "kube-api-access-nttjj") pod "2a53169d-5a5b-466c-988c-3e60ceef0111" (UID: "2a53169d-5a5b-466c-988c-3e60ceef0111"). InnerVolumeSpecName "kube-api-access-nttjj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:54:05 crc kubenswrapper[4885]: I1002 02:54:05.645466 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2a53169d-5a5b-466c-988c-3e60ceef0111-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2a53169d-5a5b-466c-988c-3e60ceef0111" (UID: "2a53169d-5a5b-466c-988c-3e60ceef0111"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:54:05 crc kubenswrapper[4885]: I1002 02:54:05.741293 4885 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a53169d-5a5b-466c-988c-3e60ceef0111-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 02:54:05 crc kubenswrapper[4885]: I1002 02:54:05.741331 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nttjj\" (UniqueName: \"kubernetes.io/projected/2a53169d-5a5b-466c-988c-3e60ceef0111-kube-api-access-nttjj\") on node \"crc\" DevicePath \"\"" Oct 02 02:54:05 crc kubenswrapper[4885]: I1002 02:54:05.741369 4885 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a53169d-5a5b-466c-988c-3e60ceef0111-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 02:54:06 crc kubenswrapper[4885]: I1002 02:54:06.074800 4885 generic.go:334] "Generic (PLEG): container finished" podID="2a53169d-5a5b-466c-988c-3e60ceef0111" containerID="2024120544b22f474f05d7b052cd999260fcba3f94e9daf154b9fdd6fcc3b883" exitCode=0 Oct 02 02:54:06 crc kubenswrapper[4885]: I1002 02:54:06.074840 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wl5qp" event={"ID":"2a53169d-5a5b-466c-988c-3e60ceef0111","Type":"ContainerDied","Data":"2024120544b22f474f05d7b052cd999260fcba3f94e9daf154b9fdd6fcc3b883"} Oct 02 02:54:06 crc kubenswrapper[4885]: I1002 02:54:06.074864 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wl5qp" event={"ID":"2a53169d-5a5b-466c-988c-3e60ceef0111","Type":"ContainerDied","Data":"917b17f8d483d829ded496a9c30d7e0163c18f9f0b4046a2313564e28eadee9f"} Oct 02 02:54:06 crc kubenswrapper[4885]: I1002 02:54:06.074881 4885 scope.go:117] "RemoveContainer" containerID="2024120544b22f474f05d7b052cd999260fcba3f94e9daf154b9fdd6fcc3b883" Oct 02 02:54:06 crc kubenswrapper[4885]: I1002 02:54:06.075008 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wl5qp" Oct 02 02:54:06 crc kubenswrapper[4885]: I1002 02:54:06.100054 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wl5qp"] Oct 02 02:54:06 crc kubenswrapper[4885]: I1002 02:54:06.103847 4885 scope.go:117] "RemoveContainer" containerID="21b94a8f784a3a8bf61ee39c3a63a2d1f45f817f9bcdaceff0b769e041c71f66" Oct 02 02:54:06 crc kubenswrapper[4885]: I1002 02:54:06.106125 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-wl5qp"] Oct 02 02:54:06 crc kubenswrapper[4885]: I1002 02:54:06.126461 4885 scope.go:117] "RemoveContainer" containerID="cb415bc03dfcfa249ab84b81298dd8fded1699a33ef5c770bc05a8f2aae446a1" Oct 02 02:54:06 crc kubenswrapper[4885]: I1002 02:54:06.176485 4885 scope.go:117] "RemoveContainer" containerID="2024120544b22f474f05d7b052cd999260fcba3f94e9daf154b9fdd6fcc3b883" Oct 02 02:54:06 crc kubenswrapper[4885]: E1002 02:54:06.177118 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2024120544b22f474f05d7b052cd999260fcba3f94e9daf154b9fdd6fcc3b883\": container with ID starting with 2024120544b22f474f05d7b052cd999260fcba3f94e9daf154b9fdd6fcc3b883 not found: ID does not exist" containerID="2024120544b22f474f05d7b052cd999260fcba3f94e9daf154b9fdd6fcc3b883" Oct 02 02:54:06 crc kubenswrapper[4885]: I1002 02:54:06.177160 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2024120544b22f474f05d7b052cd999260fcba3f94e9daf154b9fdd6fcc3b883"} err="failed to get container status \"2024120544b22f474f05d7b052cd999260fcba3f94e9daf154b9fdd6fcc3b883\": rpc error: code = NotFound desc = could not find container \"2024120544b22f474f05d7b052cd999260fcba3f94e9daf154b9fdd6fcc3b883\": container with ID starting with 2024120544b22f474f05d7b052cd999260fcba3f94e9daf154b9fdd6fcc3b883 not found: ID does not exist" Oct 02 02:54:06 crc kubenswrapper[4885]: I1002 02:54:06.177192 4885 scope.go:117] "RemoveContainer" containerID="21b94a8f784a3a8bf61ee39c3a63a2d1f45f817f9bcdaceff0b769e041c71f66" Oct 02 02:54:06 crc kubenswrapper[4885]: E1002 02:54:06.177613 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"21b94a8f784a3a8bf61ee39c3a63a2d1f45f817f9bcdaceff0b769e041c71f66\": container with ID starting with 21b94a8f784a3a8bf61ee39c3a63a2d1f45f817f9bcdaceff0b769e041c71f66 not found: ID does not exist" containerID="21b94a8f784a3a8bf61ee39c3a63a2d1f45f817f9bcdaceff0b769e041c71f66" Oct 02 02:54:06 crc kubenswrapper[4885]: I1002 02:54:06.177636 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"21b94a8f784a3a8bf61ee39c3a63a2d1f45f817f9bcdaceff0b769e041c71f66"} err="failed to get container status \"21b94a8f784a3a8bf61ee39c3a63a2d1f45f817f9bcdaceff0b769e041c71f66\": rpc error: code = NotFound desc = could not find container \"21b94a8f784a3a8bf61ee39c3a63a2d1f45f817f9bcdaceff0b769e041c71f66\": container with ID starting with 21b94a8f784a3a8bf61ee39c3a63a2d1f45f817f9bcdaceff0b769e041c71f66 not found: ID does not exist" Oct 02 02:54:06 crc kubenswrapper[4885]: I1002 02:54:06.177683 4885 scope.go:117] "RemoveContainer" containerID="cb415bc03dfcfa249ab84b81298dd8fded1699a33ef5c770bc05a8f2aae446a1" Oct 02 02:54:06 crc kubenswrapper[4885]: E1002 02:54:06.178226 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cb415bc03dfcfa249ab84b81298dd8fded1699a33ef5c770bc05a8f2aae446a1\": container with ID starting with cb415bc03dfcfa249ab84b81298dd8fded1699a33ef5c770bc05a8f2aae446a1 not found: ID does not exist" containerID="cb415bc03dfcfa249ab84b81298dd8fded1699a33ef5c770bc05a8f2aae446a1" Oct 02 02:54:06 crc kubenswrapper[4885]: I1002 02:54:06.178248 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cb415bc03dfcfa249ab84b81298dd8fded1699a33ef5c770bc05a8f2aae446a1"} err="failed to get container status \"cb415bc03dfcfa249ab84b81298dd8fded1699a33ef5c770bc05a8f2aae446a1\": rpc error: code = NotFound desc = could not find container \"cb415bc03dfcfa249ab84b81298dd8fded1699a33ef5c770bc05a8f2aae446a1\": container with ID starting with cb415bc03dfcfa249ab84b81298dd8fded1699a33ef5c770bc05a8f2aae446a1 not found: ID does not exist" Oct 02 02:54:08 crc kubenswrapper[4885]: I1002 02:54:08.058508 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2a53169d-5a5b-466c-988c-3e60ceef0111" path="/var/lib/kubelet/pods/2a53169d-5a5b-466c-988c-3e60ceef0111/volumes" Oct 02 02:54:33 crc kubenswrapper[4885]: I1002 02:54:33.679158 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-r8v6j"] Oct 02 02:54:33 crc kubenswrapper[4885]: E1002 02:54:33.683803 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a53169d-5a5b-466c-988c-3e60ceef0111" containerName="extract-content" Oct 02 02:54:33 crc kubenswrapper[4885]: I1002 02:54:33.683910 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a53169d-5a5b-466c-988c-3e60ceef0111" containerName="extract-content" Oct 02 02:54:33 crc kubenswrapper[4885]: E1002 02:54:33.683996 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a53169d-5a5b-466c-988c-3e60ceef0111" containerName="extract-utilities" Oct 02 02:54:33 crc kubenswrapper[4885]: I1002 02:54:33.684070 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a53169d-5a5b-466c-988c-3e60ceef0111" containerName="extract-utilities" Oct 02 02:54:33 crc kubenswrapper[4885]: E1002 02:54:33.684157 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a53169d-5a5b-466c-988c-3e60ceef0111" containerName="registry-server" Oct 02 02:54:33 crc kubenswrapper[4885]: I1002 02:54:33.684229 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a53169d-5a5b-466c-988c-3e60ceef0111" containerName="registry-server" Oct 02 02:54:33 crc kubenswrapper[4885]: I1002 02:54:33.684615 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a53169d-5a5b-466c-988c-3e60ceef0111" containerName="registry-server" Oct 02 02:54:33 crc kubenswrapper[4885]: I1002 02:54:33.686556 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r8v6j" Oct 02 02:54:33 crc kubenswrapper[4885]: I1002 02:54:33.701780 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-r8v6j"] Oct 02 02:54:33 crc kubenswrapper[4885]: I1002 02:54:33.730490 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f31184c4-6951-42f6-9b16-ffac70bfbad0-utilities\") pod \"certified-operators-r8v6j\" (UID: \"f31184c4-6951-42f6-9b16-ffac70bfbad0\") " pod="openshift-marketplace/certified-operators-r8v6j" Oct 02 02:54:33 crc kubenswrapper[4885]: I1002 02:54:33.730798 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mg6ts\" (UniqueName: \"kubernetes.io/projected/f31184c4-6951-42f6-9b16-ffac70bfbad0-kube-api-access-mg6ts\") pod \"certified-operators-r8v6j\" (UID: \"f31184c4-6951-42f6-9b16-ffac70bfbad0\") " pod="openshift-marketplace/certified-operators-r8v6j" Oct 02 02:54:33 crc kubenswrapper[4885]: I1002 02:54:33.731180 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f31184c4-6951-42f6-9b16-ffac70bfbad0-catalog-content\") pod \"certified-operators-r8v6j\" (UID: \"f31184c4-6951-42f6-9b16-ffac70bfbad0\") " pod="openshift-marketplace/certified-operators-r8v6j" Oct 02 02:54:33 crc kubenswrapper[4885]: I1002 02:54:33.833385 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f31184c4-6951-42f6-9b16-ffac70bfbad0-utilities\") pod \"certified-operators-r8v6j\" (UID: \"f31184c4-6951-42f6-9b16-ffac70bfbad0\") " pod="openshift-marketplace/certified-operators-r8v6j" Oct 02 02:54:33 crc kubenswrapper[4885]: I1002 02:54:33.833450 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mg6ts\" (UniqueName: \"kubernetes.io/projected/f31184c4-6951-42f6-9b16-ffac70bfbad0-kube-api-access-mg6ts\") pod \"certified-operators-r8v6j\" (UID: \"f31184c4-6951-42f6-9b16-ffac70bfbad0\") " pod="openshift-marketplace/certified-operators-r8v6j" Oct 02 02:54:33 crc kubenswrapper[4885]: I1002 02:54:33.833483 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f31184c4-6951-42f6-9b16-ffac70bfbad0-catalog-content\") pod \"certified-operators-r8v6j\" (UID: \"f31184c4-6951-42f6-9b16-ffac70bfbad0\") " pod="openshift-marketplace/certified-operators-r8v6j" Oct 02 02:54:33 crc kubenswrapper[4885]: I1002 02:54:33.834029 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f31184c4-6951-42f6-9b16-ffac70bfbad0-catalog-content\") pod \"certified-operators-r8v6j\" (UID: \"f31184c4-6951-42f6-9b16-ffac70bfbad0\") " pod="openshift-marketplace/certified-operators-r8v6j" Oct 02 02:54:33 crc kubenswrapper[4885]: I1002 02:54:33.834350 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f31184c4-6951-42f6-9b16-ffac70bfbad0-utilities\") pod \"certified-operators-r8v6j\" (UID: \"f31184c4-6951-42f6-9b16-ffac70bfbad0\") " pod="openshift-marketplace/certified-operators-r8v6j" Oct 02 02:54:33 crc kubenswrapper[4885]: I1002 02:54:33.858667 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mg6ts\" (UniqueName: \"kubernetes.io/projected/f31184c4-6951-42f6-9b16-ffac70bfbad0-kube-api-access-mg6ts\") pod \"certified-operators-r8v6j\" (UID: \"f31184c4-6951-42f6-9b16-ffac70bfbad0\") " pod="openshift-marketplace/certified-operators-r8v6j" Oct 02 02:54:34 crc kubenswrapper[4885]: I1002 02:54:34.030074 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r8v6j" Oct 02 02:54:34 crc kubenswrapper[4885]: I1002 02:54:34.634185 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-r8v6j"] Oct 02 02:54:35 crc kubenswrapper[4885]: I1002 02:54:35.386317 4885 generic.go:334] "Generic (PLEG): container finished" podID="f31184c4-6951-42f6-9b16-ffac70bfbad0" containerID="b12f5d44188caeadb1eadfce91ed55220bd52c28ccf05161456e7aa827ee3e60" exitCode=0 Oct 02 02:54:35 crc kubenswrapper[4885]: I1002 02:54:35.386806 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r8v6j" event={"ID":"f31184c4-6951-42f6-9b16-ffac70bfbad0","Type":"ContainerDied","Data":"b12f5d44188caeadb1eadfce91ed55220bd52c28ccf05161456e7aa827ee3e60"} Oct 02 02:54:35 crc kubenswrapper[4885]: I1002 02:54:35.386834 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r8v6j" event={"ID":"f31184c4-6951-42f6-9b16-ffac70bfbad0","Type":"ContainerStarted","Data":"236247fcc81d9a69871f41ed2da42d681b038f7dd215d2f1964215f5cd42c6dc"} Oct 02 02:54:36 crc kubenswrapper[4885]: I1002 02:54:36.396970 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r8v6j" event={"ID":"f31184c4-6951-42f6-9b16-ffac70bfbad0","Type":"ContainerStarted","Data":"45b4def453395f4a92e0bc9e188804eee194d507be938f0069351dc07a7ba2db"} Oct 02 02:54:37 crc kubenswrapper[4885]: I1002 02:54:37.409076 4885 generic.go:334] "Generic (PLEG): container finished" podID="f31184c4-6951-42f6-9b16-ffac70bfbad0" containerID="45b4def453395f4a92e0bc9e188804eee194d507be938f0069351dc07a7ba2db" exitCode=0 Oct 02 02:54:37 crc kubenswrapper[4885]: I1002 02:54:37.409168 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r8v6j" event={"ID":"f31184c4-6951-42f6-9b16-ffac70bfbad0","Type":"ContainerDied","Data":"45b4def453395f4a92e0bc9e188804eee194d507be938f0069351dc07a7ba2db"} Oct 02 02:54:38 crc kubenswrapper[4885]: I1002 02:54:38.419351 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r8v6j" event={"ID":"f31184c4-6951-42f6-9b16-ffac70bfbad0","Type":"ContainerStarted","Data":"dbc52a02ce790de74b3b107fe99144b2ed114930e86da4ff3d0eb976f6327c2d"} Oct 02 02:54:38 crc kubenswrapper[4885]: I1002 02:54:38.437922 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-r8v6j" podStartSLOduration=2.952052722 podStartE2EDuration="5.437907329s" podCreationTimestamp="2025-10-02 02:54:33 +0000 UTC" firstStartedPulling="2025-10-02 02:54:35.390237421 +0000 UTC m=+4064.201984820" lastFinishedPulling="2025-10-02 02:54:37.876092028 +0000 UTC m=+4066.687839427" observedRunningTime="2025-10-02 02:54:38.434927288 +0000 UTC m=+4067.246674687" watchObservedRunningTime="2025-10-02 02:54:38.437907329 +0000 UTC m=+4067.249654728" Oct 02 02:54:43 crc kubenswrapper[4885]: I1002 02:54:43.265241 4885 patch_prober.go:28] interesting pod/machine-config-daemon-rttx8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 02:54:43 crc kubenswrapper[4885]: I1002 02:54:43.265686 4885 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 02:54:44 crc kubenswrapper[4885]: I1002 02:54:44.030360 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-r8v6j" Oct 02 02:54:44 crc kubenswrapper[4885]: I1002 02:54:44.031361 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-r8v6j" Oct 02 02:54:44 crc kubenswrapper[4885]: I1002 02:54:44.085494 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-r8v6j" Oct 02 02:54:44 crc kubenswrapper[4885]: I1002 02:54:44.543738 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-r8v6j" Oct 02 02:54:44 crc kubenswrapper[4885]: I1002 02:54:44.585490 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-r8v6j"] Oct 02 02:54:46 crc kubenswrapper[4885]: I1002 02:54:46.513731 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-r8v6j" podUID="f31184c4-6951-42f6-9b16-ffac70bfbad0" containerName="registry-server" containerID="cri-o://dbc52a02ce790de74b3b107fe99144b2ed114930e86da4ff3d0eb976f6327c2d" gracePeriod=2 Oct 02 02:54:47 crc kubenswrapper[4885]: I1002 02:54:47.483134 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r8v6j" Oct 02 02:54:47 crc kubenswrapper[4885]: I1002 02:54:47.534564 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg6ts\" (UniqueName: \"kubernetes.io/projected/f31184c4-6951-42f6-9b16-ffac70bfbad0-kube-api-access-mg6ts\") pod \"f31184c4-6951-42f6-9b16-ffac70bfbad0\" (UID: \"f31184c4-6951-42f6-9b16-ffac70bfbad0\") " Oct 02 02:54:47 crc kubenswrapper[4885]: I1002 02:54:47.534647 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f31184c4-6951-42f6-9b16-ffac70bfbad0-catalog-content\") pod \"f31184c4-6951-42f6-9b16-ffac70bfbad0\" (UID: \"f31184c4-6951-42f6-9b16-ffac70bfbad0\") " Oct 02 02:54:47 crc kubenswrapper[4885]: I1002 02:54:47.534943 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f31184c4-6951-42f6-9b16-ffac70bfbad0-utilities\") pod \"f31184c4-6951-42f6-9b16-ffac70bfbad0\" (UID: \"f31184c4-6951-42f6-9b16-ffac70bfbad0\") " Oct 02 02:54:47 crc kubenswrapper[4885]: I1002 02:54:47.538386 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f31184c4-6951-42f6-9b16-ffac70bfbad0-utilities" (OuterVolumeSpecName: "utilities") pod "f31184c4-6951-42f6-9b16-ffac70bfbad0" (UID: "f31184c4-6951-42f6-9b16-ffac70bfbad0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:54:47 crc kubenswrapper[4885]: I1002 02:54:47.538810 4885 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f31184c4-6951-42f6-9b16-ffac70bfbad0-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 02:54:47 crc kubenswrapper[4885]: I1002 02:54:47.550589 4885 generic.go:334] "Generic (PLEG): container finished" podID="f31184c4-6951-42f6-9b16-ffac70bfbad0" containerID="dbc52a02ce790de74b3b107fe99144b2ed114930e86da4ff3d0eb976f6327c2d" exitCode=0 Oct 02 02:54:47 crc kubenswrapper[4885]: I1002 02:54:47.550640 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r8v6j" event={"ID":"f31184c4-6951-42f6-9b16-ffac70bfbad0","Type":"ContainerDied","Data":"dbc52a02ce790de74b3b107fe99144b2ed114930e86da4ff3d0eb976f6327c2d"} Oct 02 02:54:47 crc kubenswrapper[4885]: I1002 02:54:47.550672 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r8v6j" event={"ID":"f31184c4-6951-42f6-9b16-ffac70bfbad0","Type":"ContainerDied","Data":"236247fcc81d9a69871f41ed2da42d681b038f7dd215d2f1964215f5cd42c6dc"} Oct 02 02:54:47 crc kubenswrapper[4885]: I1002 02:54:47.550693 4885 scope.go:117] "RemoveContainer" containerID="dbc52a02ce790de74b3b107fe99144b2ed114930e86da4ff3d0eb976f6327c2d" Oct 02 02:54:47 crc kubenswrapper[4885]: I1002 02:54:47.550871 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r8v6j" Oct 02 02:54:47 crc kubenswrapper[4885]: I1002 02:54:47.560534 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f31184c4-6951-42f6-9b16-ffac70bfbad0-kube-api-access-mg6ts" (OuterVolumeSpecName: "kube-api-access-mg6ts") pod "f31184c4-6951-42f6-9b16-ffac70bfbad0" (UID: "f31184c4-6951-42f6-9b16-ffac70bfbad0"). InnerVolumeSpecName "kube-api-access-mg6ts". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:54:47 crc kubenswrapper[4885]: I1002 02:54:47.604309 4885 scope.go:117] "RemoveContainer" containerID="45b4def453395f4a92e0bc9e188804eee194d507be938f0069351dc07a7ba2db" Oct 02 02:54:47 crc kubenswrapper[4885]: I1002 02:54:47.609998 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f31184c4-6951-42f6-9b16-ffac70bfbad0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f31184c4-6951-42f6-9b16-ffac70bfbad0" (UID: "f31184c4-6951-42f6-9b16-ffac70bfbad0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:54:47 crc kubenswrapper[4885]: I1002 02:54:47.640471 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg6ts\" (UniqueName: \"kubernetes.io/projected/f31184c4-6951-42f6-9b16-ffac70bfbad0-kube-api-access-mg6ts\") on node \"crc\" DevicePath \"\"" Oct 02 02:54:47 crc kubenswrapper[4885]: I1002 02:54:47.640675 4885 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f31184c4-6951-42f6-9b16-ffac70bfbad0-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 02:54:47 crc kubenswrapper[4885]: I1002 02:54:47.671226 4885 scope.go:117] "RemoveContainer" containerID="b12f5d44188caeadb1eadfce91ed55220bd52c28ccf05161456e7aa827ee3e60" Oct 02 02:54:47 crc kubenswrapper[4885]: I1002 02:54:47.879415 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-r8v6j"] Oct 02 02:54:47 crc kubenswrapper[4885]: I1002 02:54:47.886270 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-r8v6j"] Oct 02 02:54:48 crc kubenswrapper[4885]: I1002 02:54:48.071173 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f31184c4-6951-42f6-9b16-ffac70bfbad0" path="/var/lib/kubelet/pods/f31184c4-6951-42f6-9b16-ffac70bfbad0/volumes" Oct 02 02:54:48 crc kubenswrapper[4885]: I1002 02:54:48.103686 4885 scope.go:117] "RemoveContainer" containerID="dbc52a02ce790de74b3b107fe99144b2ed114930e86da4ff3d0eb976f6327c2d" Oct 02 02:54:48 crc kubenswrapper[4885]: E1002 02:54:48.106871 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dbc52a02ce790de74b3b107fe99144b2ed114930e86da4ff3d0eb976f6327c2d\": container with ID starting with dbc52a02ce790de74b3b107fe99144b2ed114930e86da4ff3d0eb976f6327c2d not found: ID does not exist" containerID="dbc52a02ce790de74b3b107fe99144b2ed114930e86da4ff3d0eb976f6327c2d" Oct 02 02:54:48 crc kubenswrapper[4885]: I1002 02:54:48.106918 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dbc52a02ce790de74b3b107fe99144b2ed114930e86da4ff3d0eb976f6327c2d"} err="failed to get container status \"dbc52a02ce790de74b3b107fe99144b2ed114930e86da4ff3d0eb976f6327c2d\": rpc error: code = NotFound desc = could not find container \"dbc52a02ce790de74b3b107fe99144b2ed114930e86da4ff3d0eb976f6327c2d\": container with ID starting with dbc52a02ce790de74b3b107fe99144b2ed114930e86da4ff3d0eb976f6327c2d not found: ID does not exist" Oct 02 02:54:48 crc kubenswrapper[4885]: I1002 02:54:48.106943 4885 scope.go:117] "RemoveContainer" containerID="45b4def453395f4a92e0bc9e188804eee194d507be938f0069351dc07a7ba2db" Oct 02 02:54:48 crc kubenswrapper[4885]: E1002 02:54:48.108430 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"45b4def453395f4a92e0bc9e188804eee194d507be938f0069351dc07a7ba2db\": container with ID starting with 45b4def453395f4a92e0bc9e188804eee194d507be938f0069351dc07a7ba2db not found: ID does not exist" containerID="45b4def453395f4a92e0bc9e188804eee194d507be938f0069351dc07a7ba2db" Oct 02 02:54:48 crc kubenswrapper[4885]: I1002 02:54:48.108498 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"45b4def453395f4a92e0bc9e188804eee194d507be938f0069351dc07a7ba2db"} err="failed to get container status \"45b4def453395f4a92e0bc9e188804eee194d507be938f0069351dc07a7ba2db\": rpc error: code = NotFound desc = could not find container \"45b4def453395f4a92e0bc9e188804eee194d507be938f0069351dc07a7ba2db\": container with ID starting with 45b4def453395f4a92e0bc9e188804eee194d507be938f0069351dc07a7ba2db not found: ID does not exist" Oct 02 02:54:48 crc kubenswrapper[4885]: I1002 02:54:48.108543 4885 scope.go:117] "RemoveContainer" containerID="b12f5d44188caeadb1eadfce91ed55220bd52c28ccf05161456e7aa827ee3e60" Oct 02 02:54:48 crc kubenswrapper[4885]: E1002 02:54:48.113415 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b12f5d44188caeadb1eadfce91ed55220bd52c28ccf05161456e7aa827ee3e60\": container with ID starting with b12f5d44188caeadb1eadfce91ed55220bd52c28ccf05161456e7aa827ee3e60 not found: ID does not exist" containerID="b12f5d44188caeadb1eadfce91ed55220bd52c28ccf05161456e7aa827ee3e60" Oct 02 02:54:48 crc kubenswrapper[4885]: I1002 02:54:48.113480 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b12f5d44188caeadb1eadfce91ed55220bd52c28ccf05161456e7aa827ee3e60"} err="failed to get container status \"b12f5d44188caeadb1eadfce91ed55220bd52c28ccf05161456e7aa827ee3e60\": rpc error: code = NotFound desc = could not find container \"b12f5d44188caeadb1eadfce91ed55220bd52c28ccf05161456e7aa827ee3e60\": container with ID starting with b12f5d44188caeadb1eadfce91ed55220bd52c28ccf05161456e7aa827ee3e60 not found: ID does not exist" Oct 02 02:55:02 crc kubenswrapper[4885]: I1002 02:55:02.471723 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-6b5d8cdf54-gr28p_af70565f-bc37-43bc-97a3-3942c35a9fb2/barbican-api-log/0.log" Oct 02 02:55:02 crc kubenswrapper[4885]: I1002 02:55:02.548798 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-6b5d8cdf54-gr28p_af70565f-bc37-43bc-97a3-3942c35a9fb2/barbican-api/0.log" Oct 02 02:55:02 crc kubenswrapper[4885]: I1002 02:55:02.719035 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-6b9ddcb5cd-2r5wd_c416d7da-12d2-4aee-a9c3-5f171a791b43/barbican-keystone-listener/0.log" Oct 02 02:55:02 crc kubenswrapper[4885]: I1002 02:55:02.740330 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-6b9ddcb5cd-2r5wd_c416d7da-12d2-4aee-a9c3-5f171a791b43/barbican-keystone-listener-log/0.log" Oct 02 02:55:02 crc kubenswrapper[4885]: I1002 02:55:02.877971 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5c6b9856b7-vhzx6_075adfcf-950d-44e6-ab5e-826f3b06c05e/barbican-worker/0.log" Oct 02 02:55:02 crc kubenswrapper[4885]: I1002 02:55:02.946475 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5c6b9856b7-vhzx6_075adfcf-950d-44e6-ab5e-826f3b06c05e/barbican-worker-log/0.log" Oct 02 02:55:03 crc kubenswrapper[4885]: I1002 02:55:03.089130 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-rgbbj_8cba1665-9cca-44ef-a790-5cd9b174e726/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 02:55:03 crc kubenswrapper[4885]: I1002 02:55:03.265452 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_0fab80f5-cc23-4ed6-8c0b-1604d9809341/ceilometer-notification-agent/0.log" Oct 02 02:55:03 crc kubenswrapper[4885]: I1002 02:55:03.307446 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_0fab80f5-cc23-4ed6-8c0b-1604d9809341/ceilometer-central-agent/0.log" Oct 02 02:55:03 crc kubenswrapper[4885]: I1002 02:55:03.317138 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_0fab80f5-cc23-4ed6-8c0b-1604d9809341/proxy-httpd/0.log" Oct 02 02:55:03 crc kubenswrapper[4885]: I1002 02:55:03.417733 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_0fab80f5-cc23-4ed6-8c0b-1604d9809341/sg-core/0.log" Oct 02 02:55:03 crc kubenswrapper[4885]: I1002 02:55:03.562306 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_ca40a36d-76b5-4cfe-8f8c-ccb4625d6f56/cinder-api/0.log" Oct 02 02:55:03 crc kubenswrapper[4885]: I1002 02:55:03.673429 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_ca40a36d-76b5-4cfe-8f8c-ccb4625d6f56/cinder-api-log/0.log" Oct 02 02:55:03 crc kubenswrapper[4885]: I1002 02:55:03.810209 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_0609d820-e495-4845-a50f-166717c55f93/cinder-scheduler/0.log" Oct 02 02:55:03 crc kubenswrapper[4885]: I1002 02:55:03.878463 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_0609d820-e495-4845-a50f-166717c55f93/probe/0.log" Oct 02 02:55:03 crc kubenswrapper[4885]: I1002 02:55:03.990580 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-zlzq7_fd55e079-473a-43d3-aca2-ee0d91c06aca/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 02:55:04 crc kubenswrapper[4885]: I1002 02:55:04.120248 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-mv2tw_c6157c99-0cbf-4edb-9a16-04c99b7b2385/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 02:55:04 crc kubenswrapper[4885]: I1002 02:55:04.229020 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-wvh9j_93d5499a-81ec-4dbc-b769-94d3e8690fbe/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 02:55:04 crc kubenswrapper[4885]: I1002 02:55:04.422205 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-cb6ffcf87-46bgc_58f621da-0934-482b-8f10-4273021207f0/init/0.log" Oct 02 02:55:05 crc kubenswrapper[4885]: I1002 02:55:05.046978 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-cb6ffcf87-46bgc_58f621da-0934-482b-8f10-4273021207f0/init/0.log" Oct 02 02:55:05 crc kubenswrapper[4885]: I1002 02:55:05.120073 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-cb6ffcf87-46bgc_58f621da-0934-482b-8f10-4273021207f0/dnsmasq-dns/0.log" Oct 02 02:55:05 crc kubenswrapper[4885]: I1002 02:55:05.279383 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-c2g9p_c1ca4902-a7b7-4cfa-a8df-5953754142e0/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 02:55:05 crc kubenswrapper[4885]: I1002 02:55:05.316793 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_d6e0fd15-e949-444a-86e2-894963c1267d/glance-httpd/0.log" Oct 02 02:55:05 crc kubenswrapper[4885]: I1002 02:55:05.378108 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_d6e0fd15-e949-444a-86e2-894963c1267d/glance-log/0.log" Oct 02 02:55:05 crc kubenswrapper[4885]: I1002 02:55:05.499488 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_511f1564-a39d-48be-9c35-ed3343e4650c/glance-httpd/0.log" Oct 02 02:55:05 crc kubenswrapper[4885]: I1002 02:55:05.554709 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_511f1564-a39d-48be-9c35-ed3343e4650c/glance-log/0.log" Oct 02 02:55:05 crc kubenswrapper[4885]: I1002 02:55:05.806704 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-6d7c47744-vjcbs_607de89d-b68e-49e8-beb4-7664a37c6105/horizon/0.log" Oct 02 02:55:05 crc kubenswrapper[4885]: I1002 02:55:05.850500 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-prq96_bd405392-0e1a-42c5-a81a-cab2d229ea55/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 02:55:06 crc kubenswrapper[4885]: I1002 02:55:06.147680 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-p9p9g_48ef0e0a-ae52-4d11-9023-d53f4ab716ac/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 02:55:06 crc kubenswrapper[4885]: I1002 02:55:06.154511 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-6d7c47744-vjcbs_607de89d-b68e-49e8-beb4-7664a37c6105/horizon-log/0.log" Oct 02 02:55:06 crc kubenswrapper[4885]: I1002 02:55:06.748673 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_b5cb9a09-d2c8-4138-9eea-f404bd2e5034/kube-state-metrics/0.log" Oct 02 02:55:06 crc kubenswrapper[4885]: I1002 02:55:06.883180 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-59f7bf489d-7tfk4_5f5ca9c7-0cfd-4c95-a160-4d909f8f3d00/keystone-api/0.log" Oct 02 02:55:07 crc kubenswrapper[4885]: I1002 02:55:07.015551 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-x47dm_54cf89d0-a777-4ac9-968b-49836d0b13b1/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 02:55:07 crc kubenswrapper[4885]: I1002 02:55:07.340534 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-684bd7f7b9-p7lvf_ddb14564-1e3d-43d7-9c9d-2665b53124b6/neutron-api/0.log" Oct 02 02:55:07 crc kubenswrapper[4885]: I1002 02:55:07.342373 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-684bd7f7b9-p7lvf_ddb14564-1e3d-43d7-9c9d-2665b53124b6/neutron-httpd/0.log" Oct 02 02:55:07 crc kubenswrapper[4885]: I1002 02:55:07.572043 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-2hp4t_f6edca36-b630-4682-93ed-2c0c2dcadecb/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 02:55:08 crc kubenswrapper[4885]: I1002 02:55:08.037808 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_71959149-27b9-4b03-aa11-1c2164677d5c/nova-api-log/0.log" Oct 02 02:55:08 crc kubenswrapper[4885]: I1002 02:55:08.220449 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_504d46bb-b842-4362-9a00-4917a01cf509/nova-cell0-conductor-conductor/0.log" Oct 02 02:55:08 crc kubenswrapper[4885]: I1002 02:55:08.386893 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_71959149-27b9-4b03-aa11-1c2164677d5c/nova-api-api/0.log" Oct 02 02:55:08 crc kubenswrapper[4885]: I1002 02:55:08.626051 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_b6f37337-ce77-444d-9efa-852d0db3aabb/nova-cell1-conductor-conductor/0.log" Oct 02 02:55:08 crc kubenswrapper[4885]: I1002 02:55:08.734122 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_f6249fa5-32a1-4353-a521-67d1fde191d4/nova-cell1-novncproxy-novncproxy/0.log" Oct 02 02:55:08 crc kubenswrapper[4885]: I1002 02:55:08.899302 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-c6gw4_14bdd01b-529b-45b4-9ac3-188fdf3d3855/nova-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 02:55:09 crc kubenswrapper[4885]: I1002 02:55:09.017317 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_179284e1-8a2f-4987-a0b4-7b24b6e3cf94/nova-metadata-log/0.log" Oct 02 02:55:09 crc kubenswrapper[4885]: I1002 02:55:09.450295 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_50278f9c-b6fc-437c-ba3a-972b857493e3/nova-scheduler-scheduler/0.log" Oct 02 02:55:09 crc kubenswrapper[4885]: I1002 02:55:09.585125 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_87196135-8d40-47cb-b918-98aea04c7c17/mysql-bootstrap/0.log" Oct 02 02:55:09 crc kubenswrapper[4885]: I1002 02:55:09.793561 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_87196135-8d40-47cb-b918-98aea04c7c17/mysql-bootstrap/0.log" Oct 02 02:55:09 crc kubenswrapper[4885]: I1002 02:55:09.849127 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_87196135-8d40-47cb-b918-98aea04c7c17/galera/0.log" Oct 02 02:55:10 crc kubenswrapper[4885]: I1002 02:55:10.031433 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_8427e298-965b-48f0-821f-bd078fa2e96f/mysql-bootstrap/0.log" Oct 02 02:55:10 crc kubenswrapper[4885]: I1002 02:55:10.265845 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_8427e298-965b-48f0-821f-bd078fa2e96f/galera/0.log" Oct 02 02:55:10 crc kubenswrapper[4885]: I1002 02:55:10.269709 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_8427e298-965b-48f0-821f-bd078fa2e96f/mysql-bootstrap/0.log" Oct 02 02:55:10 crc kubenswrapper[4885]: I1002 02:55:10.509407 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_179284e1-8a2f-4987-a0b4-7b24b6e3cf94/nova-metadata-metadata/0.log" Oct 02 02:55:10 crc kubenswrapper[4885]: I1002 02:55:10.518002 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_ab251703-d0d6-49cf-bd2a-b74dee9e48f7/openstackclient/0.log" Oct 02 02:55:10 crc kubenswrapper[4885]: I1002 02:55:10.725012 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-fgvpt_924fb321-4073-449a-a546-811b046ed26a/ovn-controller/0.log" Oct 02 02:55:10 crc kubenswrapper[4885]: I1002 02:55:10.901295 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-sphmn_eb166198-03b5-458f-b86b-5706afa88d7b/openstack-network-exporter/0.log" Oct 02 02:55:11 crc kubenswrapper[4885]: I1002 02:55:11.019895 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-q92rb_b65df3c4-cc33-43da-9bf1-8f46a480f7eb/ovsdb-server-init/0.log" Oct 02 02:55:11 crc kubenswrapper[4885]: I1002 02:55:11.184007 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-q92rb_b65df3c4-cc33-43da-9bf1-8f46a480f7eb/ovsdb-server-init/0.log" Oct 02 02:55:11 crc kubenswrapper[4885]: I1002 02:55:11.230919 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-q92rb_b65df3c4-cc33-43da-9bf1-8f46a480f7eb/ovsdb-server/0.log" Oct 02 02:55:11 crc kubenswrapper[4885]: I1002 02:55:11.253233 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-q92rb_b65df3c4-cc33-43da-9bf1-8f46a480f7eb/ovs-vswitchd/0.log" Oct 02 02:55:11 crc kubenswrapper[4885]: I1002 02:55:11.430091 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-xkwk4_d8d7cf22-9bf7-417b-b4e2-d199cfccf949/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 02:55:11 crc kubenswrapper[4885]: I1002 02:55:11.594981 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_d02dcd41-585a-411f-ace9-97c1a662fdc8/openstack-network-exporter/0.log" Oct 02 02:55:11 crc kubenswrapper[4885]: I1002 02:55:11.623212 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_d02dcd41-585a-411f-ace9-97c1a662fdc8/ovn-northd/0.log" Oct 02 02:55:11 crc kubenswrapper[4885]: I1002 02:55:11.791339 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_079e125d-c938-4962-8f41-09b3109fb0ae/ovsdbserver-nb/0.log" Oct 02 02:55:11 crc kubenswrapper[4885]: I1002 02:55:11.817622 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_079e125d-c938-4962-8f41-09b3109fb0ae/openstack-network-exporter/0.log" Oct 02 02:55:11 crc kubenswrapper[4885]: I1002 02:55:11.987691 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_33286960-496c-4f14-a334-c6dc765ca956/openstack-network-exporter/0.log" Oct 02 02:55:12 crc kubenswrapper[4885]: I1002 02:55:12.027568 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_33286960-496c-4f14-a334-c6dc765ca956/ovsdbserver-sb/0.log" Oct 02 02:55:12 crc kubenswrapper[4885]: I1002 02:55:12.244627 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-5d65d69454-8bzbn_9bcc69ce-fb1a-4265-a6de-b2e2a10ac733/placement-api/0.log" Oct 02 02:55:12 crc kubenswrapper[4885]: I1002 02:55:12.314575 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-5d65d69454-8bzbn_9bcc69ce-fb1a-4265-a6de-b2e2a10ac733/placement-log/0.log" Oct 02 02:55:12 crc kubenswrapper[4885]: I1002 02:55:12.443796 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_79f97e3d-d825-4982-9f76-91089df7a67b/setup-container/0.log" Oct 02 02:55:12 crc kubenswrapper[4885]: I1002 02:55:12.641239 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_79f97e3d-d825-4982-9f76-91089df7a67b/rabbitmq/0.log" Oct 02 02:55:12 crc kubenswrapper[4885]: I1002 02:55:12.647587 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_79f97e3d-d825-4982-9f76-91089df7a67b/setup-container/0.log" Oct 02 02:55:12 crc kubenswrapper[4885]: I1002 02:55:12.797917 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_18642896-a2b6-412c-beb1-8010484744a4/setup-container/0.log" Oct 02 02:55:12 crc kubenswrapper[4885]: I1002 02:55:12.987248 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_18642896-a2b6-412c-beb1-8010484744a4/rabbitmq/0.log" Oct 02 02:55:13 crc kubenswrapper[4885]: I1002 02:55:13.009375 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_18642896-a2b6-412c-beb1-8010484744a4/setup-container/0.log" Oct 02 02:55:13 crc kubenswrapper[4885]: I1002 02:55:13.136941 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-2rt6h_2d16884c-61ae-4288-b629-2a7c3be23ea8/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 02:55:13 crc kubenswrapper[4885]: I1002 02:55:13.265349 4885 patch_prober.go:28] interesting pod/machine-config-daemon-rttx8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 02:55:13 crc kubenswrapper[4885]: I1002 02:55:13.265406 4885 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 02:55:13 crc kubenswrapper[4885]: I1002 02:55:13.818198 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-6nvvk_379db8cf-874f-4d6e-9e3c-795457fdddb9/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 02:55:13 crc kubenswrapper[4885]: I1002 02:55:13.868009 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-vmf4j_6602e994-3938-456d-8bb3-e092b6350c8a/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 02:55:14 crc kubenswrapper[4885]: I1002 02:55:14.092721 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-jnpmf_3093cd67-4ea2-4256-bcf4-7e2533870ba6/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 02:55:14 crc kubenswrapper[4885]: I1002 02:55:14.228758 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-njjfc_e6dc3bce-d7cf-4ce8-9577-dd1f6813bff1/ssh-known-hosts-edpm-deployment/0.log" Oct 02 02:55:14 crc kubenswrapper[4885]: I1002 02:55:14.404429 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-7474b946b7-s2w7j_3b22bebc-2f6c-4981-8166-4ccaf19f9727/proxy-httpd/0.log" Oct 02 02:55:14 crc kubenswrapper[4885]: I1002 02:55:14.485957 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-7474b946b7-s2w7j_3b22bebc-2f6c-4981-8166-4ccaf19f9727/proxy-server/0.log" Oct 02 02:55:14 crc kubenswrapper[4885]: I1002 02:55:14.610583 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-9wd49_2d471d2b-912c-4f49-a7ec-c07b85c02bd9/swift-ring-rebalance/0.log" Oct 02 02:55:14 crc kubenswrapper[4885]: I1002 02:55:14.762461 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0c8d183a-f48b-4a69-8685-90c834e678ce/account-auditor/0.log" Oct 02 02:55:14 crc kubenswrapper[4885]: I1002 02:55:14.810318 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0c8d183a-f48b-4a69-8685-90c834e678ce/account-reaper/0.log" Oct 02 02:55:14 crc kubenswrapper[4885]: I1002 02:55:14.952020 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0c8d183a-f48b-4a69-8685-90c834e678ce/account-replicator/0.log" Oct 02 02:55:14 crc kubenswrapper[4885]: I1002 02:55:14.971046 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0c8d183a-f48b-4a69-8685-90c834e678ce/account-server/0.log" Oct 02 02:55:15 crc kubenswrapper[4885]: I1002 02:55:15.011706 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0c8d183a-f48b-4a69-8685-90c834e678ce/container-auditor/0.log" Oct 02 02:55:15 crc kubenswrapper[4885]: I1002 02:55:15.152967 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0c8d183a-f48b-4a69-8685-90c834e678ce/container-server/0.log" Oct 02 02:55:15 crc kubenswrapper[4885]: I1002 02:55:15.157537 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0c8d183a-f48b-4a69-8685-90c834e678ce/container-replicator/0.log" Oct 02 02:55:15 crc kubenswrapper[4885]: I1002 02:55:15.193994 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0c8d183a-f48b-4a69-8685-90c834e678ce/container-updater/0.log" Oct 02 02:55:15 crc kubenswrapper[4885]: I1002 02:55:15.741622 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0c8d183a-f48b-4a69-8685-90c834e678ce/object-expirer/0.log" Oct 02 02:55:15 crc kubenswrapper[4885]: I1002 02:55:15.748429 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0c8d183a-f48b-4a69-8685-90c834e678ce/object-auditor/0.log" Oct 02 02:55:15 crc kubenswrapper[4885]: I1002 02:55:15.787987 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0c8d183a-f48b-4a69-8685-90c834e678ce/object-replicator/0.log" Oct 02 02:55:15 crc kubenswrapper[4885]: I1002 02:55:15.890469 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0c8d183a-f48b-4a69-8685-90c834e678ce/object-server/0.log" Oct 02 02:55:15 crc kubenswrapper[4885]: I1002 02:55:15.979557 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0c8d183a-f48b-4a69-8685-90c834e678ce/object-updater/0.log" Oct 02 02:55:15 crc kubenswrapper[4885]: I1002 02:55:15.984818 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0c8d183a-f48b-4a69-8685-90c834e678ce/rsync/0.log" Oct 02 02:55:16 crc kubenswrapper[4885]: I1002 02:55:16.053820 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0c8d183a-f48b-4a69-8685-90c834e678ce/swift-recon-cron/0.log" Oct 02 02:55:16 crc kubenswrapper[4885]: I1002 02:55:16.249963 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-v9h2z_ce63c1c1-34cc-4cba-a553-33b66431bc3c/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 02:55:16 crc kubenswrapper[4885]: I1002 02:55:16.436753 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_7a715636-06d0-4508-a9d0-fe83f4c05d01/tempest-tests-tempest-tests-runner/0.log" Oct 02 02:55:16 crc kubenswrapper[4885]: I1002 02:55:16.518890 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_9cf08307-1f33-4989-b935-67c8f4a76cc4/test-operator-logs-container/0.log" Oct 02 02:55:16 crc kubenswrapper[4885]: I1002 02:55:16.701306 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-5l99v_829f7f0e-4288-42a2-b585-99beef8cf451/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 02:55:24 crc kubenswrapper[4885]: I1002 02:55:24.960619 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_d798d658-f60b-4e23-8753-b3de580e44b0/memcached/0.log" Oct 02 02:55:43 crc kubenswrapper[4885]: I1002 02:55:43.265997 4885 patch_prober.go:28] interesting pod/machine-config-daemon-rttx8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 02:55:43 crc kubenswrapper[4885]: I1002 02:55:43.266665 4885 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 02:55:43 crc kubenswrapper[4885]: I1002 02:55:43.266749 4885 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" Oct 02 02:55:43 crc kubenswrapper[4885]: I1002 02:55:43.267700 4885 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"514bc2a11d8477ee5f25c8b041d1d9f8fe40c82d21890ace02ef6041f3710407"} pod="openshift-machine-config-operator/machine-config-daemon-rttx8" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 02:55:43 crc kubenswrapper[4885]: I1002 02:55:43.267801 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" containerName="machine-config-daemon" containerID="cri-o://514bc2a11d8477ee5f25c8b041d1d9f8fe40c82d21890ace02ef6041f3710407" gracePeriod=600 Oct 02 02:55:44 crc kubenswrapper[4885]: I1002 02:55:44.108177 4885 generic.go:334] "Generic (PLEG): container finished" podID="d169c8ba-000a-4218-8e3b-4ae53035b110" containerID="514bc2a11d8477ee5f25c8b041d1d9f8fe40c82d21890ace02ef6041f3710407" exitCode=0 Oct 02 02:55:44 crc kubenswrapper[4885]: I1002 02:55:44.108248 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" event={"ID":"d169c8ba-000a-4218-8e3b-4ae53035b110","Type":"ContainerDied","Data":"514bc2a11d8477ee5f25c8b041d1d9f8fe40c82d21890ace02ef6041f3710407"} Oct 02 02:55:44 crc kubenswrapper[4885]: I1002 02:55:44.108940 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" event={"ID":"d169c8ba-000a-4218-8e3b-4ae53035b110","Type":"ContainerStarted","Data":"e190b22b6f36adea43879419bc87517eb0762fd9a17a1bf8203497c99cfb5a7f"} Oct 02 02:55:44 crc kubenswrapper[4885]: I1002 02:55:44.108979 4885 scope.go:117] "RemoveContainer" containerID="999b3e75477511af3aa11f1e9d932c1ffb739b6f215cc19f45ac7144051ab65c" Oct 02 02:55:49 crc kubenswrapper[4885]: I1002 02:55:49.188199 4885 generic.go:334] "Generic (PLEG): container finished" podID="b9f7878c-36a4-4cdf-9db7-c4e9d3548d74" containerID="66c41f69fd3d9160849345a74f233f238af50c6c776d9b1ca8e63877e1f24181" exitCode=0 Oct 02 02:55:49 crc kubenswrapper[4885]: I1002 02:55:49.188328 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4xx9f/crc-debug-w5jvs" event={"ID":"b9f7878c-36a4-4cdf-9db7-c4e9d3548d74","Type":"ContainerDied","Data":"66c41f69fd3d9160849345a74f233f238af50c6c776d9b1ca8e63877e1f24181"} Oct 02 02:55:50 crc kubenswrapper[4885]: I1002 02:55:50.332313 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4xx9f/crc-debug-w5jvs" Oct 02 02:55:50 crc kubenswrapper[4885]: I1002 02:55:50.399514 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-4xx9f/crc-debug-w5jvs"] Oct 02 02:55:50 crc kubenswrapper[4885]: I1002 02:55:50.407868 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-4xx9f/crc-debug-w5jvs"] Oct 02 02:55:50 crc kubenswrapper[4885]: I1002 02:55:50.434389 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z2vq2\" (UniqueName: \"kubernetes.io/projected/b9f7878c-36a4-4cdf-9db7-c4e9d3548d74-kube-api-access-z2vq2\") pod \"b9f7878c-36a4-4cdf-9db7-c4e9d3548d74\" (UID: \"b9f7878c-36a4-4cdf-9db7-c4e9d3548d74\") " Oct 02 02:55:50 crc kubenswrapper[4885]: I1002 02:55:50.434538 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b9f7878c-36a4-4cdf-9db7-c4e9d3548d74-host\") pod \"b9f7878c-36a4-4cdf-9db7-c4e9d3548d74\" (UID: \"b9f7878c-36a4-4cdf-9db7-c4e9d3548d74\") " Oct 02 02:55:50 crc kubenswrapper[4885]: I1002 02:55:50.434733 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b9f7878c-36a4-4cdf-9db7-c4e9d3548d74-host" (OuterVolumeSpecName: "host") pod "b9f7878c-36a4-4cdf-9db7-c4e9d3548d74" (UID: "b9f7878c-36a4-4cdf-9db7-c4e9d3548d74"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 02:55:50 crc kubenswrapper[4885]: I1002 02:55:50.435149 4885 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b9f7878c-36a4-4cdf-9db7-c4e9d3548d74-host\") on node \"crc\" DevicePath \"\"" Oct 02 02:55:50 crc kubenswrapper[4885]: I1002 02:55:50.445082 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b9f7878c-36a4-4cdf-9db7-c4e9d3548d74-kube-api-access-z2vq2" (OuterVolumeSpecName: "kube-api-access-z2vq2") pod "b9f7878c-36a4-4cdf-9db7-c4e9d3548d74" (UID: "b9f7878c-36a4-4cdf-9db7-c4e9d3548d74"). InnerVolumeSpecName "kube-api-access-z2vq2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:55:50 crc kubenswrapper[4885]: I1002 02:55:50.538392 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z2vq2\" (UniqueName: \"kubernetes.io/projected/b9f7878c-36a4-4cdf-9db7-c4e9d3548d74-kube-api-access-z2vq2\") on node \"crc\" DevicePath \"\"" Oct 02 02:55:51 crc kubenswrapper[4885]: I1002 02:55:51.217906 4885 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8d541779741db0adf7522ed1f1a8e8b64619cb6a1a067c4131d08717ef574058" Oct 02 02:55:51 crc kubenswrapper[4885]: I1002 02:55:51.217969 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4xx9f/crc-debug-w5jvs" Oct 02 02:55:51 crc kubenswrapper[4885]: I1002 02:55:51.644485 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-4xx9f/crc-debug-csc27"] Oct 02 02:55:51 crc kubenswrapper[4885]: E1002 02:55:51.645302 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f31184c4-6951-42f6-9b16-ffac70bfbad0" containerName="extract-utilities" Oct 02 02:55:51 crc kubenswrapper[4885]: I1002 02:55:51.645322 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="f31184c4-6951-42f6-9b16-ffac70bfbad0" containerName="extract-utilities" Oct 02 02:55:51 crc kubenswrapper[4885]: E1002 02:55:51.645350 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9f7878c-36a4-4cdf-9db7-c4e9d3548d74" containerName="container-00" Oct 02 02:55:51 crc kubenswrapper[4885]: I1002 02:55:51.645358 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9f7878c-36a4-4cdf-9db7-c4e9d3548d74" containerName="container-00" Oct 02 02:55:51 crc kubenswrapper[4885]: E1002 02:55:51.645380 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f31184c4-6951-42f6-9b16-ffac70bfbad0" containerName="registry-server" Oct 02 02:55:51 crc kubenswrapper[4885]: I1002 02:55:51.645389 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="f31184c4-6951-42f6-9b16-ffac70bfbad0" containerName="registry-server" Oct 02 02:55:51 crc kubenswrapper[4885]: E1002 02:55:51.645403 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f31184c4-6951-42f6-9b16-ffac70bfbad0" containerName="extract-content" Oct 02 02:55:51 crc kubenswrapper[4885]: I1002 02:55:51.645410 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="f31184c4-6951-42f6-9b16-ffac70bfbad0" containerName="extract-content" Oct 02 02:55:51 crc kubenswrapper[4885]: I1002 02:55:51.645650 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="f31184c4-6951-42f6-9b16-ffac70bfbad0" containerName="registry-server" Oct 02 02:55:51 crc kubenswrapper[4885]: I1002 02:55:51.645675 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="b9f7878c-36a4-4cdf-9db7-c4e9d3548d74" containerName="container-00" Oct 02 02:55:51 crc kubenswrapper[4885]: I1002 02:55:51.646368 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4xx9f/crc-debug-csc27" Oct 02 02:55:51 crc kubenswrapper[4885]: I1002 02:55:51.650166 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-4xx9f"/"default-dockercfg-5ljjn" Oct 02 02:55:51 crc kubenswrapper[4885]: I1002 02:55:51.766395 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2e75e08d-a70d-4e23-a1f4-08a0ede66d6e-host\") pod \"crc-debug-csc27\" (UID: \"2e75e08d-a70d-4e23-a1f4-08a0ede66d6e\") " pod="openshift-must-gather-4xx9f/crc-debug-csc27" Oct 02 02:55:51 crc kubenswrapper[4885]: I1002 02:55:51.766749 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7pczc\" (UniqueName: \"kubernetes.io/projected/2e75e08d-a70d-4e23-a1f4-08a0ede66d6e-kube-api-access-7pczc\") pod \"crc-debug-csc27\" (UID: \"2e75e08d-a70d-4e23-a1f4-08a0ede66d6e\") " pod="openshift-must-gather-4xx9f/crc-debug-csc27" Oct 02 02:55:51 crc kubenswrapper[4885]: I1002 02:55:51.869099 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7pczc\" (UniqueName: \"kubernetes.io/projected/2e75e08d-a70d-4e23-a1f4-08a0ede66d6e-kube-api-access-7pczc\") pod \"crc-debug-csc27\" (UID: \"2e75e08d-a70d-4e23-a1f4-08a0ede66d6e\") " pod="openshift-must-gather-4xx9f/crc-debug-csc27" Oct 02 02:55:51 crc kubenswrapper[4885]: I1002 02:55:51.869295 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2e75e08d-a70d-4e23-a1f4-08a0ede66d6e-host\") pod \"crc-debug-csc27\" (UID: \"2e75e08d-a70d-4e23-a1f4-08a0ede66d6e\") " pod="openshift-must-gather-4xx9f/crc-debug-csc27" Oct 02 02:55:51 crc kubenswrapper[4885]: I1002 02:55:51.869541 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2e75e08d-a70d-4e23-a1f4-08a0ede66d6e-host\") pod \"crc-debug-csc27\" (UID: \"2e75e08d-a70d-4e23-a1f4-08a0ede66d6e\") " pod="openshift-must-gather-4xx9f/crc-debug-csc27" Oct 02 02:55:51 crc kubenswrapper[4885]: I1002 02:55:51.900460 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7pczc\" (UniqueName: \"kubernetes.io/projected/2e75e08d-a70d-4e23-a1f4-08a0ede66d6e-kube-api-access-7pczc\") pod \"crc-debug-csc27\" (UID: \"2e75e08d-a70d-4e23-a1f4-08a0ede66d6e\") " pod="openshift-must-gather-4xx9f/crc-debug-csc27" Oct 02 02:55:51 crc kubenswrapper[4885]: I1002 02:55:51.970902 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4xx9f/crc-debug-csc27" Oct 02 02:55:52 crc kubenswrapper[4885]: I1002 02:55:52.072061 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b9f7878c-36a4-4cdf-9db7-c4e9d3548d74" path="/var/lib/kubelet/pods/b9f7878c-36a4-4cdf-9db7-c4e9d3548d74/volumes" Oct 02 02:55:52 crc kubenswrapper[4885]: I1002 02:55:52.235947 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4xx9f/crc-debug-csc27" event={"ID":"2e75e08d-a70d-4e23-a1f4-08a0ede66d6e","Type":"ContainerStarted","Data":"1bafa5d3829fc77fa6cd03f596d8a14222449a877ef122828e3a85819212f205"} Oct 02 02:55:53 crc kubenswrapper[4885]: I1002 02:55:53.255491 4885 generic.go:334] "Generic (PLEG): container finished" podID="2e75e08d-a70d-4e23-a1f4-08a0ede66d6e" containerID="9e37e6c61d1ccb087097d05fb51f026bc4a88397a3de3556cb512c583e8706eb" exitCode=0 Oct 02 02:55:53 crc kubenswrapper[4885]: I1002 02:55:53.255556 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4xx9f/crc-debug-csc27" event={"ID":"2e75e08d-a70d-4e23-a1f4-08a0ede66d6e","Type":"ContainerDied","Data":"9e37e6c61d1ccb087097d05fb51f026bc4a88397a3de3556cb512c583e8706eb"} Oct 02 02:55:54 crc kubenswrapper[4885]: I1002 02:55:54.350921 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4xx9f/crc-debug-csc27" Oct 02 02:55:54 crc kubenswrapper[4885]: I1002 02:55:54.421699 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7pczc\" (UniqueName: \"kubernetes.io/projected/2e75e08d-a70d-4e23-a1f4-08a0ede66d6e-kube-api-access-7pczc\") pod \"2e75e08d-a70d-4e23-a1f4-08a0ede66d6e\" (UID: \"2e75e08d-a70d-4e23-a1f4-08a0ede66d6e\") " Oct 02 02:55:54 crc kubenswrapper[4885]: I1002 02:55:54.421769 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2e75e08d-a70d-4e23-a1f4-08a0ede66d6e-host\") pod \"2e75e08d-a70d-4e23-a1f4-08a0ede66d6e\" (UID: \"2e75e08d-a70d-4e23-a1f4-08a0ede66d6e\") " Oct 02 02:55:54 crc kubenswrapper[4885]: I1002 02:55:54.422148 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2e75e08d-a70d-4e23-a1f4-08a0ede66d6e-host" (OuterVolumeSpecName: "host") pod "2e75e08d-a70d-4e23-a1f4-08a0ede66d6e" (UID: "2e75e08d-a70d-4e23-a1f4-08a0ede66d6e"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 02:55:54 crc kubenswrapper[4885]: I1002 02:55:54.422543 4885 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2e75e08d-a70d-4e23-a1f4-08a0ede66d6e-host\") on node \"crc\" DevicePath \"\"" Oct 02 02:55:54 crc kubenswrapper[4885]: I1002 02:55:54.427075 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e75e08d-a70d-4e23-a1f4-08a0ede66d6e-kube-api-access-7pczc" (OuterVolumeSpecName: "kube-api-access-7pczc") pod "2e75e08d-a70d-4e23-a1f4-08a0ede66d6e" (UID: "2e75e08d-a70d-4e23-a1f4-08a0ede66d6e"). InnerVolumeSpecName "kube-api-access-7pczc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:55:54 crc kubenswrapper[4885]: I1002 02:55:54.523853 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7pczc\" (UniqueName: \"kubernetes.io/projected/2e75e08d-a70d-4e23-a1f4-08a0ede66d6e-kube-api-access-7pczc\") on node \"crc\" DevicePath \"\"" Oct 02 02:55:55 crc kubenswrapper[4885]: I1002 02:55:55.272791 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4xx9f/crc-debug-csc27" event={"ID":"2e75e08d-a70d-4e23-a1f4-08a0ede66d6e","Type":"ContainerDied","Data":"1bafa5d3829fc77fa6cd03f596d8a14222449a877ef122828e3a85819212f205"} Oct 02 02:55:55 crc kubenswrapper[4885]: I1002 02:55:55.272843 4885 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1bafa5d3829fc77fa6cd03f596d8a14222449a877ef122828e3a85819212f205" Oct 02 02:55:55 crc kubenswrapper[4885]: I1002 02:55:55.272875 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4xx9f/crc-debug-csc27" Oct 02 02:55:59 crc kubenswrapper[4885]: I1002 02:55:59.980332 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-4xx9f/crc-debug-csc27"] Oct 02 02:55:59 crc kubenswrapper[4885]: I1002 02:55:59.986494 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-4xx9f/crc-debug-csc27"] Oct 02 02:56:00 crc kubenswrapper[4885]: I1002 02:56:00.062537 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2e75e08d-a70d-4e23-a1f4-08a0ede66d6e" path="/var/lib/kubelet/pods/2e75e08d-a70d-4e23-a1f4-08a0ede66d6e/volumes" Oct 02 02:56:01 crc kubenswrapper[4885]: I1002 02:56:01.236834 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-4xx9f/crc-debug-v6mfc"] Oct 02 02:56:01 crc kubenswrapper[4885]: E1002 02:56:01.237824 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e75e08d-a70d-4e23-a1f4-08a0ede66d6e" containerName="container-00" Oct 02 02:56:01 crc kubenswrapper[4885]: I1002 02:56:01.237847 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e75e08d-a70d-4e23-a1f4-08a0ede66d6e" containerName="container-00" Oct 02 02:56:01 crc kubenswrapper[4885]: I1002 02:56:01.238244 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e75e08d-a70d-4e23-a1f4-08a0ede66d6e" containerName="container-00" Oct 02 02:56:01 crc kubenswrapper[4885]: I1002 02:56:01.239479 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4xx9f/crc-debug-v6mfc" Oct 02 02:56:01 crc kubenswrapper[4885]: I1002 02:56:01.243922 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-4xx9f"/"default-dockercfg-5ljjn" Oct 02 02:56:01 crc kubenswrapper[4885]: I1002 02:56:01.355230 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cc3d740d-64ac-42d7-9429-94a3ed8aec1b-host\") pod \"crc-debug-v6mfc\" (UID: \"cc3d740d-64ac-42d7-9429-94a3ed8aec1b\") " pod="openshift-must-gather-4xx9f/crc-debug-v6mfc" Oct 02 02:56:01 crc kubenswrapper[4885]: I1002 02:56:01.355487 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r9qdp\" (UniqueName: \"kubernetes.io/projected/cc3d740d-64ac-42d7-9429-94a3ed8aec1b-kube-api-access-r9qdp\") pod \"crc-debug-v6mfc\" (UID: \"cc3d740d-64ac-42d7-9429-94a3ed8aec1b\") " pod="openshift-must-gather-4xx9f/crc-debug-v6mfc" Oct 02 02:56:01 crc kubenswrapper[4885]: I1002 02:56:01.457123 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cc3d740d-64ac-42d7-9429-94a3ed8aec1b-host\") pod \"crc-debug-v6mfc\" (UID: \"cc3d740d-64ac-42d7-9429-94a3ed8aec1b\") " pod="openshift-must-gather-4xx9f/crc-debug-v6mfc" Oct 02 02:56:01 crc kubenswrapper[4885]: I1002 02:56:01.457169 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r9qdp\" (UniqueName: \"kubernetes.io/projected/cc3d740d-64ac-42d7-9429-94a3ed8aec1b-kube-api-access-r9qdp\") pod \"crc-debug-v6mfc\" (UID: \"cc3d740d-64ac-42d7-9429-94a3ed8aec1b\") " pod="openshift-must-gather-4xx9f/crc-debug-v6mfc" Oct 02 02:56:01 crc kubenswrapper[4885]: I1002 02:56:01.457371 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cc3d740d-64ac-42d7-9429-94a3ed8aec1b-host\") pod \"crc-debug-v6mfc\" (UID: \"cc3d740d-64ac-42d7-9429-94a3ed8aec1b\") " pod="openshift-must-gather-4xx9f/crc-debug-v6mfc" Oct 02 02:56:01 crc kubenswrapper[4885]: I1002 02:56:01.489352 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r9qdp\" (UniqueName: \"kubernetes.io/projected/cc3d740d-64ac-42d7-9429-94a3ed8aec1b-kube-api-access-r9qdp\") pod \"crc-debug-v6mfc\" (UID: \"cc3d740d-64ac-42d7-9429-94a3ed8aec1b\") " pod="openshift-must-gather-4xx9f/crc-debug-v6mfc" Oct 02 02:56:01 crc kubenswrapper[4885]: I1002 02:56:01.559000 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4xx9f/crc-debug-v6mfc" Oct 02 02:56:01 crc kubenswrapper[4885]: W1002 02:56:01.604192 4885 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcc3d740d_64ac_42d7_9429_94a3ed8aec1b.slice/crio-2672beac4e0c0824e56ab3099dab31af089ef67d1adb096c0c07f993e9bd3ddf WatchSource:0}: Error finding container 2672beac4e0c0824e56ab3099dab31af089ef67d1adb096c0c07f993e9bd3ddf: Status 404 returned error can't find the container with id 2672beac4e0c0824e56ab3099dab31af089ef67d1adb096c0c07f993e9bd3ddf Oct 02 02:56:02 crc kubenswrapper[4885]: I1002 02:56:02.362097 4885 generic.go:334] "Generic (PLEG): container finished" podID="cc3d740d-64ac-42d7-9429-94a3ed8aec1b" containerID="2adde60b87711425fa6819debed2aa082055f0ac44c4f674b191df107eb49970" exitCode=0 Oct 02 02:56:02 crc kubenswrapper[4885]: I1002 02:56:02.362209 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4xx9f/crc-debug-v6mfc" event={"ID":"cc3d740d-64ac-42d7-9429-94a3ed8aec1b","Type":"ContainerDied","Data":"2adde60b87711425fa6819debed2aa082055f0ac44c4f674b191df107eb49970"} Oct 02 02:56:02 crc kubenswrapper[4885]: I1002 02:56:02.362884 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4xx9f/crc-debug-v6mfc" event={"ID":"cc3d740d-64ac-42d7-9429-94a3ed8aec1b","Type":"ContainerStarted","Data":"2672beac4e0c0824e56ab3099dab31af089ef67d1adb096c0c07f993e9bd3ddf"} Oct 02 02:56:02 crc kubenswrapper[4885]: I1002 02:56:02.422763 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-4xx9f/crc-debug-v6mfc"] Oct 02 02:56:02 crc kubenswrapper[4885]: I1002 02:56:02.435808 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-4xx9f/crc-debug-v6mfc"] Oct 02 02:56:03 crc kubenswrapper[4885]: I1002 02:56:03.469444 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4xx9f/crc-debug-v6mfc" Oct 02 02:56:03 crc kubenswrapper[4885]: I1002 02:56:03.500816 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r9qdp\" (UniqueName: \"kubernetes.io/projected/cc3d740d-64ac-42d7-9429-94a3ed8aec1b-kube-api-access-r9qdp\") pod \"cc3d740d-64ac-42d7-9429-94a3ed8aec1b\" (UID: \"cc3d740d-64ac-42d7-9429-94a3ed8aec1b\") " Oct 02 02:56:03 crc kubenswrapper[4885]: I1002 02:56:03.500880 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cc3d740d-64ac-42d7-9429-94a3ed8aec1b-host\") pod \"cc3d740d-64ac-42d7-9429-94a3ed8aec1b\" (UID: \"cc3d740d-64ac-42d7-9429-94a3ed8aec1b\") " Oct 02 02:56:03 crc kubenswrapper[4885]: I1002 02:56:03.501015 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cc3d740d-64ac-42d7-9429-94a3ed8aec1b-host" (OuterVolumeSpecName: "host") pod "cc3d740d-64ac-42d7-9429-94a3ed8aec1b" (UID: "cc3d740d-64ac-42d7-9429-94a3ed8aec1b"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 02:56:03 crc kubenswrapper[4885]: I1002 02:56:03.501410 4885 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cc3d740d-64ac-42d7-9429-94a3ed8aec1b-host\") on node \"crc\" DevicePath \"\"" Oct 02 02:56:03 crc kubenswrapper[4885]: I1002 02:56:03.507146 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc3d740d-64ac-42d7-9429-94a3ed8aec1b-kube-api-access-r9qdp" (OuterVolumeSpecName: "kube-api-access-r9qdp") pod "cc3d740d-64ac-42d7-9429-94a3ed8aec1b" (UID: "cc3d740d-64ac-42d7-9429-94a3ed8aec1b"). InnerVolumeSpecName "kube-api-access-r9qdp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:56:03 crc kubenswrapper[4885]: I1002 02:56:03.603072 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r9qdp\" (UniqueName: \"kubernetes.io/projected/cc3d740d-64ac-42d7-9429-94a3ed8aec1b-kube-api-access-r9qdp\") on node \"crc\" DevicePath \"\"" Oct 02 02:56:04 crc kubenswrapper[4885]: I1002 02:56:04.061321 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc3d740d-64ac-42d7-9429-94a3ed8aec1b" path="/var/lib/kubelet/pods/cc3d740d-64ac-42d7-9429-94a3ed8aec1b/volumes" Oct 02 02:56:04 crc kubenswrapper[4885]: I1002 02:56:04.122662 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-6ff8b75857-wnxfk_8113ab96-dde5-4f9f-b922-687e7a17afa7/kube-rbac-proxy/0.log" Oct 02 02:56:04 crc kubenswrapper[4885]: I1002 02:56:04.197862 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-6ff8b75857-wnxfk_8113ab96-dde5-4f9f-b922-687e7a17afa7/manager/0.log" Oct 02 02:56:04 crc kubenswrapper[4885]: I1002 02:56:04.332492 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cbbec88b66c61e3d10a0bbc868e336160e79c262c68f9747aec27f06598t8b5_5b1a29e8-0ed9-4c6a-bcc5-65849560bd7a/util/0.log" Oct 02 02:56:04 crc kubenswrapper[4885]: I1002 02:56:04.397025 4885 scope.go:117] "RemoveContainer" containerID="2adde60b87711425fa6819debed2aa082055f0ac44c4f674b191df107eb49970" Oct 02 02:56:04 crc kubenswrapper[4885]: I1002 02:56:04.397154 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4xx9f/crc-debug-v6mfc" Oct 02 02:56:04 crc kubenswrapper[4885]: I1002 02:56:04.490527 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cbbec88b66c61e3d10a0bbc868e336160e79c262c68f9747aec27f06598t8b5_5b1a29e8-0ed9-4c6a-bcc5-65849560bd7a/util/0.log" Oct 02 02:56:04 crc kubenswrapper[4885]: I1002 02:56:04.533509 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cbbec88b66c61e3d10a0bbc868e336160e79c262c68f9747aec27f06598t8b5_5b1a29e8-0ed9-4c6a-bcc5-65849560bd7a/pull/0.log" Oct 02 02:56:04 crc kubenswrapper[4885]: I1002 02:56:04.535942 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cbbec88b66c61e3d10a0bbc868e336160e79c262c68f9747aec27f06598t8b5_5b1a29e8-0ed9-4c6a-bcc5-65849560bd7a/pull/0.log" Oct 02 02:56:04 crc kubenswrapper[4885]: I1002 02:56:04.695625 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cbbec88b66c61e3d10a0bbc868e336160e79c262c68f9747aec27f06598t8b5_5b1a29e8-0ed9-4c6a-bcc5-65849560bd7a/util/0.log" Oct 02 02:56:04 crc kubenswrapper[4885]: I1002 02:56:04.700809 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cbbec88b66c61e3d10a0bbc868e336160e79c262c68f9747aec27f06598t8b5_5b1a29e8-0ed9-4c6a-bcc5-65849560bd7a/pull/0.log" Oct 02 02:56:04 crc kubenswrapper[4885]: I1002 02:56:04.707670 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cbbec88b66c61e3d10a0bbc868e336160e79c262c68f9747aec27f06598t8b5_5b1a29e8-0ed9-4c6a-bcc5-65849560bd7a/extract/0.log" Oct 02 02:56:04 crc kubenswrapper[4885]: I1002 02:56:04.876988 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-644bddb6d8-kw92z_e5874982-d78d-4b7c-bf0a-19a7572dddad/kube-rbac-proxy/0.log" Oct 02 02:56:04 crc kubenswrapper[4885]: I1002 02:56:04.931874 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-644bddb6d8-kw92z_e5874982-d78d-4b7c-bf0a-19a7572dddad/manager/0.log" Oct 02 02:56:04 crc kubenswrapper[4885]: I1002 02:56:04.990185 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-84f4f7b77b-mw6h7_cc08d74a-adca-4ae7-98c6-16c7037bc0ca/kube-rbac-proxy/0.log" Oct 02 02:56:05 crc kubenswrapper[4885]: I1002 02:56:05.048574 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-84f4f7b77b-mw6h7_cc08d74a-adca-4ae7-98c6-16c7037bc0ca/manager/0.log" Oct 02 02:56:05 crc kubenswrapper[4885]: I1002 02:56:05.128097 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-84958c4d49-2dsv9_ff10db28-28ca-4f65-8c24-2ed813d7cc5c/kube-rbac-proxy/0.log" Oct 02 02:56:05 crc kubenswrapper[4885]: I1002 02:56:05.222560 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-84958c4d49-2dsv9_ff10db28-28ca-4f65-8c24-2ed813d7cc5c/manager/0.log" Oct 02 02:56:05 crc kubenswrapper[4885]: I1002 02:56:05.299180 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5d889d78cf-862d6_78117b64-3b3a-4892-90de-0dea12fa2602/kube-rbac-proxy/0.log" Oct 02 02:56:05 crc kubenswrapper[4885]: I1002 02:56:05.329997 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5d889d78cf-862d6_78117b64-3b3a-4892-90de-0dea12fa2602/manager/0.log" Oct 02 02:56:05 crc kubenswrapper[4885]: I1002 02:56:05.433591 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-9f4696d94-pv94t_1e64a13b-65af-4460-80a8-4a4659edd8aa/kube-rbac-proxy/0.log" Oct 02 02:56:05 crc kubenswrapper[4885]: I1002 02:56:05.514101 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-9f4696d94-pv94t_1e64a13b-65af-4460-80a8-4a4659edd8aa/manager/0.log" Oct 02 02:56:05 crc kubenswrapper[4885]: I1002 02:56:05.599104 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-9d6c5db85-9fjl5_45e282ed-8195-40dc-86cf-1177f3ebeb65/kube-rbac-proxy/0.log" Oct 02 02:56:05 crc kubenswrapper[4885]: I1002 02:56:05.722035 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-9d6c5db85-9fjl5_45e282ed-8195-40dc-86cf-1177f3ebeb65/manager/0.log" Oct 02 02:56:05 crc kubenswrapper[4885]: I1002 02:56:05.736721 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-5cd4858477-mjtzp_bbd60151-4081-46a8-85c0-0783604392a1/kube-rbac-proxy/0.log" Oct 02 02:56:05 crc kubenswrapper[4885]: I1002 02:56:05.765155 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-5cd4858477-mjtzp_bbd60151-4081-46a8-85c0-0783604392a1/manager/0.log" Oct 02 02:56:05 crc kubenswrapper[4885]: I1002 02:56:05.874305 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-5bd55b4bff-pww46_d2263844-1435-4352-b773-04e8d10f35a0/kube-rbac-proxy/0.log" Oct 02 02:56:05 crc kubenswrapper[4885]: I1002 02:56:05.974913 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-5bd55b4bff-pww46_d2263844-1435-4352-b773-04e8d10f35a0/manager/0.log" Oct 02 02:56:06 crc kubenswrapper[4885]: I1002 02:56:06.051897 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-6d68dbc695-9zf8p_bcb6c44c-c620-465c-8db3-6d770deda522/kube-rbac-proxy/0.log" Oct 02 02:56:06 crc kubenswrapper[4885]: I1002 02:56:06.092336 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-6d68dbc695-9zf8p_bcb6c44c-c620-465c-8db3-6d770deda522/manager/0.log" Oct 02 02:56:06 crc kubenswrapper[4885]: I1002 02:56:06.136750 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-88c7-nz599_0f1b84a4-59d6-45fe-9c2f-0bdff82fde6e/kube-rbac-proxy/0.log" Oct 02 02:56:06 crc kubenswrapper[4885]: I1002 02:56:06.243357 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-88c7-nz599_0f1b84a4-59d6-45fe-9c2f-0bdff82fde6e/manager/0.log" Oct 02 02:56:06 crc kubenswrapper[4885]: I1002 02:56:06.286463 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-849d5b9b84-clsdk_ae73c7b6-6b7e-4ca9-8a4e-0694ab34d9a9/kube-rbac-proxy/0.log" Oct 02 02:56:06 crc kubenswrapper[4885]: I1002 02:56:06.348656 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-849d5b9b84-clsdk_ae73c7b6-6b7e-4ca9-8a4e-0694ab34d9a9/manager/0.log" Oct 02 02:56:06 crc kubenswrapper[4885]: I1002 02:56:06.477100 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-64cd67b5cb-hwrj4_e6673002-4d33-4c32-8eb4-f1727339b76d/kube-rbac-proxy/0.log" Oct 02 02:56:06 crc kubenswrapper[4885]: I1002 02:56:06.479606 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-64cd67b5cb-hwrj4_e6673002-4d33-4c32-8eb4-f1727339b76d/manager/0.log" Oct 02 02:56:06 crc kubenswrapper[4885]: I1002 02:56:06.612244 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-7b787867f4-kcpkv_9d724768-b5a0-4a55-951a-d350127061cb/kube-rbac-proxy/0.log" Oct 02 02:56:06 crc kubenswrapper[4885]: I1002 02:56:06.626976 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-7b787867f4-kcpkv_9d724768-b5a0-4a55-951a-d350127061cb/manager/0.log" Oct 02 02:56:06 crc kubenswrapper[4885]: I1002 02:56:06.705738 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-5869cb545-wgfjr_dc61bc9b-923b-41cb-b724-62c72caf0586/kube-rbac-proxy/0.log" Oct 02 02:56:06 crc kubenswrapper[4885]: I1002 02:56:06.754194 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-5869cb545-wgfjr_dc61bc9b-923b-41cb-b724-62c72caf0586/manager/0.log" Oct 02 02:56:06 crc kubenswrapper[4885]: I1002 02:56:06.827754 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-557f5d867b-ndfjx_8b18e6b1-82f2-46cd-99c8-85af816df20f/kube-rbac-proxy/0.log" Oct 02 02:56:07 crc kubenswrapper[4885]: I1002 02:56:07.027806 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-6bcc7488c-bc4gx_02e27ea3-228f-4f78-9135-deaff888afd2/kube-rbac-proxy/0.log" Oct 02 02:56:07 crc kubenswrapper[4885]: I1002 02:56:07.240139 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-zgthg_12809968-8a60-4ffa-9506-0c89ce063382/registry-server/0.log" Oct 02 02:56:07 crc kubenswrapper[4885]: I1002 02:56:07.244954 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-6bcc7488c-bc4gx_02e27ea3-228f-4f78-9135-deaff888afd2/operator/0.log" Oct 02 02:56:07 crc kubenswrapper[4885]: I1002 02:56:07.349296 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-9976ff44c-gl72q_c540bfce-7653-4976-8d6e-36b479c3c711/kube-rbac-proxy/0.log" Oct 02 02:56:07 crc kubenswrapper[4885]: I1002 02:56:07.444653 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-9976ff44c-gl72q_c540bfce-7653-4976-8d6e-36b479c3c711/manager/0.log" Oct 02 02:56:07 crc kubenswrapper[4885]: I1002 02:56:07.487732 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-589c58c6c-r65hz_6e740063-fb49-4847-ba6d-77b6c089ba50/kube-rbac-proxy/0.log" Oct 02 02:56:07 crc kubenswrapper[4885]: I1002 02:56:07.608775 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-589c58c6c-r65hz_6e740063-fb49-4847-ba6d-77b6c089ba50/manager/0.log" Oct 02 02:56:07 crc kubenswrapper[4885]: I1002 02:56:07.708947 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-5f97d8c699-bngfl_d2768afc-17ee-4141-aa68-7366faaa0145/operator/0.log" Oct 02 02:56:07 crc kubenswrapper[4885]: I1002 02:56:07.833772 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-557f5d867b-ndfjx_8b18e6b1-82f2-46cd-99c8-85af816df20f/manager/0.log" Oct 02 02:56:08 crc kubenswrapper[4885]: I1002 02:56:08.287765 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-b8d54b5d7-26g2l_dd09a3bb-1217-4a3a-822b-2b2f04171271/kube-rbac-proxy/0.log" Oct 02 02:56:08 crc kubenswrapper[4885]: I1002 02:56:08.300544 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-84d6b4b759-6f74g_f80d2e7d-10a3-492f-a045-b48126a02490/kube-rbac-proxy/0.log" Oct 02 02:56:08 crc kubenswrapper[4885]: I1002 02:56:08.307984 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-84d6b4b759-6f74g_f80d2e7d-10a3-492f-a045-b48126a02490/manager/0.log" Oct 02 02:56:08 crc kubenswrapper[4885]: I1002 02:56:08.343695 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-b8d54b5d7-26g2l_dd09a3bb-1217-4a3a-822b-2b2f04171271/manager/0.log" Oct 02 02:56:08 crc kubenswrapper[4885]: I1002 02:56:08.479439 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-85777745bb-94284_74a07e3e-5061-43f9-b239-eb480bade999/kube-rbac-proxy/0.log" Oct 02 02:56:08 crc kubenswrapper[4885]: I1002 02:56:08.505896 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-85777745bb-94284_74a07e3e-5061-43f9-b239-eb480bade999/manager/0.log" Oct 02 02:56:08 crc kubenswrapper[4885]: I1002 02:56:08.569405 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-6b9957f54f-mb6mf_c5dd164b-699d-4f14-85d0-91f138a42c95/kube-rbac-proxy/0.log" Oct 02 02:56:08 crc kubenswrapper[4885]: I1002 02:56:08.637096 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-6b9957f54f-mb6mf_c5dd164b-699d-4f14-85d0-91f138a42c95/manager/0.log" Oct 02 02:56:15 crc kubenswrapper[4885]: I1002 02:56:15.683986 4885 scope.go:117] "RemoveContainer" containerID="46552ac844213a16ef1a7a1cee33dfab4092e42f05442399558109f44192ac62" Oct 02 02:56:26 crc kubenswrapper[4885]: I1002 02:56:26.555452 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-dpqkj_cc6bb484-0230-46d9-9507-b641d2a5f330/control-plane-machine-set-operator/0.log" Oct 02 02:56:26 crc kubenswrapper[4885]: I1002 02:56:26.648713 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-5wldz_76fb6dce-01dd-4aef-8213-e03be7a5e055/kube-rbac-proxy/0.log" Oct 02 02:56:26 crc kubenswrapper[4885]: I1002 02:56:26.687546 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-5wldz_76fb6dce-01dd-4aef-8213-e03be7a5e055/machine-api-operator/0.log" Oct 02 02:56:42 crc kubenswrapper[4885]: I1002 02:56:42.008776 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-pj866_0b6eafa9-42ce-4e5b-9ded-180a669d26e2/cert-manager-controller/0.log" Oct 02 02:56:42 crc kubenswrapper[4885]: I1002 02:56:42.139024 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-rbccc_e50cfe49-c47f-4f2a-9799-0c79123fac70/cert-manager-cainjector/0.log" Oct 02 02:56:42 crc kubenswrapper[4885]: I1002 02:56:42.159636 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-vh2n2_18c66206-ad33-47d2-a37a-b3b50c2bf984/cert-manager-webhook/0.log" Oct 02 02:56:45 crc kubenswrapper[4885]: I1002 02:56:45.506174 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-v7r52"] Oct 02 02:56:45 crc kubenswrapper[4885]: E1002 02:56:45.510309 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc3d740d-64ac-42d7-9429-94a3ed8aec1b" containerName="container-00" Oct 02 02:56:45 crc kubenswrapper[4885]: I1002 02:56:45.510341 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc3d740d-64ac-42d7-9429-94a3ed8aec1b" containerName="container-00" Oct 02 02:56:45 crc kubenswrapper[4885]: I1002 02:56:45.510708 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc3d740d-64ac-42d7-9429-94a3ed8aec1b" containerName="container-00" Oct 02 02:56:45 crc kubenswrapper[4885]: I1002 02:56:45.513761 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v7r52" Oct 02 02:56:45 crc kubenswrapper[4885]: I1002 02:56:45.548496 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-v7r52"] Oct 02 02:56:45 crc kubenswrapper[4885]: I1002 02:56:45.558286 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bzb9d\" (UniqueName: \"kubernetes.io/projected/f0ba71d4-22e1-4d21-8104-05476134a59a-kube-api-access-bzb9d\") pod \"community-operators-v7r52\" (UID: \"f0ba71d4-22e1-4d21-8104-05476134a59a\") " pod="openshift-marketplace/community-operators-v7r52" Oct 02 02:56:45 crc kubenswrapper[4885]: I1002 02:56:45.558363 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f0ba71d4-22e1-4d21-8104-05476134a59a-catalog-content\") pod \"community-operators-v7r52\" (UID: \"f0ba71d4-22e1-4d21-8104-05476134a59a\") " pod="openshift-marketplace/community-operators-v7r52" Oct 02 02:56:45 crc kubenswrapper[4885]: I1002 02:56:45.558747 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f0ba71d4-22e1-4d21-8104-05476134a59a-utilities\") pod \"community-operators-v7r52\" (UID: \"f0ba71d4-22e1-4d21-8104-05476134a59a\") " pod="openshift-marketplace/community-operators-v7r52" Oct 02 02:56:45 crc kubenswrapper[4885]: I1002 02:56:45.660771 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f0ba71d4-22e1-4d21-8104-05476134a59a-utilities\") pod \"community-operators-v7r52\" (UID: \"f0ba71d4-22e1-4d21-8104-05476134a59a\") " pod="openshift-marketplace/community-operators-v7r52" Oct 02 02:56:45 crc kubenswrapper[4885]: I1002 02:56:45.660837 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bzb9d\" (UniqueName: \"kubernetes.io/projected/f0ba71d4-22e1-4d21-8104-05476134a59a-kube-api-access-bzb9d\") pod \"community-operators-v7r52\" (UID: \"f0ba71d4-22e1-4d21-8104-05476134a59a\") " pod="openshift-marketplace/community-operators-v7r52" Oct 02 02:56:45 crc kubenswrapper[4885]: I1002 02:56:45.660872 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f0ba71d4-22e1-4d21-8104-05476134a59a-catalog-content\") pod \"community-operators-v7r52\" (UID: \"f0ba71d4-22e1-4d21-8104-05476134a59a\") " pod="openshift-marketplace/community-operators-v7r52" Oct 02 02:56:45 crc kubenswrapper[4885]: I1002 02:56:45.661328 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f0ba71d4-22e1-4d21-8104-05476134a59a-utilities\") pod \"community-operators-v7r52\" (UID: \"f0ba71d4-22e1-4d21-8104-05476134a59a\") " pod="openshift-marketplace/community-operators-v7r52" Oct 02 02:56:45 crc kubenswrapper[4885]: I1002 02:56:45.661387 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f0ba71d4-22e1-4d21-8104-05476134a59a-catalog-content\") pod \"community-operators-v7r52\" (UID: \"f0ba71d4-22e1-4d21-8104-05476134a59a\") " pod="openshift-marketplace/community-operators-v7r52" Oct 02 02:56:45 crc kubenswrapper[4885]: I1002 02:56:45.692066 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bzb9d\" (UniqueName: \"kubernetes.io/projected/f0ba71d4-22e1-4d21-8104-05476134a59a-kube-api-access-bzb9d\") pod \"community-operators-v7r52\" (UID: \"f0ba71d4-22e1-4d21-8104-05476134a59a\") " pod="openshift-marketplace/community-operators-v7r52" Oct 02 02:56:45 crc kubenswrapper[4885]: I1002 02:56:45.863382 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v7r52" Oct 02 02:56:46 crc kubenswrapper[4885]: I1002 02:56:46.391644 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-v7r52"] Oct 02 02:56:46 crc kubenswrapper[4885]: I1002 02:56:46.833198 4885 generic.go:334] "Generic (PLEG): container finished" podID="f0ba71d4-22e1-4d21-8104-05476134a59a" containerID="36dde8563debc077f548a7c57dd8483f10ef11168a58dde7ef9ef1326f74b47f" exitCode=0 Oct 02 02:56:46 crc kubenswrapper[4885]: I1002 02:56:46.833400 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v7r52" event={"ID":"f0ba71d4-22e1-4d21-8104-05476134a59a","Type":"ContainerDied","Data":"36dde8563debc077f548a7c57dd8483f10ef11168a58dde7ef9ef1326f74b47f"} Oct 02 02:56:46 crc kubenswrapper[4885]: I1002 02:56:46.833808 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v7r52" event={"ID":"f0ba71d4-22e1-4d21-8104-05476134a59a","Type":"ContainerStarted","Data":"7e164fc77966d21dba2095aee22e8f1a918117db8eab84739e084d6c07aae34a"} Oct 02 02:56:48 crc kubenswrapper[4885]: I1002 02:56:48.867315 4885 generic.go:334] "Generic (PLEG): container finished" podID="f0ba71d4-22e1-4d21-8104-05476134a59a" containerID="0a0826bc5616bd54cb5e49c0b645f2d16eaae50f16e938900b9f786b7ccbcf42" exitCode=0 Oct 02 02:56:48 crc kubenswrapper[4885]: I1002 02:56:48.867653 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v7r52" event={"ID":"f0ba71d4-22e1-4d21-8104-05476134a59a","Type":"ContainerDied","Data":"0a0826bc5616bd54cb5e49c0b645f2d16eaae50f16e938900b9f786b7ccbcf42"} Oct 02 02:56:49 crc kubenswrapper[4885]: I1002 02:56:49.882744 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v7r52" event={"ID":"f0ba71d4-22e1-4d21-8104-05476134a59a","Type":"ContainerStarted","Data":"5a4d397261c9e285b7ec8a201b3485d6daf3b14e0bfbe9ff2fbae379b0d3dc4a"} Oct 02 02:56:49 crc kubenswrapper[4885]: I1002 02:56:49.929542 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-v7r52" podStartSLOduration=2.495122342 podStartE2EDuration="4.929523408s" podCreationTimestamp="2025-10-02 02:56:45 +0000 UTC" firstStartedPulling="2025-10-02 02:56:46.835645199 +0000 UTC m=+4195.647392628" lastFinishedPulling="2025-10-02 02:56:49.270046265 +0000 UTC m=+4198.081793694" observedRunningTime="2025-10-02 02:56:49.919563434 +0000 UTC m=+4198.731310843" watchObservedRunningTime="2025-10-02 02:56:49.929523408 +0000 UTC m=+4198.741270827" Oct 02 02:56:55 crc kubenswrapper[4885]: I1002 02:56:55.863656 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-v7r52" Oct 02 02:56:55 crc kubenswrapper[4885]: I1002 02:56:55.864341 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-v7r52" Oct 02 02:56:56 crc kubenswrapper[4885]: I1002 02:56:56.434882 4885 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-v7r52" Oct 02 02:56:56 crc kubenswrapper[4885]: I1002 02:56:56.505117 4885 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-v7r52" Oct 02 02:56:57 crc kubenswrapper[4885]: I1002 02:56:57.751011 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-6b874cbd85-dvlk4_c7f8ff02-de9d-456d-ba5e-a794f568c5cf/nmstate-console-plugin/0.log" Oct 02 02:56:58 crc kubenswrapper[4885]: I1002 02:56:58.043250 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-4xzpg_4376783f-66f6-41a4-a42e-2666f1cc9978/nmstate-handler/0.log" Oct 02 02:56:58 crc kubenswrapper[4885]: I1002 02:56:58.050431 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-sr7mv_c2a4785b-6378-4f85-ae1d-46e27903af36/kube-rbac-proxy/0.log" Oct 02 02:56:58 crc kubenswrapper[4885]: I1002 02:56:58.094225 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-sr7mv_c2a4785b-6378-4f85-ae1d-46e27903af36/nmstate-metrics/0.log" Oct 02 02:56:58 crc kubenswrapper[4885]: I1002 02:56:58.213730 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-858ddd8f98-r4dhw_6f1fd2a7-4a15-46b4-818f-c24f9fcd2bcd/nmstate-operator/0.log" Oct 02 02:56:58 crc kubenswrapper[4885]: I1002 02:56:58.320057 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6cdbc54649-jk45f_f75137b4-4551-43cd-b20c-6044d13d27a1/nmstate-webhook/0.log" Oct 02 02:57:00 crc kubenswrapper[4885]: I1002 02:57:00.531876 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-v7r52"] Oct 02 02:57:00 crc kubenswrapper[4885]: I1002 02:57:00.532590 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-v7r52" podUID="f0ba71d4-22e1-4d21-8104-05476134a59a" containerName="registry-server" containerID="cri-o://5a4d397261c9e285b7ec8a201b3485d6daf3b14e0bfbe9ff2fbae379b0d3dc4a" gracePeriod=2 Oct 02 02:57:00 crc kubenswrapper[4885]: I1002 02:57:00.990654 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v7r52" Oct 02 02:57:00 crc kubenswrapper[4885]: I1002 02:57:00.993310 4885 generic.go:334] "Generic (PLEG): container finished" podID="f0ba71d4-22e1-4d21-8104-05476134a59a" containerID="5a4d397261c9e285b7ec8a201b3485d6daf3b14e0bfbe9ff2fbae379b0d3dc4a" exitCode=0 Oct 02 02:57:00 crc kubenswrapper[4885]: I1002 02:57:00.993419 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v7r52" event={"ID":"f0ba71d4-22e1-4d21-8104-05476134a59a","Type":"ContainerDied","Data":"5a4d397261c9e285b7ec8a201b3485d6daf3b14e0bfbe9ff2fbae379b0d3dc4a"} Oct 02 02:57:00 crc kubenswrapper[4885]: I1002 02:57:00.993682 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v7r52" event={"ID":"f0ba71d4-22e1-4d21-8104-05476134a59a","Type":"ContainerDied","Data":"7e164fc77966d21dba2095aee22e8f1a918117db8eab84739e084d6c07aae34a"} Oct 02 02:57:00 crc kubenswrapper[4885]: I1002 02:57:00.993796 4885 scope.go:117] "RemoveContainer" containerID="5a4d397261c9e285b7ec8a201b3485d6daf3b14e0bfbe9ff2fbae379b0d3dc4a" Oct 02 02:57:01 crc kubenswrapper[4885]: I1002 02:57:01.018077 4885 scope.go:117] "RemoveContainer" containerID="0a0826bc5616bd54cb5e49c0b645f2d16eaae50f16e938900b9f786b7ccbcf42" Oct 02 02:57:01 crc kubenswrapper[4885]: I1002 02:57:01.040003 4885 scope.go:117] "RemoveContainer" containerID="36dde8563debc077f548a7c57dd8483f10ef11168a58dde7ef9ef1326f74b47f" Oct 02 02:57:01 crc kubenswrapper[4885]: I1002 02:57:01.070014 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f0ba71d4-22e1-4d21-8104-05476134a59a-utilities\") pod \"f0ba71d4-22e1-4d21-8104-05476134a59a\" (UID: \"f0ba71d4-22e1-4d21-8104-05476134a59a\") " Oct 02 02:57:01 crc kubenswrapper[4885]: I1002 02:57:01.070360 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f0ba71d4-22e1-4d21-8104-05476134a59a-catalog-content\") pod \"f0ba71d4-22e1-4d21-8104-05476134a59a\" (UID: \"f0ba71d4-22e1-4d21-8104-05476134a59a\") " Oct 02 02:57:01 crc kubenswrapper[4885]: I1002 02:57:01.070656 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bzb9d\" (UniqueName: \"kubernetes.io/projected/f0ba71d4-22e1-4d21-8104-05476134a59a-kube-api-access-bzb9d\") pod \"f0ba71d4-22e1-4d21-8104-05476134a59a\" (UID: \"f0ba71d4-22e1-4d21-8104-05476134a59a\") " Oct 02 02:57:01 crc kubenswrapper[4885]: I1002 02:57:01.071132 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f0ba71d4-22e1-4d21-8104-05476134a59a-utilities" (OuterVolumeSpecName: "utilities") pod "f0ba71d4-22e1-4d21-8104-05476134a59a" (UID: "f0ba71d4-22e1-4d21-8104-05476134a59a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:57:01 crc kubenswrapper[4885]: I1002 02:57:01.071626 4885 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f0ba71d4-22e1-4d21-8104-05476134a59a-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 02:57:01 crc kubenswrapper[4885]: I1002 02:57:01.077435 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f0ba71d4-22e1-4d21-8104-05476134a59a-kube-api-access-bzb9d" (OuterVolumeSpecName: "kube-api-access-bzb9d") pod "f0ba71d4-22e1-4d21-8104-05476134a59a" (UID: "f0ba71d4-22e1-4d21-8104-05476134a59a"). InnerVolumeSpecName "kube-api-access-bzb9d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:57:01 crc kubenswrapper[4885]: I1002 02:57:01.143564 4885 scope.go:117] "RemoveContainer" containerID="5a4d397261c9e285b7ec8a201b3485d6daf3b14e0bfbe9ff2fbae379b0d3dc4a" Oct 02 02:57:01 crc kubenswrapper[4885]: I1002 02:57:01.145229 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f0ba71d4-22e1-4d21-8104-05476134a59a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f0ba71d4-22e1-4d21-8104-05476134a59a" (UID: "f0ba71d4-22e1-4d21-8104-05476134a59a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:57:01 crc kubenswrapper[4885]: E1002 02:57:01.145733 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5a4d397261c9e285b7ec8a201b3485d6daf3b14e0bfbe9ff2fbae379b0d3dc4a\": container with ID starting with 5a4d397261c9e285b7ec8a201b3485d6daf3b14e0bfbe9ff2fbae379b0d3dc4a not found: ID does not exist" containerID="5a4d397261c9e285b7ec8a201b3485d6daf3b14e0bfbe9ff2fbae379b0d3dc4a" Oct 02 02:57:01 crc kubenswrapper[4885]: I1002 02:57:01.145768 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5a4d397261c9e285b7ec8a201b3485d6daf3b14e0bfbe9ff2fbae379b0d3dc4a"} err="failed to get container status \"5a4d397261c9e285b7ec8a201b3485d6daf3b14e0bfbe9ff2fbae379b0d3dc4a\": rpc error: code = NotFound desc = could not find container \"5a4d397261c9e285b7ec8a201b3485d6daf3b14e0bfbe9ff2fbae379b0d3dc4a\": container with ID starting with 5a4d397261c9e285b7ec8a201b3485d6daf3b14e0bfbe9ff2fbae379b0d3dc4a not found: ID does not exist" Oct 02 02:57:01 crc kubenswrapper[4885]: I1002 02:57:01.145788 4885 scope.go:117] "RemoveContainer" containerID="0a0826bc5616bd54cb5e49c0b645f2d16eaae50f16e938900b9f786b7ccbcf42" Oct 02 02:57:01 crc kubenswrapper[4885]: E1002 02:57:01.146084 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0a0826bc5616bd54cb5e49c0b645f2d16eaae50f16e938900b9f786b7ccbcf42\": container with ID starting with 0a0826bc5616bd54cb5e49c0b645f2d16eaae50f16e938900b9f786b7ccbcf42 not found: ID does not exist" containerID="0a0826bc5616bd54cb5e49c0b645f2d16eaae50f16e938900b9f786b7ccbcf42" Oct 02 02:57:01 crc kubenswrapper[4885]: I1002 02:57:01.146141 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0a0826bc5616bd54cb5e49c0b645f2d16eaae50f16e938900b9f786b7ccbcf42"} err="failed to get container status \"0a0826bc5616bd54cb5e49c0b645f2d16eaae50f16e938900b9f786b7ccbcf42\": rpc error: code = NotFound desc = could not find container \"0a0826bc5616bd54cb5e49c0b645f2d16eaae50f16e938900b9f786b7ccbcf42\": container with ID starting with 0a0826bc5616bd54cb5e49c0b645f2d16eaae50f16e938900b9f786b7ccbcf42 not found: ID does not exist" Oct 02 02:57:01 crc kubenswrapper[4885]: I1002 02:57:01.146185 4885 scope.go:117] "RemoveContainer" containerID="36dde8563debc077f548a7c57dd8483f10ef11168a58dde7ef9ef1326f74b47f" Oct 02 02:57:01 crc kubenswrapper[4885]: E1002 02:57:01.147221 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"36dde8563debc077f548a7c57dd8483f10ef11168a58dde7ef9ef1326f74b47f\": container with ID starting with 36dde8563debc077f548a7c57dd8483f10ef11168a58dde7ef9ef1326f74b47f not found: ID does not exist" containerID="36dde8563debc077f548a7c57dd8483f10ef11168a58dde7ef9ef1326f74b47f" Oct 02 02:57:01 crc kubenswrapper[4885]: I1002 02:57:01.147245 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"36dde8563debc077f548a7c57dd8483f10ef11168a58dde7ef9ef1326f74b47f"} err="failed to get container status \"36dde8563debc077f548a7c57dd8483f10ef11168a58dde7ef9ef1326f74b47f\": rpc error: code = NotFound desc = could not find container \"36dde8563debc077f548a7c57dd8483f10ef11168a58dde7ef9ef1326f74b47f\": container with ID starting with 36dde8563debc077f548a7c57dd8483f10ef11168a58dde7ef9ef1326f74b47f not found: ID does not exist" Oct 02 02:57:01 crc kubenswrapper[4885]: I1002 02:57:01.174625 4885 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f0ba71d4-22e1-4d21-8104-05476134a59a-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 02:57:01 crc kubenswrapper[4885]: I1002 02:57:01.174654 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bzb9d\" (UniqueName: \"kubernetes.io/projected/f0ba71d4-22e1-4d21-8104-05476134a59a-kube-api-access-bzb9d\") on node \"crc\" DevicePath \"\"" Oct 02 02:57:02 crc kubenswrapper[4885]: I1002 02:57:02.005522 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v7r52" Oct 02 02:57:02 crc kubenswrapper[4885]: I1002 02:57:02.060318 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-v7r52"] Oct 02 02:57:02 crc kubenswrapper[4885]: I1002 02:57:02.061962 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-v7r52"] Oct 02 02:57:04 crc kubenswrapper[4885]: I1002 02:57:04.058565 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f0ba71d4-22e1-4d21-8104-05476134a59a" path="/var/lib/kubelet/pods/f0ba71d4-22e1-4d21-8104-05476134a59a/volumes" Oct 02 02:57:14 crc kubenswrapper[4885]: I1002 02:57:14.417836 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-djznn_47f27b52-1453-4566-954f-94e6b02b2221/kube-rbac-proxy/0.log" Oct 02 02:57:14 crc kubenswrapper[4885]: I1002 02:57:14.519984 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-djznn_47f27b52-1453-4566-954f-94e6b02b2221/controller/0.log" Oct 02 02:57:14 crc kubenswrapper[4885]: I1002 02:57:14.668964 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ggsh4_05715b27-682b-46a1-80b5-667667147ce0/cp-frr-files/0.log" Oct 02 02:57:14 crc kubenswrapper[4885]: I1002 02:57:14.816698 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ggsh4_05715b27-682b-46a1-80b5-667667147ce0/cp-frr-files/0.log" Oct 02 02:57:14 crc kubenswrapper[4885]: I1002 02:57:14.864725 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ggsh4_05715b27-682b-46a1-80b5-667667147ce0/cp-reloader/0.log" Oct 02 02:57:14 crc kubenswrapper[4885]: I1002 02:57:14.874571 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ggsh4_05715b27-682b-46a1-80b5-667667147ce0/cp-metrics/0.log" Oct 02 02:57:14 crc kubenswrapper[4885]: I1002 02:57:14.874706 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ggsh4_05715b27-682b-46a1-80b5-667667147ce0/cp-reloader/0.log" Oct 02 02:57:15 crc kubenswrapper[4885]: I1002 02:57:15.669865 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ggsh4_05715b27-682b-46a1-80b5-667667147ce0/cp-frr-files/0.log" Oct 02 02:57:15 crc kubenswrapper[4885]: I1002 02:57:15.684115 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ggsh4_05715b27-682b-46a1-80b5-667667147ce0/cp-metrics/0.log" Oct 02 02:57:15 crc kubenswrapper[4885]: I1002 02:57:15.723741 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ggsh4_05715b27-682b-46a1-80b5-667667147ce0/cp-metrics/0.log" Oct 02 02:57:15 crc kubenswrapper[4885]: I1002 02:57:15.764555 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ggsh4_05715b27-682b-46a1-80b5-667667147ce0/cp-reloader/0.log" Oct 02 02:57:15 crc kubenswrapper[4885]: I1002 02:57:15.908402 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ggsh4_05715b27-682b-46a1-80b5-667667147ce0/cp-frr-files/0.log" Oct 02 02:57:15 crc kubenswrapper[4885]: I1002 02:57:15.933992 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ggsh4_05715b27-682b-46a1-80b5-667667147ce0/cp-reloader/0.log" Oct 02 02:57:15 crc kubenswrapper[4885]: I1002 02:57:15.948072 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ggsh4_05715b27-682b-46a1-80b5-667667147ce0/cp-metrics/0.log" Oct 02 02:57:16 crc kubenswrapper[4885]: I1002 02:57:16.014227 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ggsh4_05715b27-682b-46a1-80b5-667667147ce0/controller/0.log" Oct 02 02:57:16 crc kubenswrapper[4885]: I1002 02:57:16.197668 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ggsh4_05715b27-682b-46a1-80b5-667667147ce0/frr-metrics/0.log" Oct 02 02:57:16 crc kubenswrapper[4885]: I1002 02:57:16.208494 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ggsh4_05715b27-682b-46a1-80b5-667667147ce0/kube-rbac-proxy/0.log" Oct 02 02:57:16 crc kubenswrapper[4885]: I1002 02:57:16.307876 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ggsh4_05715b27-682b-46a1-80b5-667667147ce0/kube-rbac-proxy-frr/0.log" Oct 02 02:57:16 crc kubenswrapper[4885]: I1002 02:57:16.430914 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ggsh4_05715b27-682b-46a1-80b5-667667147ce0/reloader/0.log" Oct 02 02:57:16 crc kubenswrapper[4885]: I1002 02:57:16.550621 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-64bf5d555-q7xtq_66663f53-9790-48b4-8d16-4675661f340d/frr-k8s-webhook-server/0.log" Oct 02 02:57:17 crc kubenswrapper[4885]: I1002 02:57:17.359809 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-6c54f448c5-tknpj_7dc51fda-276d-4739-975d-e88c9fcec5d3/manager/0.log" Oct 02 02:57:17 crc kubenswrapper[4885]: I1002 02:57:17.540599 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ggsh4_05715b27-682b-46a1-80b5-667667147ce0/frr/0.log" Oct 02 02:57:17 crc kubenswrapper[4885]: I1002 02:57:17.564050 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-fd987dcb7-f2dx2_e0315b09-692d-4c5f-8861-0863f90338fa/webhook-server/0.log" Oct 02 02:57:17 crc kubenswrapper[4885]: I1002 02:57:17.682486 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-5h9mk_e3d6e1b0-5963-4e8f-bf3d-eb0e88ec0395/kube-rbac-proxy/0.log" Oct 02 02:57:18 crc kubenswrapper[4885]: I1002 02:57:18.059685 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-5h9mk_e3d6e1b0-5963-4e8f-bf3d-eb0e88ec0395/speaker/0.log" Oct 02 02:57:32 crc kubenswrapper[4885]: I1002 02:57:32.447912 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2ms2kw_65b0be27-bd89-4ddd-a578-7977b676f674/util/0.log" Oct 02 02:57:32 crc kubenswrapper[4885]: I1002 02:57:32.625181 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2ms2kw_65b0be27-bd89-4ddd-a578-7977b676f674/util/0.log" Oct 02 02:57:32 crc kubenswrapper[4885]: I1002 02:57:32.696470 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2ms2kw_65b0be27-bd89-4ddd-a578-7977b676f674/pull/0.log" Oct 02 02:57:32 crc kubenswrapper[4885]: I1002 02:57:32.727460 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2ms2kw_65b0be27-bd89-4ddd-a578-7977b676f674/pull/0.log" Oct 02 02:57:32 crc kubenswrapper[4885]: I1002 02:57:32.925185 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2ms2kw_65b0be27-bd89-4ddd-a578-7977b676f674/util/0.log" Oct 02 02:57:32 crc kubenswrapper[4885]: I1002 02:57:32.930823 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2ms2kw_65b0be27-bd89-4ddd-a578-7977b676f674/pull/0.log" Oct 02 02:57:32 crc kubenswrapper[4885]: I1002 02:57:32.970940 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2ms2kw_65b0be27-bd89-4ddd-a578-7977b676f674/extract/0.log" Oct 02 02:57:33 crc kubenswrapper[4885]: I1002 02:57:33.123904 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-sn6xg_56c1dc36-0146-45ea-9396-990a988a6d9e/extract-utilities/0.log" Oct 02 02:57:33 crc kubenswrapper[4885]: I1002 02:57:33.257791 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-sn6xg_56c1dc36-0146-45ea-9396-990a988a6d9e/extract-content/0.log" Oct 02 02:57:33 crc kubenswrapper[4885]: I1002 02:57:33.282539 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-sn6xg_56c1dc36-0146-45ea-9396-990a988a6d9e/extract-utilities/0.log" Oct 02 02:57:33 crc kubenswrapper[4885]: I1002 02:57:33.295448 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-sn6xg_56c1dc36-0146-45ea-9396-990a988a6d9e/extract-content/0.log" Oct 02 02:57:33 crc kubenswrapper[4885]: I1002 02:57:33.465005 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-sn6xg_56c1dc36-0146-45ea-9396-990a988a6d9e/extract-utilities/0.log" Oct 02 02:57:33 crc kubenswrapper[4885]: I1002 02:57:33.496021 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-sn6xg_56c1dc36-0146-45ea-9396-990a988a6d9e/extract-content/0.log" Oct 02 02:57:33 crc kubenswrapper[4885]: I1002 02:57:33.691624 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-xbmgl_d8e59d2f-c3ac-4a7a-b8a1-b1de59eb0724/extract-utilities/0.log" Oct 02 02:57:33 crc kubenswrapper[4885]: I1002 02:57:33.863407 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-sn6xg_56c1dc36-0146-45ea-9396-990a988a6d9e/registry-server/0.log" Oct 02 02:57:33 crc kubenswrapper[4885]: I1002 02:57:33.897040 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-xbmgl_d8e59d2f-c3ac-4a7a-b8a1-b1de59eb0724/extract-content/0.log" Oct 02 02:57:33 crc kubenswrapper[4885]: I1002 02:57:33.914770 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-xbmgl_d8e59d2f-c3ac-4a7a-b8a1-b1de59eb0724/extract-content/0.log" Oct 02 02:57:33 crc kubenswrapper[4885]: I1002 02:57:33.921688 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-xbmgl_d8e59d2f-c3ac-4a7a-b8a1-b1de59eb0724/extract-utilities/0.log" Oct 02 02:57:34 crc kubenswrapper[4885]: I1002 02:57:34.217416 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-xbmgl_d8e59d2f-c3ac-4a7a-b8a1-b1de59eb0724/extract-content/0.log" Oct 02 02:57:34 crc kubenswrapper[4885]: I1002 02:57:34.234433 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-xbmgl_d8e59d2f-c3ac-4a7a-b8a1-b1de59eb0724/extract-utilities/0.log" Oct 02 02:57:34 crc kubenswrapper[4885]: I1002 02:57:34.447691 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ctlhn5_533f5d4b-02ed-4235-8eed-715dd68a58fd/util/0.log" Oct 02 02:57:34 crc kubenswrapper[4885]: I1002 02:57:34.632637 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ctlhn5_533f5d4b-02ed-4235-8eed-715dd68a58fd/pull/0.log" Oct 02 02:57:34 crc kubenswrapper[4885]: I1002 02:57:34.633599 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ctlhn5_533f5d4b-02ed-4235-8eed-715dd68a58fd/util/0.log" Oct 02 02:57:34 crc kubenswrapper[4885]: I1002 02:57:34.679611 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ctlhn5_533f5d4b-02ed-4235-8eed-715dd68a58fd/pull/0.log" Oct 02 02:57:34 crc kubenswrapper[4885]: I1002 02:57:34.695512 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-xbmgl_d8e59d2f-c3ac-4a7a-b8a1-b1de59eb0724/registry-server/0.log" Oct 02 02:57:34 crc kubenswrapper[4885]: I1002 02:57:34.922190 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ctlhn5_533f5d4b-02ed-4235-8eed-715dd68a58fd/pull/0.log" Oct 02 02:57:34 crc kubenswrapper[4885]: I1002 02:57:34.925314 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ctlhn5_533f5d4b-02ed-4235-8eed-715dd68a58fd/util/0.log" Oct 02 02:57:34 crc kubenswrapper[4885]: I1002 02:57:34.969732 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ctlhn5_533f5d4b-02ed-4235-8eed-715dd68a58fd/extract/0.log" Oct 02 02:57:35 crc kubenswrapper[4885]: I1002 02:57:35.089863 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-z7fcl_b0909ffb-180c-4e74-9916-98de47f2419d/marketplace-operator/0.log" Oct 02 02:57:35 crc kubenswrapper[4885]: I1002 02:57:35.102027 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-sbsjt_eb7219fa-b79f-48be-9c5d-6453330d9527/extract-utilities/0.log" Oct 02 02:57:35 crc kubenswrapper[4885]: I1002 02:57:35.306120 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-sbsjt_eb7219fa-b79f-48be-9c5d-6453330d9527/extract-utilities/0.log" Oct 02 02:57:35 crc kubenswrapper[4885]: I1002 02:57:35.319107 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-sbsjt_eb7219fa-b79f-48be-9c5d-6453330d9527/extract-content/0.log" Oct 02 02:57:35 crc kubenswrapper[4885]: I1002 02:57:35.357206 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-sbsjt_eb7219fa-b79f-48be-9c5d-6453330d9527/extract-content/0.log" Oct 02 02:57:35 crc kubenswrapper[4885]: I1002 02:57:35.484478 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-sbsjt_eb7219fa-b79f-48be-9c5d-6453330d9527/extract-content/0.log" Oct 02 02:57:35 crc kubenswrapper[4885]: I1002 02:57:35.484488 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-sbsjt_eb7219fa-b79f-48be-9c5d-6453330d9527/extract-utilities/0.log" Oct 02 02:57:35 crc kubenswrapper[4885]: I1002 02:57:35.631147 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-sbsjt_eb7219fa-b79f-48be-9c5d-6453330d9527/registry-server/0.log" Oct 02 02:57:35 crc kubenswrapper[4885]: I1002 02:57:35.714178 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-9tmpc_747efce9-9b31-4e1b-a7b5-94019bcce7f4/extract-utilities/0.log" Oct 02 02:57:35 crc kubenswrapper[4885]: I1002 02:57:35.805834 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-9tmpc_747efce9-9b31-4e1b-a7b5-94019bcce7f4/extract-utilities/0.log" Oct 02 02:57:35 crc kubenswrapper[4885]: I1002 02:57:35.812537 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-9tmpc_747efce9-9b31-4e1b-a7b5-94019bcce7f4/extract-content/0.log" Oct 02 02:57:35 crc kubenswrapper[4885]: I1002 02:57:35.829302 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-9tmpc_747efce9-9b31-4e1b-a7b5-94019bcce7f4/extract-content/0.log" Oct 02 02:57:36 crc kubenswrapper[4885]: I1002 02:57:36.428505 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-9tmpc_747efce9-9b31-4e1b-a7b5-94019bcce7f4/extract-utilities/0.log" Oct 02 02:57:36 crc kubenswrapper[4885]: I1002 02:57:36.429427 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-9tmpc_747efce9-9b31-4e1b-a7b5-94019bcce7f4/extract-content/0.log" Oct 02 02:57:37 crc kubenswrapper[4885]: I1002 02:57:37.006338 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-9tmpc_747efce9-9b31-4e1b-a7b5-94019bcce7f4/registry-server/0.log" Oct 02 02:57:43 crc kubenswrapper[4885]: I1002 02:57:43.265888 4885 patch_prober.go:28] interesting pod/machine-config-daemon-rttx8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 02:57:43 crc kubenswrapper[4885]: I1002 02:57:43.266650 4885 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 02:58:13 crc kubenswrapper[4885]: I1002 02:58:13.266118 4885 patch_prober.go:28] interesting pod/machine-config-daemon-rttx8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 02:58:13 crc kubenswrapper[4885]: I1002 02:58:13.267100 4885 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 02:58:43 crc kubenswrapper[4885]: I1002 02:58:43.280227 4885 patch_prober.go:28] interesting pod/machine-config-daemon-rttx8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 02:58:43 crc kubenswrapper[4885]: I1002 02:58:43.282435 4885 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 02:58:43 crc kubenswrapper[4885]: I1002 02:58:43.282595 4885 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" Oct 02 02:58:43 crc kubenswrapper[4885]: I1002 02:58:43.284093 4885 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e190b22b6f36adea43879419bc87517eb0762fd9a17a1bf8203497c99cfb5a7f"} pod="openshift-machine-config-operator/machine-config-daemon-rttx8" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 02:58:43 crc kubenswrapper[4885]: I1002 02:58:43.284232 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" containerName="machine-config-daemon" containerID="cri-o://e190b22b6f36adea43879419bc87517eb0762fd9a17a1bf8203497c99cfb5a7f" gracePeriod=600 Oct 02 02:58:43 crc kubenswrapper[4885]: E1002 02:58:43.433440 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:58:44 crc kubenswrapper[4885]: I1002 02:58:44.150802 4885 generic.go:334] "Generic (PLEG): container finished" podID="d169c8ba-000a-4218-8e3b-4ae53035b110" containerID="e190b22b6f36adea43879419bc87517eb0762fd9a17a1bf8203497c99cfb5a7f" exitCode=0 Oct 02 02:58:44 crc kubenswrapper[4885]: I1002 02:58:44.150915 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" event={"ID":"d169c8ba-000a-4218-8e3b-4ae53035b110","Type":"ContainerDied","Data":"e190b22b6f36adea43879419bc87517eb0762fd9a17a1bf8203497c99cfb5a7f"} Oct 02 02:58:44 crc kubenswrapper[4885]: I1002 02:58:44.151167 4885 scope.go:117] "RemoveContainer" containerID="514bc2a11d8477ee5f25c8b041d1d9f8fe40c82d21890ace02ef6041f3710407" Oct 02 02:58:44 crc kubenswrapper[4885]: I1002 02:58:44.152010 4885 scope.go:117] "RemoveContainer" containerID="e190b22b6f36adea43879419bc87517eb0762fd9a17a1bf8203497c99cfb5a7f" Oct 02 02:58:44 crc kubenswrapper[4885]: E1002 02:58:44.152521 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:58:59 crc kubenswrapper[4885]: I1002 02:58:59.068532 4885 scope.go:117] "RemoveContainer" containerID="e190b22b6f36adea43879419bc87517eb0762fd9a17a1bf8203497c99cfb5a7f" Oct 02 02:58:59 crc kubenswrapper[4885]: E1002 02:58:59.069586 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:59:13 crc kubenswrapper[4885]: I1002 02:59:13.046782 4885 scope.go:117] "RemoveContainer" containerID="e190b22b6f36adea43879419bc87517eb0762fd9a17a1bf8203497c99cfb5a7f" Oct 02 02:59:13 crc kubenswrapper[4885]: E1002 02:59:13.047857 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:59:25 crc kubenswrapper[4885]: I1002 02:59:25.046617 4885 scope.go:117] "RemoveContainer" containerID="e190b22b6f36adea43879419bc87517eb0762fd9a17a1bf8203497c99cfb5a7f" Oct 02 02:59:25 crc kubenswrapper[4885]: E1002 02:59:25.047505 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:59:28 crc kubenswrapper[4885]: I1002 02:59:28.741677 4885 generic.go:334] "Generic (PLEG): container finished" podID="a06901b1-5b99-4529-909b-2dfa182f0196" containerID="2b6f9985b46f54a5c49e86cbf2a3ce430da3cd71ff838ecae9d919b670ed9e7c" exitCode=0 Oct 02 02:59:28 crc kubenswrapper[4885]: I1002 02:59:28.741825 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4xx9f/must-gather-fbjpr" event={"ID":"a06901b1-5b99-4529-909b-2dfa182f0196","Type":"ContainerDied","Data":"2b6f9985b46f54a5c49e86cbf2a3ce430da3cd71ff838ecae9d919b670ed9e7c"} Oct 02 02:59:28 crc kubenswrapper[4885]: I1002 02:59:28.743090 4885 scope.go:117] "RemoveContainer" containerID="2b6f9985b46f54a5c49e86cbf2a3ce430da3cd71ff838ecae9d919b670ed9e7c" Oct 02 02:59:29 crc kubenswrapper[4885]: I1002 02:59:29.659806 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-4xx9f_must-gather-fbjpr_a06901b1-5b99-4529-909b-2dfa182f0196/gather/0.log" Oct 02 02:59:36 crc kubenswrapper[4885]: I1002 02:59:36.046968 4885 scope.go:117] "RemoveContainer" containerID="e190b22b6f36adea43879419bc87517eb0762fd9a17a1bf8203497c99cfb5a7f" Oct 02 02:59:36 crc kubenswrapper[4885]: E1002 02:59:36.047743 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 02:59:41 crc kubenswrapper[4885]: I1002 02:59:41.173390 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-4xx9f/must-gather-fbjpr"] Oct 02 02:59:41 crc kubenswrapper[4885]: I1002 02:59:41.174276 4885 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-4xx9f/must-gather-fbjpr" podUID="a06901b1-5b99-4529-909b-2dfa182f0196" containerName="copy" containerID="cri-o://4d031c535ae33d7da37cabe51f3a634fe02db4affc1d117cb08c8012dcf9cdbf" gracePeriod=2 Oct 02 02:59:41 crc kubenswrapper[4885]: I1002 02:59:41.182231 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-4xx9f/must-gather-fbjpr"] Oct 02 02:59:41 crc kubenswrapper[4885]: E1002 02:59:41.207996 4885 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda06901b1_5b99_4529_909b_2dfa182f0196.slice/crio-4d031c535ae33d7da37cabe51f3a634fe02db4affc1d117cb08c8012dcf9cdbf.scope\": RecentStats: unable to find data in memory cache]" Oct 02 02:59:41 crc kubenswrapper[4885]: I1002 02:59:41.687343 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-4xx9f_must-gather-fbjpr_a06901b1-5b99-4529-909b-2dfa182f0196/copy/0.log" Oct 02 02:59:41 crc kubenswrapper[4885]: I1002 02:59:41.688281 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4xx9f/must-gather-fbjpr" Oct 02 02:59:41 crc kubenswrapper[4885]: I1002 02:59:41.837685 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q4pqr\" (UniqueName: \"kubernetes.io/projected/a06901b1-5b99-4529-909b-2dfa182f0196-kube-api-access-q4pqr\") pod \"a06901b1-5b99-4529-909b-2dfa182f0196\" (UID: \"a06901b1-5b99-4529-909b-2dfa182f0196\") " Oct 02 02:59:41 crc kubenswrapper[4885]: I1002 02:59:41.838131 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/a06901b1-5b99-4529-909b-2dfa182f0196-must-gather-output\") pod \"a06901b1-5b99-4529-909b-2dfa182f0196\" (UID: \"a06901b1-5b99-4529-909b-2dfa182f0196\") " Oct 02 02:59:41 crc kubenswrapper[4885]: I1002 02:59:41.852176 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a06901b1-5b99-4529-909b-2dfa182f0196-kube-api-access-q4pqr" (OuterVolumeSpecName: "kube-api-access-q4pqr") pod "a06901b1-5b99-4529-909b-2dfa182f0196" (UID: "a06901b1-5b99-4529-909b-2dfa182f0196"). InnerVolumeSpecName "kube-api-access-q4pqr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:59:41 crc kubenswrapper[4885]: I1002 02:59:41.925908 4885 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-4xx9f_must-gather-fbjpr_a06901b1-5b99-4529-909b-2dfa182f0196/copy/0.log" Oct 02 02:59:41 crc kubenswrapper[4885]: I1002 02:59:41.926320 4885 generic.go:334] "Generic (PLEG): container finished" podID="a06901b1-5b99-4529-909b-2dfa182f0196" containerID="4d031c535ae33d7da37cabe51f3a634fe02db4affc1d117cb08c8012dcf9cdbf" exitCode=143 Oct 02 02:59:41 crc kubenswrapper[4885]: I1002 02:59:41.926379 4885 scope.go:117] "RemoveContainer" containerID="4d031c535ae33d7da37cabe51f3a634fe02db4affc1d117cb08c8012dcf9cdbf" Oct 02 02:59:41 crc kubenswrapper[4885]: I1002 02:59:41.926522 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4xx9f/must-gather-fbjpr" Oct 02 02:59:41 crc kubenswrapper[4885]: I1002 02:59:41.940159 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q4pqr\" (UniqueName: \"kubernetes.io/projected/a06901b1-5b99-4529-909b-2dfa182f0196-kube-api-access-q4pqr\") on node \"crc\" DevicePath \"\"" Oct 02 02:59:41 crc kubenswrapper[4885]: I1002 02:59:41.956078 4885 scope.go:117] "RemoveContainer" containerID="2b6f9985b46f54a5c49e86cbf2a3ce430da3cd71ff838ecae9d919b670ed9e7c" Oct 02 02:59:41 crc kubenswrapper[4885]: I1002 02:59:41.969623 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a06901b1-5b99-4529-909b-2dfa182f0196-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "a06901b1-5b99-4529-909b-2dfa182f0196" (UID: "a06901b1-5b99-4529-909b-2dfa182f0196"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:59:42 crc kubenswrapper[4885]: I1002 02:59:42.032960 4885 scope.go:117] "RemoveContainer" containerID="4d031c535ae33d7da37cabe51f3a634fe02db4affc1d117cb08c8012dcf9cdbf" Oct 02 02:59:42 crc kubenswrapper[4885]: E1002 02:59:42.033401 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4d031c535ae33d7da37cabe51f3a634fe02db4affc1d117cb08c8012dcf9cdbf\": container with ID starting with 4d031c535ae33d7da37cabe51f3a634fe02db4affc1d117cb08c8012dcf9cdbf not found: ID does not exist" containerID="4d031c535ae33d7da37cabe51f3a634fe02db4affc1d117cb08c8012dcf9cdbf" Oct 02 02:59:42 crc kubenswrapper[4885]: I1002 02:59:42.033476 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d031c535ae33d7da37cabe51f3a634fe02db4affc1d117cb08c8012dcf9cdbf"} err="failed to get container status \"4d031c535ae33d7da37cabe51f3a634fe02db4affc1d117cb08c8012dcf9cdbf\": rpc error: code = NotFound desc = could not find container \"4d031c535ae33d7da37cabe51f3a634fe02db4affc1d117cb08c8012dcf9cdbf\": container with ID starting with 4d031c535ae33d7da37cabe51f3a634fe02db4affc1d117cb08c8012dcf9cdbf not found: ID does not exist" Oct 02 02:59:42 crc kubenswrapper[4885]: I1002 02:59:42.033511 4885 scope.go:117] "RemoveContainer" containerID="2b6f9985b46f54a5c49e86cbf2a3ce430da3cd71ff838ecae9d919b670ed9e7c" Oct 02 02:59:42 crc kubenswrapper[4885]: E1002 02:59:42.033838 4885 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2b6f9985b46f54a5c49e86cbf2a3ce430da3cd71ff838ecae9d919b670ed9e7c\": container with ID starting with 2b6f9985b46f54a5c49e86cbf2a3ce430da3cd71ff838ecae9d919b670ed9e7c not found: ID does not exist" containerID="2b6f9985b46f54a5c49e86cbf2a3ce430da3cd71ff838ecae9d919b670ed9e7c" Oct 02 02:59:42 crc kubenswrapper[4885]: I1002 02:59:42.033860 4885 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b6f9985b46f54a5c49e86cbf2a3ce430da3cd71ff838ecae9d919b670ed9e7c"} err="failed to get container status \"2b6f9985b46f54a5c49e86cbf2a3ce430da3cd71ff838ecae9d919b670ed9e7c\": rpc error: code = NotFound desc = could not find container \"2b6f9985b46f54a5c49e86cbf2a3ce430da3cd71ff838ecae9d919b670ed9e7c\": container with ID starting with 2b6f9985b46f54a5c49e86cbf2a3ce430da3cd71ff838ecae9d919b670ed9e7c not found: ID does not exist" Oct 02 02:59:42 crc kubenswrapper[4885]: I1002 02:59:42.041979 4885 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/a06901b1-5b99-4529-909b-2dfa182f0196-must-gather-output\") on node \"crc\" DevicePath \"\"" Oct 02 02:59:42 crc kubenswrapper[4885]: I1002 02:59:42.059930 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a06901b1-5b99-4529-909b-2dfa182f0196" path="/var/lib/kubelet/pods/a06901b1-5b99-4529-909b-2dfa182f0196/volumes" Oct 02 02:59:50 crc kubenswrapper[4885]: I1002 02:59:50.046486 4885 scope.go:117] "RemoveContainer" containerID="e190b22b6f36adea43879419bc87517eb0762fd9a17a1bf8203497c99cfb5a7f" Oct 02 02:59:50 crc kubenswrapper[4885]: E1002 02:59:50.047747 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 03:00:00 crc kubenswrapper[4885]: I1002 03:00:00.185184 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322900-6ww5g"] Oct 02 03:00:00 crc kubenswrapper[4885]: E1002 03:00:00.196755 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0ba71d4-22e1-4d21-8104-05476134a59a" containerName="extract-content" Oct 02 03:00:00 crc kubenswrapper[4885]: I1002 03:00:00.196821 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0ba71d4-22e1-4d21-8104-05476134a59a" containerName="extract-content" Oct 02 03:00:00 crc kubenswrapper[4885]: E1002 03:00:00.196927 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a06901b1-5b99-4529-909b-2dfa182f0196" containerName="gather" Oct 02 03:00:00 crc kubenswrapper[4885]: I1002 03:00:00.196948 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="a06901b1-5b99-4529-909b-2dfa182f0196" containerName="gather" Oct 02 03:00:00 crc kubenswrapper[4885]: E1002 03:00:00.197014 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a06901b1-5b99-4529-909b-2dfa182f0196" containerName="copy" Oct 02 03:00:00 crc kubenswrapper[4885]: I1002 03:00:00.197031 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="a06901b1-5b99-4529-909b-2dfa182f0196" containerName="copy" Oct 02 03:00:00 crc kubenswrapper[4885]: E1002 03:00:00.197103 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0ba71d4-22e1-4d21-8104-05476134a59a" containerName="registry-server" Oct 02 03:00:00 crc kubenswrapper[4885]: I1002 03:00:00.197123 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0ba71d4-22e1-4d21-8104-05476134a59a" containerName="registry-server" Oct 02 03:00:00 crc kubenswrapper[4885]: E1002 03:00:00.197152 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0ba71d4-22e1-4d21-8104-05476134a59a" containerName="extract-utilities" Oct 02 03:00:00 crc kubenswrapper[4885]: I1002 03:00:00.197169 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0ba71d4-22e1-4d21-8104-05476134a59a" containerName="extract-utilities" Oct 02 03:00:00 crc kubenswrapper[4885]: I1002 03:00:00.198495 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="a06901b1-5b99-4529-909b-2dfa182f0196" containerName="copy" Oct 02 03:00:00 crc kubenswrapper[4885]: I1002 03:00:00.198610 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0ba71d4-22e1-4d21-8104-05476134a59a" containerName="registry-server" Oct 02 03:00:00 crc kubenswrapper[4885]: I1002 03:00:00.198699 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="a06901b1-5b99-4529-909b-2dfa182f0196" containerName="gather" Oct 02 03:00:00 crc kubenswrapper[4885]: I1002 03:00:00.201249 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322900-6ww5g" Oct 02 03:00:00 crc kubenswrapper[4885]: I1002 03:00:00.206075 4885 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 02 03:00:00 crc kubenswrapper[4885]: I1002 03:00:00.206928 4885 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 02 03:00:00 crc kubenswrapper[4885]: I1002 03:00:00.229598 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322900-6ww5g"] Oct 02 03:00:00 crc kubenswrapper[4885]: I1002 03:00:00.380160 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/674edbdf-8473-45e6-809b-0b2843ab5337-secret-volume\") pod \"collect-profiles-29322900-6ww5g\" (UID: \"674edbdf-8473-45e6-809b-0b2843ab5337\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322900-6ww5g" Oct 02 03:00:00 crc kubenswrapper[4885]: I1002 03:00:00.380229 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/674edbdf-8473-45e6-809b-0b2843ab5337-config-volume\") pod \"collect-profiles-29322900-6ww5g\" (UID: \"674edbdf-8473-45e6-809b-0b2843ab5337\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322900-6ww5g" Oct 02 03:00:00 crc kubenswrapper[4885]: I1002 03:00:00.380295 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cbdsf\" (UniqueName: \"kubernetes.io/projected/674edbdf-8473-45e6-809b-0b2843ab5337-kube-api-access-cbdsf\") pod \"collect-profiles-29322900-6ww5g\" (UID: \"674edbdf-8473-45e6-809b-0b2843ab5337\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322900-6ww5g" Oct 02 03:00:00 crc kubenswrapper[4885]: I1002 03:00:00.483203 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/674edbdf-8473-45e6-809b-0b2843ab5337-secret-volume\") pod \"collect-profiles-29322900-6ww5g\" (UID: \"674edbdf-8473-45e6-809b-0b2843ab5337\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322900-6ww5g" Oct 02 03:00:00 crc kubenswrapper[4885]: I1002 03:00:00.483588 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/674edbdf-8473-45e6-809b-0b2843ab5337-config-volume\") pod \"collect-profiles-29322900-6ww5g\" (UID: \"674edbdf-8473-45e6-809b-0b2843ab5337\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322900-6ww5g" Oct 02 03:00:00 crc kubenswrapper[4885]: I1002 03:00:00.483800 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cbdsf\" (UniqueName: \"kubernetes.io/projected/674edbdf-8473-45e6-809b-0b2843ab5337-kube-api-access-cbdsf\") pod \"collect-profiles-29322900-6ww5g\" (UID: \"674edbdf-8473-45e6-809b-0b2843ab5337\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322900-6ww5g" Oct 02 03:00:00 crc kubenswrapper[4885]: I1002 03:00:00.485097 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/674edbdf-8473-45e6-809b-0b2843ab5337-config-volume\") pod \"collect-profiles-29322900-6ww5g\" (UID: \"674edbdf-8473-45e6-809b-0b2843ab5337\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322900-6ww5g" Oct 02 03:00:00 crc kubenswrapper[4885]: I1002 03:00:00.493142 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/674edbdf-8473-45e6-809b-0b2843ab5337-secret-volume\") pod \"collect-profiles-29322900-6ww5g\" (UID: \"674edbdf-8473-45e6-809b-0b2843ab5337\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322900-6ww5g" Oct 02 03:00:00 crc kubenswrapper[4885]: I1002 03:00:00.513392 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cbdsf\" (UniqueName: \"kubernetes.io/projected/674edbdf-8473-45e6-809b-0b2843ab5337-kube-api-access-cbdsf\") pod \"collect-profiles-29322900-6ww5g\" (UID: \"674edbdf-8473-45e6-809b-0b2843ab5337\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322900-6ww5g" Oct 02 03:00:00 crc kubenswrapper[4885]: I1002 03:00:00.525207 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322900-6ww5g" Oct 02 03:00:01 crc kubenswrapper[4885]: I1002 03:00:01.046292 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322900-6ww5g"] Oct 02 03:00:01 crc kubenswrapper[4885]: I1002 03:00:01.164974 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29322900-6ww5g" event={"ID":"674edbdf-8473-45e6-809b-0b2843ab5337","Type":"ContainerStarted","Data":"ee6cc9e3a2d93292ff5f021fad838ee2a9738a312c9537b2659c79f8a902bb99"} Oct 02 03:00:01 crc kubenswrapper[4885]: W1002 03:00:01.571918 4885 helpers.go:245] readString: Failed to read "/sys/fs/cgroup/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod674edbdf_8473_45e6_809b_0b2843ab5337.slice/crio-conmon-98ba2bf6f2246c702a34828fd502a6b9b8ad918b37d749041fd4bbb46ecd4ffb.scope/cpuset.cpus.effective": read /sys/fs/cgroup/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod674edbdf_8473_45e6_809b_0b2843ab5337.slice/crio-conmon-98ba2bf6f2246c702a34828fd502a6b9b8ad918b37d749041fd4bbb46ecd4ffb.scope/cpuset.cpus.effective: no such device Oct 02 03:00:01 crc kubenswrapper[4885]: E1002 03:00:01.799928 4885 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod674edbdf_8473_45e6_809b_0b2843ab5337.slice/crio-conmon-98ba2bf6f2246c702a34828fd502a6b9b8ad918b37d749041fd4bbb46ecd4ffb.scope\": RecentStats: unable to find data in memory cache]" Oct 02 03:00:02 crc kubenswrapper[4885]: I1002 03:00:02.062598 4885 scope.go:117] "RemoveContainer" containerID="e190b22b6f36adea43879419bc87517eb0762fd9a17a1bf8203497c99cfb5a7f" Oct 02 03:00:02 crc kubenswrapper[4885]: E1002 03:00:02.065993 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 03:00:02 crc kubenswrapper[4885]: I1002 03:00:02.180067 4885 generic.go:334] "Generic (PLEG): container finished" podID="674edbdf-8473-45e6-809b-0b2843ab5337" containerID="98ba2bf6f2246c702a34828fd502a6b9b8ad918b37d749041fd4bbb46ecd4ffb" exitCode=0 Oct 02 03:00:02 crc kubenswrapper[4885]: I1002 03:00:02.180160 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29322900-6ww5g" event={"ID":"674edbdf-8473-45e6-809b-0b2843ab5337","Type":"ContainerDied","Data":"98ba2bf6f2246c702a34828fd502a6b9b8ad918b37d749041fd4bbb46ecd4ffb"} Oct 02 03:00:03 crc kubenswrapper[4885]: I1002 03:00:03.693691 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322900-6ww5g" Oct 02 03:00:03 crc kubenswrapper[4885]: I1002 03:00:03.866532 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/674edbdf-8473-45e6-809b-0b2843ab5337-config-volume\") pod \"674edbdf-8473-45e6-809b-0b2843ab5337\" (UID: \"674edbdf-8473-45e6-809b-0b2843ab5337\") " Oct 02 03:00:03 crc kubenswrapper[4885]: I1002 03:00:03.866606 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cbdsf\" (UniqueName: \"kubernetes.io/projected/674edbdf-8473-45e6-809b-0b2843ab5337-kube-api-access-cbdsf\") pod \"674edbdf-8473-45e6-809b-0b2843ab5337\" (UID: \"674edbdf-8473-45e6-809b-0b2843ab5337\") " Oct 02 03:00:03 crc kubenswrapper[4885]: I1002 03:00:03.866637 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/674edbdf-8473-45e6-809b-0b2843ab5337-secret-volume\") pod \"674edbdf-8473-45e6-809b-0b2843ab5337\" (UID: \"674edbdf-8473-45e6-809b-0b2843ab5337\") " Oct 02 03:00:03 crc kubenswrapper[4885]: I1002 03:00:03.868759 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/674edbdf-8473-45e6-809b-0b2843ab5337-config-volume" (OuterVolumeSpecName: "config-volume") pod "674edbdf-8473-45e6-809b-0b2843ab5337" (UID: "674edbdf-8473-45e6-809b-0b2843ab5337"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 03:00:03 crc kubenswrapper[4885]: I1002 03:00:03.880815 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/674edbdf-8473-45e6-809b-0b2843ab5337-kube-api-access-cbdsf" (OuterVolumeSpecName: "kube-api-access-cbdsf") pod "674edbdf-8473-45e6-809b-0b2843ab5337" (UID: "674edbdf-8473-45e6-809b-0b2843ab5337"). InnerVolumeSpecName "kube-api-access-cbdsf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:00:03 crc kubenswrapper[4885]: I1002 03:00:03.882336 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/674edbdf-8473-45e6-809b-0b2843ab5337-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "674edbdf-8473-45e6-809b-0b2843ab5337" (UID: "674edbdf-8473-45e6-809b-0b2843ab5337"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:00:03 crc kubenswrapper[4885]: I1002 03:00:03.970938 4885 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/674edbdf-8473-45e6-809b-0b2843ab5337-config-volume\") on node \"crc\" DevicePath \"\"" Oct 02 03:00:03 crc kubenswrapper[4885]: I1002 03:00:03.970982 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cbdsf\" (UniqueName: \"kubernetes.io/projected/674edbdf-8473-45e6-809b-0b2843ab5337-kube-api-access-cbdsf\") on node \"crc\" DevicePath \"\"" Oct 02 03:00:03 crc kubenswrapper[4885]: I1002 03:00:03.970998 4885 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/674edbdf-8473-45e6-809b-0b2843ab5337-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 02 03:00:04 crc kubenswrapper[4885]: I1002 03:00:04.207314 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29322900-6ww5g" event={"ID":"674edbdf-8473-45e6-809b-0b2843ab5337","Type":"ContainerDied","Data":"ee6cc9e3a2d93292ff5f021fad838ee2a9738a312c9537b2659c79f8a902bb99"} Oct 02 03:00:04 crc kubenswrapper[4885]: I1002 03:00:04.207785 4885 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ee6cc9e3a2d93292ff5f021fad838ee2a9738a312c9537b2659c79f8a902bb99" Oct 02 03:00:04 crc kubenswrapper[4885]: I1002 03:00:04.208011 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322900-6ww5g" Oct 02 03:00:04 crc kubenswrapper[4885]: I1002 03:00:04.792795 4885 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322855-22v6b"] Oct 02 03:00:04 crc kubenswrapper[4885]: I1002 03:00:04.819716 4885 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322855-22v6b"] Oct 02 03:00:06 crc kubenswrapper[4885]: I1002 03:00:06.070389 4885 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6954232a-a827-4d3b-abf7-e552100fd000" path="/var/lib/kubelet/pods/6954232a-a827-4d3b-abf7-e552100fd000/volumes" Oct 02 03:00:15 crc kubenswrapper[4885]: I1002 03:00:15.047655 4885 scope.go:117] "RemoveContainer" containerID="e190b22b6f36adea43879419bc87517eb0762fd9a17a1bf8203497c99cfb5a7f" Oct 02 03:00:15 crc kubenswrapper[4885]: E1002 03:00:15.048977 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 03:00:15 crc kubenswrapper[4885]: I1002 03:00:15.895549 4885 scope.go:117] "RemoveContainer" containerID="69dff0f53d137fa8c49e2b7a285c1e7cf30a27dfaf3772c78aa36403cd87322d" Oct 02 03:00:16 crc kubenswrapper[4885]: I1002 03:00:16.000872 4885 scope.go:117] "RemoveContainer" containerID="66c41f69fd3d9160849345a74f233f238af50c6c776d9b1ca8e63877e1f24181" Oct 02 03:00:30 crc kubenswrapper[4885]: I1002 03:00:30.057046 4885 scope.go:117] "RemoveContainer" containerID="e190b22b6f36adea43879419bc87517eb0762fd9a17a1bf8203497c99cfb5a7f" Oct 02 03:00:30 crc kubenswrapper[4885]: E1002 03:00:30.058416 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 03:00:41 crc kubenswrapper[4885]: I1002 03:00:41.047833 4885 scope.go:117] "RemoveContainer" containerID="e190b22b6f36adea43879419bc87517eb0762fd9a17a1bf8203497c99cfb5a7f" Oct 02 03:00:41 crc kubenswrapper[4885]: E1002 03:00:41.049061 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 03:00:52 crc kubenswrapper[4885]: I1002 03:00:52.061074 4885 scope.go:117] "RemoveContainer" containerID="e190b22b6f36adea43879419bc87517eb0762fd9a17a1bf8203497c99cfb5a7f" Oct 02 03:00:52 crc kubenswrapper[4885]: E1002 03:00:52.064533 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 03:01:00 crc kubenswrapper[4885]: I1002 03:01:00.169201 4885 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29322901-bsn7l"] Oct 02 03:01:00 crc kubenswrapper[4885]: E1002 03:01:00.171273 4885 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="674edbdf-8473-45e6-809b-0b2843ab5337" containerName="collect-profiles" Oct 02 03:01:00 crc kubenswrapper[4885]: I1002 03:01:00.171372 4885 state_mem.go:107] "Deleted CPUSet assignment" podUID="674edbdf-8473-45e6-809b-0b2843ab5337" containerName="collect-profiles" Oct 02 03:01:00 crc kubenswrapper[4885]: I1002 03:01:00.171679 4885 memory_manager.go:354] "RemoveStaleState removing state" podUID="674edbdf-8473-45e6-809b-0b2843ab5337" containerName="collect-profiles" Oct 02 03:01:00 crc kubenswrapper[4885]: I1002 03:01:00.177865 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29322901-bsn7l" Oct 02 03:01:00 crc kubenswrapper[4885]: I1002 03:01:00.182867 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29322901-bsn7l"] Oct 02 03:01:00 crc kubenswrapper[4885]: I1002 03:01:00.258722 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7bb19cf9-8a8f-4871-b015-2d5a96c81f3e-config-data\") pod \"keystone-cron-29322901-bsn7l\" (UID: \"7bb19cf9-8a8f-4871-b015-2d5a96c81f3e\") " pod="openstack/keystone-cron-29322901-bsn7l" Oct 02 03:01:00 crc kubenswrapper[4885]: I1002 03:01:00.258826 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g4grx\" (UniqueName: \"kubernetes.io/projected/7bb19cf9-8a8f-4871-b015-2d5a96c81f3e-kube-api-access-g4grx\") pod \"keystone-cron-29322901-bsn7l\" (UID: \"7bb19cf9-8a8f-4871-b015-2d5a96c81f3e\") " pod="openstack/keystone-cron-29322901-bsn7l" Oct 02 03:01:00 crc kubenswrapper[4885]: I1002 03:01:00.259034 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bb19cf9-8a8f-4871-b015-2d5a96c81f3e-combined-ca-bundle\") pod \"keystone-cron-29322901-bsn7l\" (UID: \"7bb19cf9-8a8f-4871-b015-2d5a96c81f3e\") " pod="openstack/keystone-cron-29322901-bsn7l" Oct 02 03:01:00 crc kubenswrapper[4885]: I1002 03:01:00.259448 4885 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7bb19cf9-8a8f-4871-b015-2d5a96c81f3e-fernet-keys\") pod \"keystone-cron-29322901-bsn7l\" (UID: \"7bb19cf9-8a8f-4871-b015-2d5a96c81f3e\") " pod="openstack/keystone-cron-29322901-bsn7l" Oct 02 03:01:00 crc kubenswrapper[4885]: I1002 03:01:00.362230 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7bb19cf9-8a8f-4871-b015-2d5a96c81f3e-config-data\") pod \"keystone-cron-29322901-bsn7l\" (UID: \"7bb19cf9-8a8f-4871-b015-2d5a96c81f3e\") " pod="openstack/keystone-cron-29322901-bsn7l" Oct 02 03:01:00 crc kubenswrapper[4885]: I1002 03:01:00.362352 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g4grx\" (UniqueName: \"kubernetes.io/projected/7bb19cf9-8a8f-4871-b015-2d5a96c81f3e-kube-api-access-g4grx\") pod \"keystone-cron-29322901-bsn7l\" (UID: \"7bb19cf9-8a8f-4871-b015-2d5a96c81f3e\") " pod="openstack/keystone-cron-29322901-bsn7l" Oct 02 03:01:00 crc kubenswrapper[4885]: I1002 03:01:00.362423 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bb19cf9-8a8f-4871-b015-2d5a96c81f3e-combined-ca-bundle\") pod \"keystone-cron-29322901-bsn7l\" (UID: \"7bb19cf9-8a8f-4871-b015-2d5a96c81f3e\") " pod="openstack/keystone-cron-29322901-bsn7l" Oct 02 03:01:00 crc kubenswrapper[4885]: I1002 03:01:00.362528 4885 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7bb19cf9-8a8f-4871-b015-2d5a96c81f3e-fernet-keys\") pod \"keystone-cron-29322901-bsn7l\" (UID: \"7bb19cf9-8a8f-4871-b015-2d5a96c81f3e\") " pod="openstack/keystone-cron-29322901-bsn7l" Oct 02 03:01:00 crc kubenswrapper[4885]: I1002 03:01:00.371253 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bb19cf9-8a8f-4871-b015-2d5a96c81f3e-combined-ca-bundle\") pod \"keystone-cron-29322901-bsn7l\" (UID: \"7bb19cf9-8a8f-4871-b015-2d5a96c81f3e\") " pod="openstack/keystone-cron-29322901-bsn7l" Oct 02 03:01:00 crc kubenswrapper[4885]: I1002 03:01:00.371582 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7bb19cf9-8a8f-4871-b015-2d5a96c81f3e-config-data\") pod \"keystone-cron-29322901-bsn7l\" (UID: \"7bb19cf9-8a8f-4871-b015-2d5a96c81f3e\") " pod="openstack/keystone-cron-29322901-bsn7l" Oct 02 03:01:00 crc kubenswrapper[4885]: I1002 03:01:00.375902 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7bb19cf9-8a8f-4871-b015-2d5a96c81f3e-fernet-keys\") pod \"keystone-cron-29322901-bsn7l\" (UID: \"7bb19cf9-8a8f-4871-b015-2d5a96c81f3e\") " pod="openstack/keystone-cron-29322901-bsn7l" Oct 02 03:01:00 crc kubenswrapper[4885]: I1002 03:01:00.388865 4885 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g4grx\" (UniqueName: \"kubernetes.io/projected/7bb19cf9-8a8f-4871-b015-2d5a96c81f3e-kube-api-access-g4grx\") pod \"keystone-cron-29322901-bsn7l\" (UID: \"7bb19cf9-8a8f-4871-b015-2d5a96c81f3e\") " pod="openstack/keystone-cron-29322901-bsn7l" Oct 02 03:01:00 crc kubenswrapper[4885]: I1002 03:01:00.529992 4885 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29322901-bsn7l" Oct 02 03:01:01 crc kubenswrapper[4885]: I1002 03:01:01.069717 4885 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29322901-bsn7l"] Oct 02 03:01:01 crc kubenswrapper[4885]: I1002 03:01:01.925506 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29322901-bsn7l" event={"ID":"7bb19cf9-8a8f-4871-b015-2d5a96c81f3e","Type":"ContainerStarted","Data":"aa8aa3f7c1b9b1ed85546a48c85fb6ae525872f3c2fe2ca1639872ee6ec3f73a"} Oct 02 03:01:01 crc kubenswrapper[4885]: I1002 03:01:01.925643 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29322901-bsn7l" event={"ID":"7bb19cf9-8a8f-4871-b015-2d5a96c81f3e","Type":"ContainerStarted","Data":"d1c9045a4e3a6de80fb4c8b48a73fbe9da6f1a1605c9f4b69cad331b589c5365"} Oct 02 03:01:01 crc kubenswrapper[4885]: I1002 03:01:01.954079 4885 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29322901-bsn7l" podStartSLOduration=1.954058407 podStartE2EDuration="1.954058407s" podCreationTimestamp="2025-10-02 03:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 03:01:01.949978653 +0000 UTC m=+4450.761726082" watchObservedRunningTime="2025-10-02 03:01:01.954058407 +0000 UTC m=+4450.765805816" Oct 02 03:01:03 crc kubenswrapper[4885]: I1002 03:01:03.951532 4885 generic.go:334] "Generic (PLEG): container finished" podID="7bb19cf9-8a8f-4871-b015-2d5a96c81f3e" containerID="aa8aa3f7c1b9b1ed85546a48c85fb6ae525872f3c2fe2ca1639872ee6ec3f73a" exitCode=0 Oct 02 03:01:03 crc kubenswrapper[4885]: I1002 03:01:03.952587 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29322901-bsn7l" event={"ID":"7bb19cf9-8a8f-4871-b015-2d5a96c81f3e","Type":"ContainerDied","Data":"aa8aa3f7c1b9b1ed85546a48c85fb6ae525872f3c2fe2ca1639872ee6ec3f73a"} Oct 02 03:01:04 crc kubenswrapper[4885]: I1002 03:01:04.046878 4885 scope.go:117] "RemoveContainer" containerID="e190b22b6f36adea43879419bc87517eb0762fd9a17a1bf8203497c99cfb5a7f" Oct 02 03:01:04 crc kubenswrapper[4885]: E1002 03:01:04.047300 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 03:01:05 crc kubenswrapper[4885]: I1002 03:01:05.365462 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29322901-bsn7l" Oct 02 03:01:05 crc kubenswrapper[4885]: I1002 03:01:05.477101 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7bb19cf9-8a8f-4871-b015-2d5a96c81f3e-config-data\") pod \"7bb19cf9-8a8f-4871-b015-2d5a96c81f3e\" (UID: \"7bb19cf9-8a8f-4871-b015-2d5a96c81f3e\") " Oct 02 03:01:05 crc kubenswrapper[4885]: I1002 03:01:05.477595 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bb19cf9-8a8f-4871-b015-2d5a96c81f3e-combined-ca-bundle\") pod \"7bb19cf9-8a8f-4871-b015-2d5a96c81f3e\" (UID: \"7bb19cf9-8a8f-4871-b015-2d5a96c81f3e\") " Oct 02 03:01:05 crc kubenswrapper[4885]: I1002 03:01:05.477813 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g4grx\" (UniqueName: \"kubernetes.io/projected/7bb19cf9-8a8f-4871-b015-2d5a96c81f3e-kube-api-access-g4grx\") pod \"7bb19cf9-8a8f-4871-b015-2d5a96c81f3e\" (UID: \"7bb19cf9-8a8f-4871-b015-2d5a96c81f3e\") " Oct 02 03:01:05 crc kubenswrapper[4885]: I1002 03:01:05.478051 4885 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7bb19cf9-8a8f-4871-b015-2d5a96c81f3e-fernet-keys\") pod \"7bb19cf9-8a8f-4871-b015-2d5a96c81f3e\" (UID: \"7bb19cf9-8a8f-4871-b015-2d5a96c81f3e\") " Oct 02 03:01:05 crc kubenswrapper[4885]: I1002 03:01:05.485885 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7bb19cf9-8a8f-4871-b015-2d5a96c81f3e-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "7bb19cf9-8a8f-4871-b015-2d5a96c81f3e" (UID: "7bb19cf9-8a8f-4871-b015-2d5a96c81f3e"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:01:05 crc kubenswrapper[4885]: I1002 03:01:05.488209 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb19cf9-8a8f-4871-b015-2d5a96c81f3e-kube-api-access-g4grx" (OuterVolumeSpecName: "kube-api-access-g4grx") pod "7bb19cf9-8a8f-4871-b015-2d5a96c81f3e" (UID: "7bb19cf9-8a8f-4871-b015-2d5a96c81f3e"). InnerVolumeSpecName "kube-api-access-g4grx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:01:05 crc kubenswrapper[4885]: I1002 03:01:05.529360 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7bb19cf9-8a8f-4871-b015-2d5a96c81f3e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7bb19cf9-8a8f-4871-b015-2d5a96c81f3e" (UID: "7bb19cf9-8a8f-4871-b015-2d5a96c81f3e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:01:05 crc kubenswrapper[4885]: I1002 03:01:05.573445 4885 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7bb19cf9-8a8f-4871-b015-2d5a96c81f3e-config-data" (OuterVolumeSpecName: "config-data") pod "7bb19cf9-8a8f-4871-b015-2d5a96c81f3e" (UID: "7bb19cf9-8a8f-4871-b015-2d5a96c81f3e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:01:05 crc kubenswrapper[4885]: I1002 03:01:05.581518 4885 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7bb19cf9-8a8f-4871-b015-2d5a96c81f3e-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 02 03:01:05 crc kubenswrapper[4885]: I1002 03:01:05.581812 4885 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7bb19cf9-8a8f-4871-b015-2d5a96c81f3e-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 03:01:05 crc kubenswrapper[4885]: I1002 03:01:05.582150 4885 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bb19cf9-8a8f-4871-b015-2d5a96c81f3e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 03:01:05 crc kubenswrapper[4885]: I1002 03:01:05.583196 4885 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g4grx\" (UniqueName: \"kubernetes.io/projected/7bb19cf9-8a8f-4871-b015-2d5a96c81f3e-kube-api-access-g4grx\") on node \"crc\" DevicePath \"\"" Oct 02 03:01:05 crc kubenswrapper[4885]: I1002 03:01:05.979359 4885 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29322901-bsn7l" Oct 02 03:01:05 crc kubenswrapper[4885]: I1002 03:01:05.979348 4885 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29322901-bsn7l" event={"ID":"7bb19cf9-8a8f-4871-b015-2d5a96c81f3e","Type":"ContainerDied","Data":"d1c9045a4e3a6de80fb4c8b48a73fbe9da6f1a1605c9f4b69cad331b589c5365"} Oct 02 03:01:05 crc kubenswrapper[4885]: I1002 03:01:05.980004 4885 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d1c9045a4e3a6de80fb4c8b48a73fbe9da6f1a1605c9f4b69cad331b589c5365" Oct 02 03:01:19 crc kubenswrapper[4885]: I1002 03:01:19.048371 4885 scope.go:117] "RemoveContainer" containerID="e190b22b6f36adea43879419bc87517eb0762fd9a17a1bf8203497c99cfb5a7f" Oct 02 03:01:19 crc kubenswrapper[4885]: E1002 03:01:19.049494 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" Oct 02 03:01:31 crc kubenswrapper[4885]: I1002 03:01:31.047058 4885 scope.go:117] "RemoveContainer" containerID="e190b22b6f36adea43879419bc87517eb0762fd9a17a1bf8203497c99cfb5a7f" Oct 02 03:01:31 crc kubenswrapper[4885]: E1002 03:01:31.048136 4885 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rttx8_openshift-machine-config-operator(d169c8ba-000a-4218-8e3b-4ae53035b110)\"" pod="openshift-machine-config-operator/machine-config-daemon-rttx8" podUID="d169c8ba-000a-4218-8e3b-4ae53035b110" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515067365432024460 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015067365432017375 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015067354264016521 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015067354264015471 5ustar corecore